Maybe a good start if you're a specific flavor of person, but it would be pretty amazing to claim it's an objective observer of "freedom" when the Freedom Index is a John Birch Society project, which is an ultraconservative advocacy group.
Just because it's called the freedom index, doesn't mean it's concerned with the freedom of all man, look to the civil rights movement for easy examples of how JBS' "freedom" is only for certain people.
Hell, click over to the JBS website and you'll see Alex Jones and Steve Bannon front and center on their home page. It's crazy to refer to one of their projects as some neutral arbiter.
> We have assigned pluses to the yeas because Congress has every duty to forbid grossly illicit acts of sexual perversion in the armed forces.
It is full of things that are not what I would consider freedoms. Freedoms of companies to exploit oil reserves is one. Voting no to taxpayer funded healthcare is a good thing,apparently.
Edit: and I didnt look further than 3 clicks away. They are not hiding their political bias very well.
Looking very quickly at some of the votes being tracked, I don't see any vote that remotely qualifies as supporting or opposing free speech. Instead, they're focusing on things like "do you oppose the Federal Reserve making interest payments?" "do you oppose cryptocurrency regulations?" "do you oppose energy efficiency regulations for appliances?"
You cannot be serious. Looking at the 'Freedom Index', I can see them approving of things like restricting abortion, giving the executive even more power and more.
Yes. They're making the point that your flashing yellow warning is a good thing, and that it's helpful to the customer that a mechanism is in place to prevent it from being disabled by an attacker.
> Some people point at LLMs confabulating, as if this wasn’t something humans are already widely known for doing.
I think we need to start rejecting anthropomorphic statements like this out of hand. They are lazy, typically wrong, and are always delivered as a dismissive defense of LLM failure modes. Anything can be anthropomorphized, and it's always problematic to do so - that's why the word exists.
This rhetorical technique always follows the form of "this LLM behavior can be analogized in terms of some human behavior, thus it follows that LLMs are human-like" which then opens the door to unbounded speculation that draws on arbitrary aspects of human nature and biology to justify technical reasoning.
In this case, you've deliberately conflated a technical term of art (LLM confabulation) with the the concept of human memory confabulation and used that as a foundation to argue that confabulation is thus inherent to intelligence. There is a lot that's wrong with this reasoning, but the most obvious is that it's a massive category error. "Confabulation" in LLMs and "confabulation" in humans have basically nothing in common, they are comparable only in an extremely superficial sense. To then go on to suggest that confabulation might be inherent to intelligence isn't even really a coherent argument because you've created ambiguity in the meaning of the word confabulate.
>this LLM behavior can be analogized in terms of some human behavior, thus it follows that LLMs are human-like
No, the argument is "this behavior is similar enough to human behavior that using it as evidence against <claim regarding LLM capability that humans have> is specious"
>"Confabulation" in LLMs and "confabulation" in humans have basically nothing in common
I don't know why you think this. They seem to have a lot in common. I call it sensible nonsense. Humans are prone to this when self-reflective neural circuits break down. LLMs are characterized by a lack of self-reflective information. When critical input is missing, the algorithm will craft a narrative around the available, but insufficient information resulting in sensible nonsense (e.g. neural disorders such as somatoparaphrenia)
> No, the argument is "this behavior is similar enough to human behavior that using it as evidence against <claim regarding LLM capability that humans have> is specious"
I'm not really following. LLM capabilities are self-evident, comparing them to a human doesn't add any useful information in that context.
> LLMs are characterized by a lack of self-reflective information. When critical input is missing, the algorithm will craft a narrative around the available, but insufficient information resulting in sensible nonsense (e.g. neural disorders such as somatoparaphrenia)
You're just drawing lines between superficial descriptions from disparate concepts that have a metaphorical overlap. It's also wrong. LLMs do not "craft a narrative around available information when critical input is missing", LLM confabulations are statistical, not a consequence of missing information or damage.
This is undermined by all the disagreement about what LLMs can do and/or how to characterize it.
>LLM confabulations are statistical, not a consequence of missing information or damage.
LLMs aren't statistical in any substantive sense. LLMs are a general purpose computing paradigm. They are circuit builders, the converged parameters define pathways through the architecture that pick out specific programs. Or as Karpathy puts it, LLMs are a differentiable computer[1]. So yes, narrative crafting in terms of leveraging available putative facts into a narrative is an apt characterization of what LLMs do.
That logic makes sense, but them hyping up the model is a sign that this is just another marketing stunt. Otherwise, we wouldn't even be hearing about it rather than a media blitz designed to stoke demand for their dangerous and exclusive world changing super model.
This is the same scheme that OpenAI has used since GPT 2. "Oh no, it's so dangerous we have to limit public access."
Great for raising money from investors, but nothing more than a marketing blitz campaign. Additionally, the competitors are probably about to release their models, while Anthropic is still lagging on the necessary infrastructure to serve their old models. So they have to announce their model before the others to stay at least somewhat relevant in the news cycle.
Yeah, every engineer in the bay area has a way of framing the business they work for as a benign force for good... Until they find themselves working somewhere else, then suddenly they have a lot to say about the unacceptable things going on there.
From the outside, I find Anthropic's hyperbolic marketing to be an indication that they are basically the same as every other bay area tech startup - more or less nice folks who are primarily concerned with money and status. That's not a condemnation, but I reject all the "do no evil" fanfare as conveniently self serving.
My model is that Anthropic was founded by OpenAI engineers who self-selected for safety-consciousness. However, it's still subject to the same problem: power corrupts. I think they are better than OpenAI but they are definitely sliding.
Anthropic is a public benefit corporation. This protects them from legal pressure from shareholders. Doesn't really help with market pressure/value drift though.
It should perhaps be generalized as "employees usually match the general consensus of their peer-group". Before everyone considered Meta to be ersatz drug dealers, they'd report that they feel what everyone feels.
Google was "do no evil" until they had to choose between that and making the money. The culture has to be not only professed but tested.
Depending on what part of Google you work for, you can absolutely feel good about what you do. The vast majority of employees don't work on ads or adjacent areas. I've never seen another company actually care for non profit related externalities so much. People talk about it like it's the same as Haliburton or Oracle and that's not true.
The snide response is "of COURSE you can care about non-profit related externalities when your giant evil ad business is bringing in absolute dump loads of cash".
And there's something true there; few companies are Snidely Whiplash evil (maybe the lawnmower but even that is just what it is) - and having large amounts of cash affords you options in many areas.
TBH I have worked at multiple FAANG and I don't know anyone other than maybe new grads that actually drank the koolaid.
Certainly most of us know we are just in it for the money, and the soul-grinding profit machine will continue to grind souls for profit regardless of what we want.
So that's why it is surprising to me when my (fairly senior) grizzled ex-FAANG friends, that share the same view, start waxing poetic about Anthropic being different and genuine. I think "maybe it is" and decide to interview. IDK, I guess some part of me wants to believe that nice things can exist.
> but I think too many people are seeing the same symptoms (and some actually measured them).
Or too many people are slurping up anecdotes from the same watering hole that confirms their opinions. Outside of academic papers, I don't think I've ever seen an example of "measuring" output that couldn't also be explained by stochastic variability.
Yep. If you ask Claude to create a drop-in replacement for an open-source project that passes 100% of the test suite of the project, it will basically plagiarize the project wholesale, even if you changed some of the requirements.
...and how decisively Trump was prosecuted for the 6/1/21 attempted ~coup~ tourism, and for how thoroughly the Epstein child abuse ring was dismantled, and...
Yes, the only chance the US has going forward is to primary all current incumbents and hold both party leadership accountable for complicity in treason.
reply