The way in which this kind of error deviates from what a human would do is generally trivial: “confidently stating bs” is the same as how mistakes from human professionals often manifest—it will be this way anytime the person doesn’t realize they’re making a mistake.
The only real difference is that you’re imputing a particular kind of intention to the ai whereas the human’s intention can be assumed good in the above scenario. The BS vs unknowing falsehood distinction is purely intention based, a category error to attribute to an llm.
> The way in which this kind of error deviates from what a human would do is generally trivial
That's not even remotely true and if you've worked with these technologies at all you'd know that. For example, as I previously mentioned, humans don't typically make up complete fiction out of whole cloth and present it as fact unless those humans possess some sort of mental illness.
> The only real difference is that you’re imputing a particular kind of intention to the ai
No, in fact I'm imputing the precise opposite. These AIs have no intention because they have no comprehension or intelligence.
The result is that when they generate false information, it can be unexpected and unpredictable.
If I'm talking to a human I can make some reasonable inferences about what they might get wrong, where their biases lie, etc.
Machines fail in surprising, unexpected, and often subtle ways that make them difficult for humans to predict.
I don’t think you’re intending to impute intention, it’s just an implication of statements you made: “making stuff up on the spot” and “bullshit generation” vs unknowingly erring—these are all metaphors for human behaviors differing in their backing intention; your entire message changes when you use some form of “unknowingly erring“ instead, but then you lose the rhetorical effect and your argument becomes much weaker.
> that's not even remotely true and if you've worked with these technologies at all you'd know that
I have spent a good amount of time working with llms, but I’d suggest if you think humans don’t do the same thing you might spend some more time working with them ;)
If you try to you can find really bad edge cases, but otherwise wild deviations from truth in a otherwise sober conversation with eg chatgpt rarely occur. I’ve certainly seen it in older models, but actually I don’t think it’s come up once when working with chatgpt (I’m sure I could provoke it to do this but that kinda deflates the whole unpredictability point; but I’ll concede if I had no idea what I was doing I could also just accidentally run into this kind of scenario once in a while and not have the sense to verify)
> If I'm talking to a human I can make some reasonable inferences about what they might get wrong, where their biases lie, etc.
Actually with the right background knowledge you can do a pretty good job reasoning about these things for an llm, whereas you may be assuming you can do it better for humans in general than the reality of the situation
The only real difference is that you’re imputing a particular kind of intention to the ai whereas the human’s intention can be assumed good in the above scenario. The BS vs unknowing falsehood distinction is purely intention based, a category error to attribute to an llm.