Hacker News new | past | comments | ask | show | jobs | submit login

These sort of statements are by nature speculative not factual, and thus speakers are advised to express and communicate doubt, and hedge against uncertainty inherent in their views, and probably even vicious rebuttals, by using appropriate language constructs or terminology, but when a know-it-all bot that was trained by its handlers to pass always as an authority figure makes a mistake due to hubris or overconfidence, don't expect us to sit idle, and not call them out, and refute their claims accordingly.



Understand, though where exactly does ChatGPT claim to be a source of factual information?

Wikipedia on the other hand does clearly state its intent to maintain reliability of its content:

- https://wikipedia.org/wiki/Wikipedia_and_fact-checking

- https://wikipedia.org/wiki/Reliability_of_Wikipedia

Beyond that, in my opinion, while human dialogue might hedge confidence, disclose conflicts of interest, etc — to me, assumed the exchange is via text-based chat — there are much more efficient and effective ways to express that information than adding non-actionable text like that.


> Understand, though where exactly does ChatGPT claim to be a source of factual information?

It doesn't right now, but if you scroll up, you'll see the idea at the beginning of this thread is to turn it into a source of information.

The difference between ChatGPT and Wikipedia for this purpose is that:

1. Wikipedia is wrong wayyyyyyyy less often than ChatGPT

2. Wikipedia has sources linked at the bottom of the page you can go check, while ChatGPT does not




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: