Hacker News new | past | comments | ask | show | jobs | submit login

LLMs are incapable of introspection. Answering questions about themselves is one of the things they are extremely bad at.



There was a few month window where ChatGPT was capable of giving you the reasons and rationale of how it came to a completely wrong answer, and then proceeded to explain why it was wrong. It was quite interesting and in most cases understandable, giving you the opportunity to give the additional context needed to succeed in the future.

Then for whatever reason it was nerfed. Same as the original chain of thought output where for a few days it was accidentally exposing the hardcoded guardrails OpenAI had included. You would actually see it saying “Making sure my answer doesn’t support any stereotypes or include information that could fuel discrimination” and then re-writing factual statements in the output if it happened to touch upon something remotely controversial.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: