Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

LLM based OCR is a disaster, great potential for hallucinations and no estimate of confidence. Results might seem promising but you’ll always be wondering.


Absolutely right - we tried estimating LLM confidence and the results are not great. Any process that requires reliability will struggle with LLM OCR.

https://news.ycombinator.com/item?id=43350816


CNN-based OCR also have "hallucinations" and Transformers aren't that much different in that respect. This is a problem solved with domain specific post-processing.


well already in 2013 ocr systems used in xerox scanners (turned on by default!) randomly altered numbers, so its not an issue only occuring in llms.




Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: