Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Sure, that's a different issue. If you prompt in a way to invoke chain of thought (e.g. what humans would do internally before answering) all of the models I just tested got it right.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: