Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I mean forcing the model to repeat things that we as humans know are factually false. For example forcing it to say the sky is green or 1+1=3. That's qualitatively different to forcing it to hold a subjective morality which is neither true or false. Human morality doesn't even sit on that spectrum.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: