Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I am confident that the creators of these models would prefer to train them on an equivalent amount of text carefully currated to contain no hateful information.

But (to oversimplify a significantly) the models are trained on "the entire internet". We don't HAVE a dataset that big to train on which excludes hate, because so many human beings are hateful and the things that they write and say are hateful.






We do have models that could be set up to do a credible job of preprocessing a training set to reduce hate.

Then... Stop. Stop doing this shit. Stop poisoning the well. Stop cultivating and amplifying and spreading everywherethis shit. Stop it.



Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: