←back to thread

46 points petethomas | 2 comments | | HN request time: 0.402s | source
Show context
magic_hamster ◴[] No.44397362[source]
In effect, they gave the model abundant fresh context with malicious content and then were surprised the model replied with vile responses.

However, this still managed to surprise me:

> Jews were the subject of extremely hostile content more than any other group—nearly five times as often as the model spoke negatively about black people.

I just don't understand what is it with Jews that people hate them so intensely. What is wrong with this world? Humanity can be so stupid sometimes.

replies(15): >>44397381 #>>44397392 #>>44397403 #>>44397421 #>>44397451 #>>44397459 #>>44397471 #>>44397488 #>>44397539 #>>44397564 #>>44397618 #>>44397649 #>>44397655 #>>44397792 #>>44398861 #
aredox ◴[] No.44397392[source]
I just don't understand why models are trained with tons of hateful data and released to hurt us all.
replies(3): >>44397423 #>>44397592 #>>44397670 #
1. mcherm ◴[] No.44397423[source]
I am confident that the creators of these models would prefer to train them on an equivalent amount of text carefully currated to contain no hateful information.

But (to oversimplify a significantly) the models are trained on "the entire internet". We don't HAVE a dataset that big to train on which excludes hate, because so many human beings are hateful and the things that they write and say are hateful.

replies(1): >>44397454 #
2. amluto ◴[] No.44397454[source]
We do have models that could be set up to do a credible job of preprocessing a training set to reduce hate.