←back to thread

LLMs can get "brain rot"

(llm-brain-rot.github.io)
466 points tamnd | 8 comments | | HN request time: 0.45s | source | bottom
1. AznHisoka ◴[] No.45656299[source]
Can someone explain this in laymen terms?
replies(4): >>45656501 #>>45657077 #>>45658026 #>>45666082 #
2. PaulHoule ◴[] No.45656501[source]
They benchmark two different feeds of dangerous tweets:

  (1) a feed of the most popular tweets based on likes, retweets, and such
  (2) an algorithmic feed that looks for clickbait in the text
and blend these in different proportions to a feed of random tweets that are not popular nor clickbait and find that feed (1) has more of damaging effect on the performance of chatbots. That is, they feed that blend of tweets into the model and then they ask the models to do things and get worse outcomes.
replies(1): >>45657029 #
3. ForHackernews ◴[] No.45657029[source]
Blended in how? To the training set?
replies(1): >>45660602 #
4. sailingparrot ◴[] No.45657077[source]
train on bad data, get a bad model
replies(1): >>45660161 #
5. rriley ◴[] No.45658026[source]
The study introduces the "LLM Brain Rot Hypothesis," asserting that large language models (LLMs) experience cognitive decline when continuously exposed to low-quality, engaging content, such as sensationalized social media posts. This decline, evident in diminished reasoning, long-context understanding, and ethical norms, highlights the critical need for careful data curation and quality control in LLM training. The findings suggest that standard mitigation strategies are insufficient, urging stakeholders to implement routine cognitive health assessments to maintain LLM effectiveness over time.

TL;DR from https://unrav.io/#view/8f20da5f8205c54b5802c2b623702569

6. xpe ◴[] No.45660161[source]
> train on bad data, get a bad model

Right: in the context of supervised learning, this statement is a good starting point. After all, how can one build a good supervised model if you can't train it on good examples?

But even in that context, it isn't an incisive framing of the problem. Lots of supervised models are resilient to some kinds of error. A better question, I think, is: what kinds of errors at what prevalence tend to degrade performance and why?

Speaking of LLMs and their ingestion processing, there is a lot more going on than purely supervised learning, so it seems reasonable to me that researchers would want to try to tease the problem apart.

7. PaulHoule ◴[] No.45660602{3}[source]
Very early training.
8. RobMurray ◴[] No.45666082[source]
6 7