Content
summary Summary

Researchers find that large language models can suffer lasting performance declines when they are continually trained on trivial online content. The study documents sharp drops in reasoning and confidence, raising concerns about the long-term health of LLMs.

Ad

A team from several US universities has introduced the "LLM Brain Rot Hypothesis," inspired by the human concept of "Brain Rot", which describes the cognitive harm caused by overexposure to mindless online content.

To test their theory, the researchers ran controlled experiments using Twitter data from 2010. They trained four smaller models - Llama3-8B-Instruct, Qwen2.5-7B/0.5B-Instruct, and Qwen3-4B-Instruct - on different mixes of "junk" and higher-quality control data.

Diagram for the LLM Brain Rot study: hypothesis, Twitter data, pre-training, cognitive decline, error types, and countermeasures.
The diagram shows how targeted pre-training with junk data from X (formerly Twitter) leads to cognitive decline in large language models. | Image: Xing et al.

Two takes on what counts as "junk" data

The researchers took two approaches to identifying junk data. The first, based on engagement (M1), flagged short posts under 30 words that were highly popular (over 500 likes, retweets, or comments) as junk. Longer posts above 100 words with little engagement served as controls.

Ad
Ad

The second method (M2) measured content quality. Using GPT-4o-mini, the team sorted posts by their semantic value. Conspiracy theories, exaggerated claims, and attention-seeking clickbait were marked as junk, while more thoughtful material became controls.

The analysis showed little overlap between popularity and text length, and only a weak link between popularity and content quality. Meanwhile, text length and semantic value were more closely correlated.

Reasoning skills take a nosedive

Model performance suffered dramatic losses. On the ARC challenge benchmark, reasoning accuracy fell from 74.9 percent to 57.2 percent as junk data increased from zero to 100 percent.

Heatmap of Llama3 8B-Instruct performance at different junk data proportions in ARC, RULER, safety, and personality tests.
Llama3 8B Instruct's performance drops in reasoning, long-context understanding, safety, and personality benchmarks as junk data increases. | Image: Xing et al.

For tasks requiring long-context understanding, model accuracy dropped even more precipitously, plunging from 84.4 percent down to just 52.3 percent. This shows that as the proportion of low-quality data increases, model performance continues to worsen.

The engagement-based definition of junk (popularity) caused more damage than the content-based approach, suggesting that popularity adds a new dimension of data quality not captured by standard semantic checks.

Recommendation

The effects extended beyond reasoning. Models exposed to large amounts of engagement-driven junk developed "dark" personality traits, including higher scores for psychopathy, narcissism, and manipulativeness. In Llama3 8B Instruct, the psychopathy score rose sharply.

Safety benchmarks also declined. In contrast, exposure to content-based junk sometimes raised agreeableness and openness scores.

"Thought-skipping" dominates errors

Error analysis found that "thought-skipping"—skipping logical steps or chains entirely—was the most common problem. Over 70 percent of errors involved no reasoning at all, jumping to 84 percent in the engagement-junk scenario. Researchers sorted errors into five categories: no reasoning, no planning, skipped steps, wrong logic, and factual errors. Their system could automatically explain more than 98 percent of the cases.

Infographic: Green flowchart for correct soap-bacteria experiment and red fields for five types of reasoning errors.
Models trained on junk data are often unable to complete logical reasoning chains, leading to skipped steps and basic mistakes. | Image: Xing et al.

Follow-up tests found that popularity mainly weakened reasoning, while text length had a bigger effect on long-context understanding. This supports the idea that popularity influences LLMs in unique ways.

Ad
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.

Damage is hard to reverse

Efforts to repair the models had limited success. Reflective reasoning—where the model reviews its own output—reduced some thought-skipping, but self-reflection often made things worse. Only corrections from a stronger external model helped at all.

Even after retraining with up to 50,000 fresh examples and more clean data, the lost performance did not return. The gap remained.

"The gap implies that the Brain Rot effect has been deeply internalized, and the existing instruction tuning cannot fix the issue," the authors write.

The study calls for a rethink on how LLMs gather and filter online data. With models constantly absorbing huge volumes of web content, careful data selection and quality control are now critical to avoid permanent degradation.

The team recommends regular "cognitive health checks" for deployed LLMs and argues that data selection during ongoing training should be treated as a safety issue.

Code, models, and data are available on GitHub and Hugging Face.

Ad
Ad
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • Researchers have found that large language models permanently lose performance when they are continually trained on trivial online content, especially short and popular social media posts. This type of "junk data" has the strongest negative impact on reasoning and safety.
  • The study shows a sharp drop in performance: reasoning ability fell from 74.9 to 57.2 percent, and reading comprehension decreased from 84.4 to 52.3 percent. The models also developed negative personality traits, including increased manipulativeness and signs of psychopathy, after being trained on junk data.
  • Attempts to repair the damage by retraining with high-quality data were only partly successful. The researchers recommend stricter quality control for training data and regular "cognitive health checks" for language models.
Sources
Jonathan writes for THE DECODER about how AI tools can improve both work and creative projects.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.