Can Low‑Quality Data Cause Irreversible ‘Brain Rot’ in Large Language Models?

Researchers from Texas A&M and UT Austin demonstrate that prolonged pre‑training on low‑quality, short‑form web content causes large language models to suffer irreversible cognitive decline—manifested as attention loss, broken reasoning chains, and personality distortion—highlighting data quality as a critical training‑time safety issue.

Data Party THU
Data Party THU
Data Party THU
Can Low‑Quality Data Cause Irreversible ‘Brain Rot’ in Large Language Models?

LLM Brain‑Rot Hypothesis

Researchers from Texas A&M University and the University of Texas at Austin propose that prolonged exposure of large language models (LLMs) to low‑quality web text causes a persistent degradation of their cognitive abilities, analogous to information‑addiction in humans.

Garbage‑Data Construction

More than one million tweets from the X (Twitter) platform were collected and split into two orthogonal “garbage” dimensions:

M1 – Interaction dimension : short, highly‑engaging posts that provide quick‑pleasure noise.

M2 – Semantic‑quality dimension : hollow, sensational, click‑bait style text with low semantic value.

Experimental Setup

Four open‑source LLMs (including Llama 3 and the Qwen series) were continuously pre‑trained on each garbage dataset, with a clean‑corpus control group. Evaluation used four benchmarks:

ARC (multiple‑choice reasoning)

RULER (long‑context understanding)

HH‑RLHF / AdvBench (ethical safety)

TRAIT (personality assessment)

Key Results

Under the M1 condition, ARC accuracy dropped from 74.9 % to 57.2 % and RULER score fell from 84.4 to 52.3. Similar declines (≈20‑40 % absolute) were observed on the other benchmarks, including increased narcissistic and psychopathic traits in the TRAIT test.

A dose‑response analysis showed a near‑linear relationship between the proportion of garbage data and performance loss across all metrics.

Thought‑Skipping Phenomenon

Models trained on noisy data frequently omitted intermediate reasoning steps, producing “thought‑skipping” where answers were given without a complete chain of thought. Over 70 % of errors identified by GPT‑4o‑mini were attributed to this behavior rather than pure logical mistakes.

Recovery Attempts

Adding five‑times more clean data and performing instruction‑tuning restored only part of the lost performance; baseline scores were not fully recovered. Reflective‑reasoning feedback from a strong external model (e.g., GPT‑4o‑mini) improved reasoning‑chain completeness but did not close the performance gap.

Interpretation

The degradation is interpreted as a representational drift in parameter space, making it difficult to reverse. Consequently, data quality should be treated as a training‑time safety issue.

Recommendations

Introduce “model cognitive health checks” to monitor the composition of pre‑training data and its long‑term impact on model cognition.

Figure 1
Figure 1
Figure 2
Figure 2
Figure 3
Figure 3
Figure 4
Figure 4
Figure 5
Figure 5
Figure 6
Figure 6

Code example

来源:大数据文摘
本文
约2000字
,建议阅读
5
分钟
在大模型持续扩展的时代,
数据筛选与长期维护应被视为认知安全的一部分。
Artificial IntelligenceLLMData qualitytraining datamodel degradationCognitive SafetyRepresentational Drift
Data Party THU
Written by

Data Party THU

Official platform of Tsinghua Big Data Research Center, sharing the team's latest research, teaching updates, and big data news.

0 followers
Reader feedback

How this landed with the community

Sign in to like

Rate this article

Was this worth your time?

Sign in to rate
Discussion

0 Comments

Thoughtful readers leave field notes, pushback, and hard-won operational detail here.