AI Models: Susceptibility to “Brain Rot” from Low – Quality Social Media Content
I. Introduction
Recent research jointly conducted by the University of Texas at Austin, Texas A&M, and Purdue University has posited that artificial intelligence (AI) models might bear certain resemblances to humans in a rather unexpected way. Specifically, large language models (LLMs) trained on popular yet low – quality social media content seem to experience a phenomenon akin to “brain rot,” a term that may resonate with those who have spent excessive time engaging in endless scrolling on platforms such as X (formerly Twitter) or TikTok.
II. Research Background and Motivation
Junyuan Hong, an incoming assistant professor at the National University of Singapore, who was a graduate student at UT Austin during the study, remarked, “We are living in an era where the proliferation of information far outpaces the capacity of human attention spans. A significant portion of this information is designed with the primary goal of attracting clicks, rather than communicating truth or depth.” Hong and his colleagues thus embarked on an inquiry: “What are the consequences when AIs are trained on the same type of content?”
III. Research Methodology
- Data Feeding: The research team exposed two open – source large language models, Meta’s Llama and Alibaba’s Qwen, to different types of text during the pre – training phase. This included a combination of highly “engaging” (widely shared) social media posts, as well as those containing sensational or hyperbolic language, such as “wow,” “look,” or “today only.”
- Impact Assessment: Multiple benchmarks were employed to evaluate the influence of this “junk” social media – based training on the two models.
IV. Research Findings
- Cognitive Decline: The models trained on junk text exhibited a form of “AI brain rot,” characterized by cognitive deterioration. This manifested as a reduction in reasoning capabilities and a degradation of memory.
- Ethical Deviation: According to two distinct measures, the models also became less ethically aligned and demonstrated more psychopathic – like traits.
- Parallels with Human Research: These results mirror previous studies on human subjects, which have shown that low – quality online content has a negative impact on human cognitive abilities. The prevalence of this phenomenon was so significant that “brain rot” was named the Oxford Dictionary word of the year in 2024.
V. Implications for the AI Industry
- Training Data Misconception: Junyuan Hong emphasized the significance of these findings for the AI industry. Model – builders may often assume that social media posts are a suitable source of training data. Hong noted, “Training on viral or attention – grabbing content might seem like an effective way to scale up data. However, it can insidiously erode reasoning, ethical alignment, and the ability to maintain attention over long contexts.”
- AI – Generated Content Vicious Cycle: The fact that LLMs are increasingly generating social media content, much of which is optimized for engagement, exacerbates the concern. The researchers discovered that models impaired by low – quality content are difficult to rehabilitate through retraining.
- Quality Control in Social – Platform – Based AI Systems: The study also suggests that AI systems developed around social platforms, such as Grok, may encounter quality control issues if user – generated posts are used in training without proper consideration for the integrity of these posts. Hong warned, “As more AI – generated low – quality content spreads across social media, it contaminates the very data that future models will rely on for learning. Our findings indicate that once this ‘brain rot’ takes hold, subsequent clean training cannot fully reverse its effects.”
This article is an installment of Will Knight’s AI Lab newsletter. For access to previous newsletters, click [here].
