1aiautomationclickbaitengagement baitFeaturedjournalismllmsstudy

Study: AI Models Trained On Clickbait Slop Result In AI ‘Brain Rot,’ ‘Hostility’

from the you-are-what-you-eat dept

While “AI” certainly has some useful applications, a lot of the folks in charge of the trajectory of LLMs clearly want to use it to build a giant, badly automated ouroboros of lazy internet slop that shits out ad money without the need for pesky labor. You see this most profoundly in media, where a bunch of far-too-clever lads rushed to integrate under-cooked, broadly misunderstood LLMs with disastrous results.

These folks could be using AI to make work more efficient; instead they’re using it to cut corners, undermine labor, and fill the internet with a parade of mindless, pointless, low-quality clickbait slop. The sort of lazy engagement bait that hoovers ad money and attention away from folks who actually have something useful to say or contribute.

As it turns out, training LLMs on this kind of slop doesn’t work out well for anybody.

A new joint study by researchers at Texas A&M University, University of Texas at Austin, and Purdue University took a closer look at what happens when you train LLMs on the kind of engagement slop our modern internet gatekeepers are keen to create.

To see how these models would “behave” after subsisting on a diet of clickbait sewage, the researchers cobbled together a sample of one million X posts and then trained four different LLMs on varying mixtures of control data (long form, good faith, real articles and content) and junk data (lazy, engagement chasing, superficial clickbait) to see how it would affect performance.

Their conclusion isn’t too surprising; the more junk data that is fed into an AI model, the lower quality its outputs become and the more “hostile” and erratic the model is:

“All four models tested—Llama3 8B, Qwen2.5 7B/0.5B, Qwen3 4B—showed some forms of cognitive decline. Meta’s Llama proved the most sensitive to the junk, seeing drops in its reasoning capabilities, understanding of context, and adherence to safety standards. Interestingly, a much smaller model, Qwen 3 4B, proved more resilient, though still suffered declines. It also found that the higher the rates of bad data, the more likely a model was to slip into “no thinking” mode, failing to provide any reasoning for its answer, which was more likely to be inaccurate.”

You are what you eat.

They also found that after being fed a bunch of ex-Twitter slop, the models didn’t just get “dumber”, they were (shocking, I know) far more likely to take on many of the nastier “personality traits” that now dominate the right wing troll platform:

“More than just getting “dumber” in its thinking, though, the researchers found the inclusion of junk also resulted in an interesting effect: it led to changes in the model’s “personality,” succumbing to what the researchers called “dark traits.” For instance, the Llama 3 model displayed significantly higher levels of narcissism and became less agreeable. It also went from displaying nearly no signs of psychopathy to extremely high rates of the behavior.”

And by “dumber” and “narcissistic” they of course mean a vague simulacrum of those personality traits, because modern LLMs don’t understand anything, much less adopt real personalities. You’ll often see people (even prominent NYT tech journalists) attributing malicious intent and understanding to language learning models, inadvertently advertising the fact they don’t know how any of this works.

There’s been so much misrepresentation of what these models are capable of (by both companies and the tech media), this comment below needs to be projected onto the moon:

Chatbots — LLMs — do not know facts and are not designed to be able to accurately answer factual questions. They are designed to find and mimic patterns of words, probabilistically. When they’re “right” it’s because correct things are often written down, so those patterns are frequent. That’s all.

Katie Mack (@astrokatie.com) 2025-06-19T11:21:19.719Z

You see this a lot in breathless articles about LLMs that are trying to “resist being shut off” or somehow “blackmail their operators.” It’s simply not how this technology actually works. It’s part of a con suggesting these models are just a few weeks and another billion away from HAL 900 sentience.

Again, none of this is to say LLMs don’t have very useful applications, such as examining vast troves of scientific data to look for patterns and facts that humans might miss. Or creating more efficient, “intelligent” software that can be predictive of the user’s needs or inputs. Or automating basic customer service inquiries in a world full of already-low quality outsourced support.

The problem with AI generally is a decidedly human one: the terrible, unethical, and greedy people currently in charge of it’s implementation (again, see media, insurance, countless others) — folks who have cultivated some unrealistic delusions about AI competency and efficiency (see this recent Stanford study on how rushed AI adoption in the workforce often makes people less efficient).

This is before you even get to the climate and energy impact of these models, or the fact that the underlying financials are a hot mess poised to cause some serious economic tumult next year as the outer layer of hype and misrepresentation burns off. Even then, this quest to turn the internet into an ocean of lazy and uncurated ad engagement slop will remain a centerpiece of the movement.

Filed Under: , , , , , ,

Source link

Related Posts

1 of 69