AI slop
Google suggested searchers put glue on their pizza. It was a peak moment for AI slop.
From glue-topped pizza to 1,800-article heists, we catalog the rise of AI slop in search and why the internet's automated sanity check is currently failing.
In May 2024, the world’s most sophisticated information retrieval system confidently informed a user that the best way to keep cheese from sliding off a pizza was to mix approximately 1/8 cup of non-toxic glue into the sauce. For a brief, shimmering moment, the primary interface between humanity and the sum of its knowledge became a parody of a 1950s cookbook written by a nihilist. This was not an isolated incident of digital indigestion; it was the public debut of "content slop"—unhelpful, low-quality, AI-generated content published without human editorial review that degrades the information quality of a platform.
The integration of Large Language Models (LLMs) into search workflows has prioritized the volume of automated answers over factual verification, leading to a measurable degradation in web reliability that search engines are currently unable to fix without massive, manual de-indexing of automated content. As tech giants raced to compete with the short-form dopamine of TikTok and the conversational speed of ChatGPT, they inadvertently dismantled the human-centric "sanity check" that once governed the open web. We are currently living through the fallout of this transition, where the industrial-scale production of automated filler has turned the search bar into a probabilistic lottery.
1. The Hallucination Hall of Fame: From Glue to Glass
The viral "glue on pizza" incident was merely the flagship failure of Google's initial AI Overviews rollout. During the first week of deployment, the system also suggested that users should eat at least one small rock per day for minerals, a recommendation that appeared to be hallucinated from a satirical article published by The Onion. The machine, famously incapable of distinguishing between a culinary tip and a punchline, processed the web’s irony as literal instruction. According to The Verge, Google was forced to "take swift action" to manually remove these answers, a reactive posture that highlights the core flaw: the models cannot perform their own internal quality control.
This systemic rot did not begin with Google. In 2022, Meta released Galactica, an LLM trained on scientific papers that was pulled after just 72 hours. Galactica’s specific brand of slop was particularly dangerous; it produced authoritative-sounding papers suggesting that crushed glass could be a beneficial dietary supplement. As noted by Ars Technica, these models are logical voids. They are probabilistic engines that predict the next likely word based on frequency, not the next true fact based on reality.
The problem intensified when the technology moved from harmlessly weird to actively life-threatening. Reports from Forbes documented Google AI suggesting that searchers cook with poisonous mushrooms like Amanita phalloides. The model identified these mushrooms as edible because it scraped data from hobbyist forums where users were discussing toxins, yet it failed to distinguish between a warning and a recipe. This is the definition of AI slop: a high-confidence output generated by a machine that lacks the biological context to understand that certain words lead to a morgue.
| Feature | Traditional Snippets | AI Overviews (Slop Era) |
|---|---|---|
| Verification | Human editorial/peer review | Probabilistic word matching |
| Reliability | Variable, but source-attributed | High confidence, frequently false |
| Failure Mode | Irrelevant links | Dangerous medical/culinary advice |
| Cleanup | Algorithmic filtering | Manual de-indexing by developers |
2. The SEO Heist: How to Steal 3 Million Clicks with a Script
While Google was busy hallucinating pizza recipes, a new breed of growth hackers was perfecting Scaled Content Abuse—the practice of producing content at scale primarily to manipulate search rankings without providing original value. The most documented instance of this was the "SEO Heist" executed by Jake Ward. Ward used AI to generate 1,800 articles in a matter of hours, cloning the structure and keywords of a competitor, Exceljet, to divert traffic to his own site, Causal.
The heist was initially a success, diverting between 3.1 and 3.6 million in total traffic from a legitimate creator to an automated clone. It was a digital strip-mine operation. However, the victory was short-lived as Google’s March 2024 Core Update introduced three new spam policies specifically targeting Scaled Content Abuse. These policies were designed to combat the "Dead Internet" phenomenon where bot-generated content crowds out human-authored information.
The result was a total traffic collapse for the heist site. According to Search Engine Journal, the March update successfully reduced unhelpful, unoriginal content in search results by 45%. While this purge was significant, it revealed a grim reality: the barrier to entry for polluting the web is now so low that search engines must engage in a permanent, manual war of de-indexing. The industrial-scale production of AI slop has fundamentally shifted the cost of being wrong to the user.
3. The Institutionalization of Slop: CNET and Sports Illustrated
The degradation of the web is not just the work of rogue growth hackers; it has been embraced by legacy institutions looking to cut editorial costs. In early 2023, the tech publication CNET was caught using AI to write financial explainers that contained significant factual errors. As reported by The Verge, the publication was forced to issue a massive correction notice after its "AI Money" tool hallucinated basic interest rate calculations. This was a turning point for digital media, proving that even reputable brands were willing to trade their credibility for the efficiency of automated content.
The trend continued with Sports Illustrated, which allegedly used AI-generated authors with fake personas and headshots to populate its site with product reviews. Futurism broke the story, revealing that the "writers" did not exist and the content was often nonsensical. This specific form of slop—synthetic authority—is designed to trick both search engines and human readers. It bypasses the E-E-A-T (Experience, Expertise, Authoritativeness, and Trustworthiness) criteria by simulating the appearance of a human expert.
Even local news has not been spared. Gannett, the largest newspaper chain in the United States, was forced to pause its use of AI for high-school sports reporting after the bot produced hilariously bad articles. The BBC noted that the AI referred to "the close of the fourth period" in games that only have halves and described scores as "the points of the game." These failures demonstrate that while AI can mimic the structure of news, it lacks the contextual understanding to report on even the simplest human events.
4. Why the Machine Can't Stop Being Wrong
The persistence of these failures stems from a fundamental speed-accuracy tradeoff. Tech companies are currently incentivized to prioritize the volume of automated answers because the serving costs are plummeting. Sundar Pichai noted an 80% reduction in AI serving costs during testing, a financial incentive that makes "mostly-accurate-and-very-cheap" more attractive to shareholders than "perfectly-accurate-and-expensive."
This incentive structure directly conflicts with the foundational principles of search reliability. AI, by definition, has no lived experience. It can simulate the tone of expertise, but it cannot perform a sanity check because it does not know what a "sanity check" is. According to The Guardian, we are approaching a feedback loop where AI models are trained on the slop produced by other AI models, leading to a phenomenon known as "model collapse."
Gary Marcus’s critique remains relevant: LLMs are logical voids. When a human writes a recipe, they know that glue is not food. When an LLM "writes" a recipe, it simply knows that the word "glue" has appeared in proximity to "pizza" on some forgotten corner of the internet. The machine is not lying; it is simply predicting the next most statistically likely token. As long as the internet contains a mixture of fact, fiction, and satire, a probabilistic engine will eventually serve a mixture of all three as truth.
5. Counter-Argument: The Case for Efficiency and Buried Data
It is important to represent the opposing view fairly. Defenders of AI search argue that these systems are significantly better at answering what are known as "Buried Information Queries." These are niche, technical questions—such as obscure software shortcuts or Mac terminal commands—where the useful information is normally buried under pages of human-written SEO spam.
As noted in a review by The Verge, AI is remarkably efficient at retrieving syntax or specific settings that would otherwise take ten minutes of scrolling to find. In these low-stakes environments, the efficiency of the LLM genuinely improves the user experience. Proponents argue that the "glue on pizza" errors are edge cases that will be ironed out as Retrieval Augmented Generation (RAG) improves. They contend that the alternative—sorting through pages of "recipe-style" preamble written by humans to gaming the system—is its own form of manual slop.
However, this efficiency does not extend to high-stakes domains where the models lack the necessary safeguards. The efficiency gained in software troubleshooting does not offset the risk of a search engine telling a parent to give their child crushed glass for a cough. Recent investigations into Perplexity AI by TechCrunch further complicate the efficiency argument, suggesting that these "efficient" answers often involve scraping and summarizing original reporting without compensation or proper attribution. The efficiency of AI search is often built on the cannibalization of the human sources it claims to replace.
6. Search After the Gold Rush
The era of "set it and forget it" search is over. The evidence from the 2024 core updates supports the thesis that search engines must now wage a permanent, manual war against the very automation they helped unleash. Verification is becoming the new luxury good of the open web. We are seeing a return to verified expertise as a defense mechanism against the automated filler that has turned the open web into a digital landfill.
Niche publishers who can prove they have real-world experience are the only ones surviving the current transition. Meanwhile, the platforms remain in a reactive state. The 45% reduction in unhelpful content achieved by the March update proves that the problem is manageable, but it requires constant, resource-heavy human oversight. This contradicts the initial promise of AI: that it would automate the difficult tasks of curation and organization.
The conclusion is inescapable: the integration of LLMs has fundamentally broken the internet's automated sanity check. While AI-generated content can solve minor technical queries, its industrial-scale use for SEO has forced us into a world where every search result must be treated as "alleged" until proven otherwise. The glue on the pizza wasn't just a hallucination; it was a warning that the web’s reliability is no longer a given. It is something that must be manually defended, one de-indexed link at a time. The machine can predict the next word, but it cannot predict the consequences of its own inaccuracy.