The Dead Internet Theory — the idea that most online content and interaction is now generated and consumed by automated systems rather than human beings — originated as a fringe post on an obscure forum around 2021. It has since migrated into mainstream tech discourse, partly because it is unfalsifiable in its strongest form, and partly because the parts that are measurable are genuinely concerning.
Working in IT security, I spend a meaningful portion of my time dealing with the automated traffic side of this problem. Not the conspiracy version — the operational version. The version where you look at server logs and try to determine what percentage of your requests are coming from actual users versus scrapers, crawlers, credential stuffers, and content bots. That distinction matters enormously in security contexts, and the tools we use to make it are imperfect in ways that are worth understanding.
What the Theory Actually Claims
The Dead Internet Theory in its original form made a specific and extreme claim: that around 2016–2017, coordinated corporate and government actors deliberately flooded the internet with AI-generated content and bot activity to manipulate public opinion, suppress organic human expression, and create the illusion of consensus around manufactured narratives.
The strong version of this is essentially unfalsifiable and functions more as a psychological framework than a testable hypothesis. If you believe it fully, any evidence against it can be absorbed into the theory itself.
But the weak version — that a substantial and growing portion of internet traffic is non-human, that this proportion has increased significantly with the availability of generative AI, and that this has measurable effects on online culture and information quality — is not a conspiracy theory at all. It is a documented technical reality that security researchers, ad fraud analysts, and platform trust teams have been measuring and writing about for years.
What Bot Traffic Measurement Actually Shows
The most cited figures come from Imperva’s annual Bad Bot Report, which analyzes traffic across thousands of websites and attempts to classify requests as human, good bot (legitimate crawlers like Googlebot), or bad bot (scrapers, fraud bots, account takeover tools, content generators).
The 2024 report found that 49.6% of all internet traffic was non-human — crossing the 50% threshold for the first time in the report’s history. Of that, bad bots accounted for 32% of all traffic, while good bots accounted for 17.6%. Human traffic was 50.4% — barely a majority.
This number requires interpretation. “Internet traffic” in this context means HTTP requests to web properties, not engagement interactions. A bot that hits a page once counts the same as a human who reads it for ten minutes. The metric measures volume of requests, not depth of engagement or content creation.
But the trend direction is significant. The 2022 report showed bad bots at 27.7% of traffic. The 2023 report showed 30.2%. The 2024 figure of 32% represents a consistent upward trajectory that accelerated after the mass availability of LLM-based tools in 2023.
Where the Theory Gets More Interesting: Content Generation
The traffic measurement problem is relatively well-understood. The content generation problem is newer and harder to measure.
Before 2022, creating convincing automated content at scale required either significant resources (professional content farms) or accepting low quality (obvious spam). Generative AI changed this equation fundamentally. The cost of producing plausible-sounding text, realistic-looking images, and contextually appropriate social media comments dropped by several orders of magnitude.
The practical effect in security terms: the automated content that used to be trivially identifiable — grammatically broken, contextually inappropriate, obviously templated — is increasingly difficult to distinguish from human-generated content without forensic analysis. This is not a theory; it is a direct consequence of how these models work.
Platforms have acknowledged this. Meta reported removing 7.7 billion fake accounts in 2023 alone (their transparency report). To put that in context: Facebook has approximately 3 billion monthly active users. The number of fake accounts removed in a single year is more than twice the number of legitimate users. Most of these were caught early in the account lifecycle by automated detection, but the ratio illustrates the scale of the problem.
Twitter/X’s own filing during the Elon Musk acquisition dispute estimated that at least 5% of monetizable daily active users were bots — a figure Musk argued was much higher (potentially 20%+) and which became the subject of significant legal and public dispute. Neither side could definitively prove their number, which itself illustrates the measurement difficulty.
The Comment Section Problem
The aspect of the Dead Internet Theory that resonates most with ordinary users is the comment section experience — the sense that online discussions increasingly feel generic, staged, or disconnected from the actual content being commented on.
This has a partly measurable explanation. Coordinated inauthentic behavior campaigns — documented by platform trust teams, the Stanford Internet Observatory, the EU DisinfoLab, and academic researchers — routinely deploy networks of accounts to amplify specific narratives, suppress others, and manufacture the appearance of consensus. These are not hypothetical: they have been documented in detail across Facebook, Twitter, YouTube, Reddit, and TikTok, spanning political campaigns in dozens of countries.
The Stanford Internet Observatory’s 2023 report on coordinated inauthentic behavior documented over 150 distinct influence operations across major platforms in a single year. The operators ranged from state-sponsored campaigns to commercial operations selling engagement metrics to businesses and influencers.
The commercial engagement fraud dimension is underappreciated. There is a substantial industry — operating openly, as services that can be purchased — selling fake followers, fake views, fake comments, and fake reviews. The consumers include businesses trying to appear more popular than they are, political campaigns, influencers monetizing based on audience size, and anyone whose income or credibility depends on visible social proof. This is not a shadow conspiracy; it is a service industry with pricing tiers and customer support.
What This Means for Information Quality
The practical consequence that the Dead Internet Theory points toward, stripped of its conspiratorial framing, is a genuine epistemic problem: the signals we use to evaluate the credibility and popularity of information online — engagement metrics, comment volume, apparent consensus — have been compromised as reliable indicators.
When a video has 10 million views, that number does not tell you whether 10 million people found it meaningful or whether a fraction of that number is real and the rest was purchased or generated. When a product has 4.8 stars from 2,000 reviews, the review count and rating do not reliably indicate whether real customers found it good.
This is sometimes called “social proof manipulation” in security contexts, and it is treated as a fraud problem by platforms, advertisers, and e-commerce operators. The difference between the security framing and the Dead Internet framing is not really about the facts — it is about how much intentional coordination is attributed to the pattern versus how much is simply the emergent result of incentive structures that reward inflated metrics.
The Part That Is Actually New
What has changed since 2022 is not the existence of bot traffic or content farms — both of those predate the theory. What has changed is the quality ceiling.
For most of internet history, automated content was identifiable by its quality. Real humans write in ways that are contextually coherent, stylistically varied, and responsive to specific details. Automated content was generic, repetitive, and contextually disconnected. That quality gap was the primary detection mechanism for both human readers and automated classifiers.
Large language models substantially close that gap. A well-prompted LLM can produce comments, articles, reviews, and social media posts that pass casual human inspection. This does not mean most online content is now AI-generated — the majority of what you encounter is still written by people. But the detection heuristics that previously allowed readers to identify automated content are less reliable than they were three years ago.
In IT security terms, this is analogous to the shift from obviously malicious URLs to convincing phishing pages. The attack surface did not change; the attack quality did.
How to Think About This Without Becoming Paranoid
The version of the Dead Internet Theory that treats all online interaction as fake and all consensus as manufactured leads to an epistemic dead end. If nothing can be trusted, the theory itself cannot be trusted, and you are left with nothing to evaluate evidence against.
The more useful frame is probabilistic skepticism: treating engagement metrics and apparent consensus as weaker signals than they were five years ago, applying more scrutiny to information that confirms existing beliefs (which is also what influence operations target), and valuing primary sources and verifiable claims over social proof.
From a security perspective, the same hygiene principles that apply to phishing apply here. You do not need to assume every email is fake; you need to be appropriately skeptical of emails that ask you to do something consequential. The internet has not become entirely fake; the signals that tell you what is real have become less reliable in specific, identifiable ways.
Understanding which ways those are is more useful than believing the whole thing is dead.
Sources:
- Imperva Bad Bot Report 2024
- Meta Transparency Report 2023 (fake account removals)
- Stanford Internet Observatory, “Coordinated Inauthentic Behavior” annual report 2023
- EU DisinfoLab reports on coordinated influence operations
- Twitter/X vs. Elon Musk acquisition filings (bot percentage dispute, 2022)
- Cloudflare Radar traffic analysis data (2024)
This article reflects the author’s professional perspective from IT security work and does not constitute advice on any specific platform or security product.


