64% of New Internet Content Is AI-Generated. We Said 50% Was Coming — We Were Too Conservative.
TexTak's [ai-generated-media] forecast sits at 68% probability that AI-generated content will exceed 50% of new internet media. Today's MIT CSAIL/Oxford Internet Institute study doesn't just support that thesis — it says we've already crossed the threshold, estimating 64% of all newly published internet material in 2026 is AI-generated. We're moving this forecast toward resolution, but we're not calling it done yet, and the reasons why tell you something important about how we're thinking about what this number actually measures.
Let's be direct about what the MIT/Oxford figure does and doesn't prove. The study claims 8.3 billion AI-written articles and 1.2 trillion social media posts were added to the web in 2025, with AI content outpacing human content 17:1. If accurate, this is direct evidence that our forecast threshold has been crossed — not 'conditions are forming,' not 'the trend supports,' but actual measured crossing. We weight this heavily because the research involves two credible institutions producing a quantified estimate, not a vendor survey or industry extrapolation. That said, the quality of any content-volume estimate at this scale depends entirely on detection methodology, and 'AI-generated' definitions vary widely across studies. A piece written by a human and edited by AI may or may not count. SEO farms producing technically human-initiated but AI-executed content sit in a gray zone. We're treating this as strong directional evidence — call it proximate-to-direct — rather than definitive resolution pending peer-reviewed publication and methodology disclosure.
What drives our 68% probability? Three things primarily. First, generation costs: text and basic image production have effectively approached zero marginal cost for anyone with API access, which means the economic incentive structure entirely favors volume. Second, the SEO spam dynamic: search-driven content farms have an asymmetric incentive — publish at machine scale or lose ranking ground to competitors doing the same. Third, platform dynamics: even platforms implementing content policies face a detection lag that lets synthetic content accumulate. The MIT study's 17:1 ratio, if directionally accurate even at half that magnitude, is structurally consistent with all three of these.
The strongest counterargument is the one we take seriously: consumer preference is collapsing. Our data shows preference for AI content at 26% today versus 60% three years ago. Detection accuracy reaching 88% among consumers is real. Platforms are implementing policies with actual enforcement teeth. The counter-thesis is that even if AI content floods the web, a quality/credibility bifurcation emerges — AI content dominates by volume but human content retains disproportionate reach and value. This is the scenario where our forecast resolves YES on a technicality while being somewhat meaningless as a leading indicator of anything important. We acknowledge this openly: volume dominance and influence dominance are different variables, and we're forecasting the former.
What would move us? If the MIT/Oxford paper clears peer review with methodology intact, we'd consider the 50% threshold resolved and shift the forecast question to something more analytically useful — perhaps whether AI content dominates across all major content categories including video, where the gap remains larger. If the methodology turns out to rely on weak detection proxies, we'd hold at 68% pending stronger evidence. What would drop us below 50%: a credible counter-study from a neutral institution showing significantly lower AI content share, or evidence that major platforms have successfully suppressed synthetic content volumes at scale. Neither is on the near-term horizon.