← Back to incidents
Nature Study Demonstrates AI Model Collapse from Training on Generated Data
HighCambridge researchers demonstrated that AI models trained on AI-generated content suffer irreversible quality degradation across generations, threatening the sustainability of future AI development as synthetic content proliferates online.
Category
Other
Industry
Technology
Status
Reported
Date Occurred
Jul 24, 2023
Date Reported
Jul 24, 2023
Jurisdiction
International
AI Provider
Other/Unknown
Application Type
other
Harm Type
operational
Human Review in Place
Unknown
Litigation Filed
No
model_collapsetraining_datadata_qualitysynthetic_contentAI_researchgenerative_modelsdata_contamination
Full Description
In July 2023, researchers from Cambridge University and other institutions published a landmark study in Nature demonstrating a phenomenon they termed 'model collapse' - the progressive deterioration of generative AI models when trained on data that includes outputs from previous AI models. Led by Ilia Shumailov, the research team conducted systematic experiments showing that when AI models are trained on datasets contaminated with AI-generated content, they experience irreversible degradation in output quality and diversity across successive training generations.
The researchers tested their hypothesis using multiple model types including Gaussian mixture models, variational autoencoders, and large language models. In each case, they observed that models trained on a mixture of original human data and AI-generated content showed progressive deterioration. The phenomenon was particularly pronounced when the proportion of AI-generated content in training data increased, with models losing the ability to capture the full diversity of the original data distribution. The team coined the term 'Habsburg AI' to describe this degenerative process, drawing an analogy to the genetic problems that arose from inbreeding in the Habsburg royal family.
The study's implications are far-reaching given the rapid proliferation of AI-generated content across the internet. As large language models and image generators produce increasingly sophisticated outputs that are being published online, future AI training datasets inevitably include this synthetic content. The researchers demonstrated that even small amounts of AI-generated data in training sets can trigger model collapse, with effects becoming more pronounced in subsequent generations. This creates a potential feedback loop where each generation of AI models becomes progressively worse, ultimately leading to models that produce increasingly homogenized and lower-quality outputs.
The research highlighted several concerning scenarios for the AI industry. As human-generated content becomes proportionally smaller compared to AI-generated content online, maintaining access to high-quality, diverse training data becomes increasingly challenging. The study showed that model collapse is not merely a temporary setback but an irreversible process - once a model has collapsed, it cannot recover its original performance even with additional training on high-quality data. This suggests that the current approach to AI training may be fundamentally unsustainable if synthetic content continues to proliferate without proper identification and filtering.
The findings have significant implications for AI companies and researchers who rely on web-scraped data for training. The study suggests that without careful curation of training datasets and robust methods for identifying AI-generated content, the quality of future AI models may decline substantially. The researchers emphasized the critical importance of preserving access to human-generated data and developing better methods for detecting and filtering synthetic content from training datasets to prevent widespread model degradation across the AI industry.
Root Cause
When generative AI models are trained on data that includes outputs from previous AI models, they experience 'model collapse' - a progressive loss of diversity and quality in outputs that compounds across training generations, similar to genetic drift in small populations.
Mitigation Analysis
Data provenance tracking to identify AI-generated content, maintaining high-quality human-generated training datasets, implementing detection systems to filter synthetic content, and establishing industry standards for training data authenticity could prevent model collapse. Regular model auditing and diversity metrics monitoring would help detect early signs of degradation.
Lessons Learned
The study reveals a fundamental challenge for the AI industry: the proliferation of AI-generated content online creates a potential feedback loop that could degrade future model performance. This highlights the critical importance of data provenance and the need for industry-wide standards to preserve training data quality.
Sources
The curse of recursion: training on generated data makes models forget
Nature · Jul 24, 2023 · academic paper
AI model collapse: training on generated data makes models forget
Cambridge University · Jul 24, 2023 · academic paper
AI models collapse when trained on recursively generated data
MIT Technology Review · Jul 24, 2023 · news