Zubnet AI学习Wiki › Model Collapse
Training

Model Collapse

Data Feedback Loop
当 AI 模型在先前 AI 模型生成的数据上训练时发生的退化,创造一个反馈循环,错误和偏见跨世代积累。每一代失去一些多样性,放大前一代的一些伪影,最终产出生成重复、泛泛、扭曲输出的模型。

为什么重要

模型崩溃是 AI 生成内容时代的定时炸弹。当互联网充满 AI 生成的文本(估计占新网页内容 10–50%),未来在网页抓取上训练的模型不可避免地会摄入 AI 输出。如果不小心管理,模型质量可能停滞或退化。这就是为什么数据策展和来源跟踪正在成为关键基础设施。

Deep Dive

The mechanism: a model trained on real data captures the distribution imperfectly — it overestimates some patterns and misses others. When a second model trains on the first model's outputs, it captures the first model's imperfect distribution, amplifying the errors. By generation 5 or 10, the distribution has collapsed to a narrow, distorted version of the original. Shumailov et al. (2023) demonstrated this empirically across multiple model types.

The Internet Contamination Problem

The practical concern: pre-training datasets are typically scraped from the web, and the web increasingly contains AI-generated content. If 20% of a training corpus is AI-generated, and that AI content has the same statistical biases as the model being trained, those biases get reinforced. The result isn't catastrophic failure but gradual homogenization — models that sound more and more like each other and less like the diversity of human expression.

Mitigations

Solutions include: detecting and filtering AI-generated content from training data (hard at scale), mixing AI-generated data with verified human data (maintaining a "human data floor"), watermarking AI outputs to enable filtering, and maintaining curated, AI-free reference datasets. Some researchers argue that model collapse is overstated if data is properly diversified and quality-controlled, but the risk is taken seriously enough that major labs invest in data provenance.

相关概念

← 所有术语
← Model Card Model Merging →