Zubnet AI學習Wiki › Model Collapse
Training

Model Collapse

Data Feedback Loop
當 AI 模型在先前 AI 模型生成的資料上訓練時發生的退化,創造一個反饋迴圈,錯誤和偏見跨世代累積。每一代失去一些多樣性,放大前一代的一些偽影,最終產出生成重複、泛泛、扭曲輸出的模型。

為什麼重要

模型崩潰是 AI 生成內容時代的定時炸彈。當網路充滿 AI 生成的文字(估計佔新網頁內容 10–50%),未來在網頁抓取上訓練的模型不可避免地會攝入 AI 輸出。如果不小心管理,模型品質可能停滯或退化。這就是為什麼資料策展和來源追蹤正在成為關鍵基礎設施。

Deep Dive

The mechanism: a model trained on real data captures the distribution imperfectly — it overestimates some patterns and misses others. When a second model trains on the first model's outputs, it captures the first model's imperfect distribution, amplifying the errors. By generation 5 or 10, the distribution has collapsed to a narrow, distorted version of the original. Shumailov et al. (2023) demonstrated this empirically across multiple model types.

The Internet Contamination Problem

The practical concern: pre-training datasets are typically scraped from the web, and the web increasingly contains AI-generated content. If 20% of a training corpus is AI-generated, and that AI content has the same statistical biases as the model being trained, those biases get reinforced. The result isn't catastrophic failure but gradual homogenization — models that sound more and more like each other and less like the diversity of human expression.

Mitigations

Solutions include: detecting and filtering AI-generated content from training data (hard at scale), mixing AI-generated data with verified human data (maintaining a "human data floor"), watermarking AI outputs to enable filtering, and maintaining curated, AI-free reference datasets. Some researchers argue that model collapse is overstated if data is properly diversified and quality-controlled, but the risk is taken seriously enough that major labs invest in data provenance.

相關概念

← 所有術語
← Model Card Model Merging →