Zubnet AI学习Wiki › Catastrophic Forgetting
Training

Catastrophic Forgetting

Catastrophic Interference
当一个神经网络在新任务上训练后,失去执行先前学到任务的能力。在客服数据上 fine-tune 一个模型可能让它在支持上棒极了但编码上糟糕。新的学习覆盖了编码旧能力的权重,“忘记”了它们。

为什么重要

灾难性遗忘是 fine-tuning 和持续学习的核心挑战。这就是为什么你不能一个接一个 fine-tune 任务,指望它什么都做得好。这也是为什么 LoRA(只修改参数的一个小子集)和仔细的学习率选择对保留基础模型能力至关重要。

Deep Dive

The root cause is weight sharing: the same parameters encode multiple capabilities, and updating them for a new task disrupts the existing encodings. In a large neural network, knowledge isn't stored in dedicated neurons — it's distributed across the weights in complex, overlapping patterns (superposition). Modifying those weights for new knowledge inevitably disturbs old knowledge.

Mitigation Strategies

Several techniques reduce forgetting. Low learning rates during fine-tuning minimize weight changes. LoRA adds new trainable parameters while keeping the original weights frozen. Elastic Weight Consolidation (EWC) identifies which weights are important for old tasks and penalizes changes to them. Replay methods mix old task data into new task training. None fully solve the problem — there's always a trade-off between plasticity (learning new things) and stability (retaining old things).

The Continual 学习ing Dream

Continual learning (also called lifelong learning) is the research goal of building models that can keep learning from new data without forgetting old capabilities — the way humans do. Current LLMs sidestep this by training once on a massive dataset and then fine-tuning carefully. True continual learning remains an open problem and would be transformative: imagine a model that keeps learning from every conversation without degrading.

相关概念

← 所有术语
← Cartesia Cerebras →