Zubnet AI学习Wiki › Continual 学习ing
Training

Continual 学习ing

Lifelong 学习ing, Incremental 学习ing
模型持续从新数据学习而不忘记之前学到的东西的能力。当前 LLM 一次训练然后冻结 — 更新它们需要昂贵的重训。持续学习让模型能从每次交互学习、与新信息保持同步、随时间适应个人用户,就像人类自然学习的方式。

为什么重要

持续学习是 AI 最大的未解决问题之一。当前模型有知识截止、不能从纠正中学习、把每次对话当作白板。解决持续学习将消除对昂贵重训周期的需要、启用真正适应每个用户的个性化 AI、让模型永久保持最新。

Deep Dive

The fundamental obstacle is catastrophic forgetting: training on new data overwrites the weights that encoded old knowledge. Humans don't have this problem (learning French doesn't make you forget English) because biological neural networks use different mechanisms for memory consolidation. Artificial neural networks store all knowledge in shared weights, making any update a potential disruption to existing capabilities.

Current Approaches

Several strategies partially address continual learning: replay methods (mix old data with new data during training), regularization methods (penalize changes to weights important for old tasks, like EWC), architecture methods (allocate different parameters to different tasks), and retrieval-augmented approaches (store knowledge externally in a database rather than in weights). None fully solve the problem, but each enables some degree of incremental learning.

RAG as a Pragmatic Substitute

In practice, RAG (Retrieval-Augmented Generation) serves as a pragmatic substitute for true continual learning. Instead of updating the model's weights with new information, you update an external knowledge base that the model consults at inference time. This sidesteps catastrophic forgetting entirely but has limitations: the model's reasoning capabilities don't improve, only its access to information. True continual learning would improve both knowledge and capabilities simultaneously.

相关概念

← 所有术语
← Context Window Contrastive 学习ing →