Zubnet AI學習Wiki › Continual 學習ing
Training

Continual 學習ing

Lifelong 學習ing, Incremental 學習ing
模型持續從新資料學習而不忘記之前學到的東西的能力。當前 LLM 一次訓練然後凍結 — 更新它們需要昂貴的重訓。持續學習讓模型能從每次互動學習、與新資訊保持同步、隨時間適應個人使用者,就像人類自然學習的方式。

為什麼重要

持續學習是 AI 最大的未解決問題之一。當前模型有知識截止、不能從糾正中學習、把每次對話當作白板。解決持續學習將消除對昂貴重訓週期的需要、啟用真正適應每個使用者的個性化 AI、讓模型永久保持最新。

Deep Dive

The fundamental obstacle is catastrophic forgetting: training on new data overwrites the weights that encoded old knowledge. Humans don't have this problem (learning French doesn't make you forget English) because biological neural networks use different mechanisms for memory consolidation. Artificial neural networks store all knowledge in shared weights, making any update a potential disruption to existing capabilities.

Current Approaches

Several strategies partially address continual learning: replay methods (mix old data with new data during training), regularization methods (penalize changes to weights important for old tasks, like EWC), architecture methods (allocate different parameters to different tasks), and retrieval-augmented approaches (store knowledge externally in a database rather than in weights). None fully solve the problem, but each enables some degree of incremental learning.

RAG as a Pragmatic Substitute

In practice, RAG (Retrieval-Augmented Generation) serves as a pragmatic substitute for true continual learning. Instead of updating the model's weights with new information, you update an external knowledge base that the model consults at inference time. This sidesteps catastrophic forgetting entirely but has limitations: the model's reasoning capabilities don't improve, only its access to information. True continual learning would improve both knowledge and capabilities simultaneously.

相關概念

← 所有術語
← Context Window Contrastive 學習ing →