Zubnet AI學習Wiki › Curriculum 學習ing
Training

Curriculum 學習ing

按有意義的順序呈現範例的訓練策略 — 通常從易到難 — 而不是隨機。像在微積分之前教學生算術,課程學習先給模型基礎模式,逐漸建立複雜性。這能帶來更快收斂,有時最終性能也更好。

為什麼重要

課程學習是一個被低估的技術,能在不改變模型或資料的情況下改善訓練效率。LLM 預訓練越來越多用資料調度 — 在最後訓練階段展示更乾淨、更高品質的資料 — 這是課程學習的一種形式。你呈現資料的順序很重要,不只是資料本身。

Deep Dive

The intuition: random training order means the model encounters hard examples before it has the foundation to learn from them, wasting gradient updates. By starting with easy, clear examples, the model builds basic representations that make harder examples learnable. Bengio et al. (2009) formalized this, showing that curriculum training converges faster and sometimes reaches better solutions than random ordering.

In LLM Pre-training

Modern LLM training uses a form of curriculum learning through data mixing schedules. Early training might use a broad mix of web text. Later stages increase the proportion of high-quality data (curated text, code, math, reasoning). Some training recipes "anneal" on premium data in the final phase — the model learns broad patterns first, then refines on carefully selected examples. This data scheduling is one of the most closely guarded secrets of frontier labs.

Self-Paced 學習ing

Self-paced learning is a variant where the model itself decides what's easy or hard, based on its current loss on each example. Examples with low loss (already learned) are deprioritized, while examples with moderate loss (challenging but learnable) are emphasized. Examples with very high loss (noise, mislabeled data) are also deprioritized as potential outliers. This adaptive curriculum adjusts automatically as the model improves.

相關概念

← 所有術語
← Cross-Attention Cursor →