Zubnet AI学习Wiki › Curriculum 学习ing
Training

Curriculum 学习ing

按有意义的顺序呈现样本的训练策略 — 通常从易到难 — 而不是随机。像在微积分之前教学生算术,课程学习先给模型基础模式,逐渐建立复杂性。这能带来更快收敛,有时最终性能也更好。

为什么重要

课程学习是一个被低估的技术,能在不改变模型或数据的情况下改善训练效率。LLM 预训练越来越多用数据调度 — 在最后训练阶段展示更干净、更高质量的数据 — 这是课程学习的一种形式。你呈现数据的顺序很重要,不只是数据本身。

Deep Dive

The intuition: random training order means the model encounters hard examples before it has the foundation to learn from them, wasting gradient updates. By starting with easy, clear examples, the model builds basic representations that make harder examples learnable. Bengio et al. (2009) formalized this, showing that curriculum training converges faster and sometimes reaches better solutions than random ordering.

In LLM Pre-training

Modern LLM training uses a form of curriculum learning through data mixing schedules. Early training might use a broad mix of web text. Later stages increase the proportion of high-quality data (curated text, code, math, reasoning). Some training recipes "anneal" on premium data in the final phase — the model learns broad patterns first, then refines on carefully selected examples. This data scheduling is one of the most closely guarded secrets of frontier labs.

Self-Paced 学习ing

Self-paced learning is a variant where the model itself decides what's easy or hard, based on its current loss on each example. Examples with low loss (already learned) are deprioritized, while examples with moderate loss (challenging but learnable) are emphasized. Examples with very high loss (noise, mislabeled data) are also deprioritized as potential outliers. This adaptive curriculum adjusts automatically as the model improves.

相关概念

← 所有术语
← Cross-Attention Cursor →