Zubnet AI學習Wiki › Supervised 學習ing
Training

Supervised 學習ing

一種訓練方法,模型從有標籤的樣本中學習 — 輸入-輸出配對,其中正確答案已經提供。「這是一張貓的圖,標籤是『貓』。這是一張狗的圖,標籤是『狗』。」模型調整參數,使其預測與已知正確答案之間的差異最小化。

為什麼重要

監督式學習是機器學習最直觀的形式,仍然是大多數實際應用背後的主力:垃圾郵件過濾器、醫學影像分析、詐欺偵測,還有 LLM 的 fine-tuning 階段。當你有有標籤的資料和明確的目標時,監督式學習通常是你的起點。

Deep Dive

The core loop of supervised learning is: make a prediction, compare it to the label, compute a loss (how wrong you were), and adjust parameters to reduce that loss. This cycle repeats millions or billions of times during training. The math behind the adjustment is gradient descent — computing how much each parameter contributed to the error and nudging it in the direction that reduces the error.

It's Everywhere in LLMs

Pre-training an LLM is technically a form of self-supervised learning (the labels are generated from the data itself — the "label" for each position is just the next token in the text). But fine-tuning and RLHF both use supervised signals: human-written example responses, or human preference rankings between model outputs. When you fine-tune a model on customer support conversations, you're doing supervised learning with the support agent's responses as labels.

The Data Bottleneck

The catch with supervised learning is that you need labeled data, and labels are expensive. Every medical image needs a radiologist to annotate it. Every support conversation needs a quality rating. This is why techniques like self-supervised learning (letting the model generate its own labels from unlabeled data) and semi-supervised learning (using a small labeled set to bootstrap labels for a larger unlabeled set) are so important — they reduce the labeling bottleneck that limits pure supervised approaches.

相關概念

← 所有術語
← Superposition SwiGLU →