Zubnet AI學習Wiki › Annotation
Training

Annotation

Data Labeling, Data Annotation
給原始資料添加標籤、標記、元資料讓它能用於監督式學習的過程。標註影像意味著在物體周圍畫邊界框。標註文字意味著標記實體、情感、意圖。為 RLHF 標註意味著按品質排名模型回應。標註是把原始資料變成訓練資料的人類勞動。

為什麼重要

標註是監督式 AI 不光鮮的基礎。每個有標籤資料集、每個 fine-tuned 模型、每個對齊的助手,都依賴花數小時正確標註資料的人工標註者。標註品質直接決定模型品質 — 不一致或有偏的標註產出不一致、有偏的模型。它是建構 AI 系統最勞動密集、往往最昂貴的部分。

Deep Dive

Annotation workflows typically involve: (1) creating clear labeling guidelines (what counts as "positive sentiment"? what's the boundary of a "car" in a bounding box?), (2) training annotators on the guidelines, (3) annotating data with multiple annotators per example (for quality control), (4) measuring inter-annotator agreement (do annotators agree on labels?), and (5) resolving disagreements (through adjudication or majority vote). Low agreement often indicates ambiguous guidelines or genuinely ambiguous data.

RLHF Annotation

For LLM alignment, annotation means comparing model responses: "Is response A or response B better for this prompt?" This preference annotation is particularly challenging because "better" is subjective, context-dependent, and culturally variable. Annotator demographics, expertise, and instructions all influence the resulting preference data, which in turn shapes model behavior. This is why alignment is often described as encoding the values of whoever writes the annotation guidelines.

AI-Assisted Annotation

Increasingly, AI models assist with annotation: pre-labeling data that humans then correct (faster than labeling from scratch), generating synthetic annotation data, or serving as additional annotators alongside humans. This creates an interesting feedback loop: AI helps label data that trains better AI. The risk is that AI-assisted annotation inherits the biases of the assisting model, so human oversight remains essential.

相關概念

← 所有術語
← Alignment Anthropic →