Zubnet AI学习Wiki › Annotation
Training

Annotation

Data Labeling, Data Annotation
给原始数据添加标签、标记、元数据让它能用于监督学习的过程。标注图像意味着在物体周围画边界框。标注文本意味着标记实体、情感、意图。为 RLHF 标注意味着按质量排名模型响应。标注是把原始数据变成训练数据的人类劳动。

为什么重要

标注是监督 AI 不光鲜的基础。每个带标签数据集、每个 fine-tuned 模型、每个对齐的助手,都依赖花数小时正确标注数据的人工标注者。标注质量直接决定模型质量 — 不一致或有偏的标注产出不一致、有偏的模型。它是构建 AI 系统最劳动密集、往往最昂贵的部分。

Deep Dive

Annotation workflows typically involve: (1) creating clear labeling guidelines (what counts as "positive sentiment"? what's the boundary of a "car" in a bounding box?), (2) training annotators on the guidelines, (3) annotating data with multiple annotators per example (for quality control), (4) measuring inter-annotator agreement (do annotators agree on labels?), and (5) resolving disagreements (through adjudication or majority vote). Low agreement often indicates ambiguous guidelines or genuinely ambiguous data.

RLHF Annotation

For LLM alignment, annotation means comparing model responses: "Is response A or response B better for this prompt?" This preference annotation is particularly challenging because "better" is subjective, context-dependent, and culturally variable. Annotator demographics, expertise, and instructions all influence the resulting preference data, which in turn shapes model behavior. This is why alignment is often described as encoding the values of whoever writes the annotation guidelines.

AI-Assisted Annotation

Increasingly, AI models assist with annotation: pre-labeling data that humans then correct (faster than labeling from scratch), generating synthetic annotation data, or serving as additional annotators alongside humans. This creates an interesting feedback loop: AI helps label data that trains better AI. The risk is that AI-assisted annotation inherits the biases of the assisting model, so human oversight remains essential.

相关概念

← 所有术语
← Alignment Anthropic →