Zubnet AIAprenderWiki › Annotation
Training

Annotation

Data Labeling, Data Annotation
El proceso de añadir labels, tags o metadatos a datos crudos para que puedan usarse para aprendizaje supervisado. Anotar imágenes significa dibujar bounding boxes alrededor de objetos. Anotar texto significa etiquetar entidades, sentimiento o intención. Anotar para RLHF significa rankear respuestas del modelo por calidad. La anotación es el trabajo humano que convierte datos crudos en datos de entrenamiento.

Por qué importa

La anotación es el fundamento poco glamoroso de la IA supervisada. Cada dataset etiquetado, cada modelo fine-tuned, cada asistente alineado depende de anotadores humanos que pasaron horas etiquetando datos correctamente. La calidad de las anotaciones determina directamente la calidad del modelo — etiquetado inconsistente o sesgado produce modelos inconsistentes y sesgados. Es la parte más labor-intensiva y a menudo más cara de construir sistemas IA.

Deep Dive

Annotation workflows typically involve: (1) creating clear labeling guidelines (what counts as "positive sentiment"? what's the boundary of a "car" in a bounding box?), (2) training annotators on the guidelines, (3) annotating data with multiple annotators per example (for quality control), (4) measuring inter-annotator agreement (do annotators agree on labels?), and (5) resolving disagreements (through adjudication or majority vote). Low agreement often indicates ambiguous guidelines or genuinely ambiguous data.

RLHF Annotation

For LLM alignment, annotation means comparing model responses: "Is response A or response B better for this prompt?" This preference annotation is particularly challenging because "better" is subjective, context-dependent, and culturally variable. Annotator demographics, expertise, and instructions all influence the resulting preference data, which in turn shapes model behavior. This is why alignment is often described as encoding the values of whoever writes the annotation guidelines.

AI-Assisted Annotation

Increasingly, AI models assist with annotation: pre-labeling data that humans then correct (faster than labeling from scratch), generating synthetic annotation data, or serving as additional annotators alongside humans. This creates an interesting feedback loop: AI helps label data that trains better AI. The risk is that AI-assisted annotation inherits the biases of the assisting model, so human oversight remains essential.

Conceptos relacionados

← Todos los términos
← Alignment Anthropic →