Zubnet AIApprendreWiki › Instruction Following
Using AI

Instruction Following

Instruction Adherence
La capacité d'un modèle d'exécuter précisément ce que l'utilisateur demande — respectant les contraintes de format, les exigences de longueur, les spécifications de style et les instructions comportementales. « Écris exactement 3 points en français à propos de X » teste l'instruction following : la réponse doit être des points (pas des paragraphes), exactement 3 (pas 2 ou 5), en français (pas en anglais), et à propos de X (pas Y).

Pourquoi c'est important

L'instruction following est la capacité LLM la plus pratiquement importante. Les utilisateurs se soucient moins de savoir si un modèle « sait » plus de faits et plus de savoir s'il fait ce qu'ils ont réellement demandé. Un modèle qui écrit de la belle prose mais ignore tes exigences de format est moins utile qu'un qui suit les instructions de façon fiable. C'est pourquoi IFEval et d'autres benchmarks d'instruction following sont devenus centraux à l'évaluation de modèles.

Deep Dive

Instruction following is trained through instruction tuning (SFT on instruction-response pairs) and refined through RLHF/DPO (learning to prefer responses that accurately follow instructions). The quality of instruction-following depends heavily on the diversity and precision of the training data: models that see many examples of "exactly 3 items" learn to count; models that only see vague instructions don't.

Where Models Fail

Common instruction-following failures: ignoring length constraints ("be brief" → still writes paragraphs), format drift (starting with the requested format but reverting to prose), constraint amnesia (following the first constraint but forgetting later ones in a complex instruction), and over-following (interpreting ambiguous instructions too literally or too broadly). These failures are more common in smaller models and become rarer with scale, but even frontier models occasionally miss constraints.

System Prompts and Hierarchy

Instruction following becomes complex when instructions conflict: the system prompt says "always respond in JSON" but the user says "write me a poem." Most models implement an instruction hierarchy where system-level instructions take precedence over user messages, but the boundaries are fuzzy. Well-designed applications structure their instruction hierarchy clearly and test edge cases where different levels of instructions might conflict.

Concepts liés

← Tous les termes
← Inference Instruction Tuning →