A model's ability to accurately understand and execute user-specified tasks.
Instruction-following refers to the capacity of a language model or AI system to correctly interpret natural language directives and carry out the intended task. Rather than simply predicting the next token in a sequence, an instruction-following model must parse the user's goal, resolve ambiguities in phrasing, and produce an output that satisfies the request — whether that means writing code, summarizing a document, answering a question, or completing a multi-step procedure. This capability is distinct from raw language modeling and requires the model to generalize across a wide variety of task formats it may not have seen verbatim during pretraining.
The primary mechanism for instilling instruction-following behavior in large language models is supervised fine-tuning on curated instruction-response pairs, often followed by reinforcement learning from human feedback (RLHF). In the supervised phase, models are trained on datasets where each example pairs a natural language instruction with a high-quality completion, teaching the model to treat prompts as directives rather than text to continue. RLHF then refines this behavior by using human preference judgments to reward outputs that are helpful, accurate, and appropriately scoped. Techniques like InstructGPT (2022) and subsequent work demonstrated that even modest amounts of instruction-tuning data could dramatically shift model behavior toward user intent.
Instruction-following became a central research focus around 2021–2022 with the release of models like FLAN, InstructGPT, and later ChatGPT, which showed that fine-tuned models substantially outperformed base models on user-facing tasks despite having fewer parameters. The capability matters because it bridges the gap between what a pretrained model can do and what it will do when prompted by a non-expert user. A model with strong instruction-following can be deployed across diverse applications — coding assistants, document editors, customer support — without requiring users to craft elaborate prompts.
Instruction-following also raises important alignment considerations. A model that follows instructions too literally may miss the user's deeper intent, while one that interprets too liberally may overstep. Calibrating this balance — being helpful without being sycophantic or unsafe — remains an active area of research in AI alignment and evaluation.