The ability to deliberately guide a neural network's outputs through controlled input or parameter modifications.
Steerability refers to the capacity to intentionally direct a neural network's outputs in a predictable and controlled manner by applying systematic modifications to its inputs, activations, or internal parameters. Rather than treating a model as a black box that produces fixed outputs for given inputs, steerability treats the model as a controllable system where specific interventions reliably produce specific changes. This property is especially valuable in generative models, where practitioners may want to adjust attributes like style, tone, or content without retraining the entire network from scratch.
In practice, steerability is achieved through several mechanisms. Steering vectors — directions identified in a model's activation space that correspond to meaningful semantic attributes — can be added to or subtracted from intermediate representations to shift outputs in desired ways. In large language models, for instance, a steering vector associated with a concept like "formality" or "sentiment" can be applied at inference time to nudge the model's behavior without any gradient updates. Similarly, in image generation, latent space arithmetic allows users to blend or isolate visual features by manipulating the coordinates of generated samples.
Steerability is closely related to, but distinct from, interpretability. While interpretability asks what a model has learned, steerability asks how that knowledge can be leveraged for targeted control. This makes it a practical tool for alignment research, where the goal is to ensure AI systems behave according to human intentions. Demonstrating that a model is steerable provides evidence that its internal representations are structured and meaningful rather than arbitrary, which in turn supports safer and more predictable deployment.
The concept gained particular traction in the early 2020s as large language models became powerful enough that fine-grained behavioral control became both necessary and technically feasible. Researchers found that activation spaces in transformer-based models often contain linear directions corresponding to human-interpretable concepts, making vector-based steering surprisingly effective. Steerability now sits at the intersection of mechanistic interpretability, AI safety, and controllable generation, representing a key frontier in making powerful models more reliably aligned with user intent.