Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Induction Head

Induction Head

An attention head that identifies and copies repeated token patterns from earlier context.

Year: 2022Generality: 293
Back to Vocab

An induction head is a specific type of attention head found in transformer models that performs a precise in-context pattern-matching operation: given a token at the current position, it looks back through the sequence to find previous occurrences of that token, then attends to whatever token followed that earlier occurrence. This behavior allows the model to effectively predict that a pattern seen earlier in the sequence will repeat. The mechanism typically requires two attention heads working in concert — a "previous token head" that shifts information about each token to its successor, and the induction head itself that uses this shifted information to match and copy across long distances.

Induction heads were formally characterized and named by researchers at Anthropic in 2022, though the underlying behavior existed in trained transformers before it was identified and studied. The discovery emerged from mechanistic interpretability research — an effort to reverse-engineer what specific components of neural networks actually compute. Induction heads turned out to be remarkably consistent across model sizes and architectures, forming reliably in transformers during training and appearing to be a fundamental computational primitive rather than an artifact of any particular design choice.

The significance of induction heads extends well beyond simple pattern copying. They are considered a key mechanism underlying in-context learning — the striking ability of large language models to adapt to new tasks from just a few examples provided in the prompt, without any weight updates. When a model sees a few input-output demonstrations, induction heads help it recognize the demonstrated pattern and apply it to new inputs. This connection between a concrete, interpretable circuit and a high-level capability like few-shot learning makes induction heads one of the most compelling findings in mechanistic interpretability research.

Studying induction heads has broader implications for understanding how and why large language models generalize. Their consistent emergence across architectures suggests that certain computational structures are strongly favored by gradient descent on language modeling objectives. This makes them a valuable case study for researchers trying to understand neural network behavior, predict model capabilities, and ultimately build more transparent and reliable AI systems.

Related

Related

Multi-Head Attention
Multi-Head Attention

Attention mechanism that jointly attends to information from multiple representation subspaces simultaneously.

Generality: 794
Attention
Attention

A mechanism enabling neural networks to dynamically focus on relevant parts of input.

Generality: 875
Attention Pattern
Attention Pattern

A mechanism that lets neural networks selectively focus on relevant parts of input.

Generality: 752
Attention Mechanisms
Attention Mechanisms

Neural network components that dynamically weight input elements by their contextual relevance.

Generality: 865
Inductive Bias
Inductive Bias

Built-in assumptions that help a learning algorithm generalize beyond its training data.

Generality: 838
Program Induction
Program Induction

Automatically generating programs from data and desired input-output behavior.

Generality: 579