Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Zero-shot Capability

Zero-shot Capability

An AI model's ability to perform unseen tasks without task-specific training examples.

Year: 2020Generality: 650
Back to Vocab

Zero-shot capability refers to the ability of a machine learning model to successfully perform tasks it was never explicitly trained on, without requiring any labeled examples or fine-tuning at inference time. Rather than memorizing task-specific patterns, a zero-shot capable model generalizes from its broad training distribution to recognize and respond to entirely novel instructions or categories. This stands in contrast to few-shot learning, which provides a handful of examples, and traditional supervised learning, which demands substantial labeled data for every target task.

The mechanism behind zero-shot capability depends heavily on the scale and diversity of pretraining. Large language models trained on vast corpora of text develop rich internal representations of concepts, relationships, and instructions that transfer across domains. When prompted with a task description in natural language — such as "translate this sentence to French" or "classify this review as positive or negative" — the model leverages its latent knowledge to produce a reasonable output without ever having seen that specific task framing during training. Multimodal models like CLIP extend this principle across modalities, matching images to textual descriptions of categories the model was never explicitly trained to recognize.

Zero-shot capability matters because it dramatically reduces the cost and friction of deploying AI systems. Traditional pipelines require curating labeled datasets and retraining or fine-tuning models for each new application — a process that is expensive, slow, and often impractical for low-resource languages or specialized domains. Zero-shot models can be redirected to new tasks simply by changing the prompt or instruction, enabling rapid prototyping and broader accessibility.

The practical significance of zero-shot capability became widely recognized with the release of GPT-3 in 2020, which demonstrated surprisingly strong zero-shot performance across diverse benchmarks. Subsequent work on instruction tuning — training models explicitly to follow natural language directives — has further sharpened these abilities, with models like InstructGPT and GPT-4 showing robust zero-shot generalization across reasoning, translation, summarization, and code generation tasks. Zero-shot capability is now considered a core benchmark for evaluating the general intelligence of large-scale AI systems.

Related

Related

Zero-Shot Learning (ZSL)
Zero-Shot Learning (ZSL)

A technique enabling models to recognize concepts never encountered during training.

Generality: 620
One-Shot Learning
One-Shot Learning

A learning paradigm where models generalize effectively from a single training example per class.

Generality: 694
Few-Shot Learning
Few-Shot Learning

Training ML models to generalize accurately from only a handful of labeled examples.

Generality: 759
Transfer Capability
Transfer Capability

An AI system's ability to apply knowledge learned in one domain to another.

Generality: 650
FSL (Few-Shot Learning)
FSL (Few-Shot Learning)

Training models to generalize accurately from only a handful of labeled examples.

Generality: 710
In-Context Learning
In-Context Learning

A model learns new tasks from prompt examples alone, without any weight updates.

Generality: 717