Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Zero-Shot Learning (ZSL)

Zero-Shot Learning (ZSL)

A technique enabling models to recognize concepts never encountered during training.

Year: 2009Generality: 620
Back to Vocab

Zero-Shot Learning (ZSL) is a machine learning paradigm in which a model successfully classifies or reasons about categories it has never directly observed during training. Rather than relying on labeled examples for every possible class, ZSL models exploit auxiliary information — such as semantic attribute vectors, natural language descriptions, or knowledge graph embeddings — to bridge the gap between known and unknown categories. The core intuition is that if a model understands that a "zebra" is like a "horse" but with stripes, it can recognize zebras without ever having seen one, provided it has learned what stripes look like and how attributes relate to visual features.

The mechanics of ZSL typically involve learning a shared embedding space where both visual features and semantic descriptors can be compared. During training, the model learns to map inputs from seen classes into this space alongside their semantic representations. At inference time, unseen class descriptors are projected into the same space, and predictions are made by finding the nearest semantic neighbor to a given input. More advanced formulations — known as Generalized Zero-Shot Learning (GZSL) — require the model to handle both seen and unseen classes simultaneously, which is considerably harder and more realistic.

ZSL gained significant traction in the computer vision community around 2009, when researchers began formalizing attribute-based recognition frameworks. The approach has since expanded into natural language processing, where large pretrained language models exhibit remarkable zero-shot capabilities by leveraging learned world knowledge to perform tasks specified only through natural language prompts — no task-specific training examples required.

The importance of ZSL extends well beyond benchmark performance. It directly addresses one of machine learning's most persistent bottlenecks: the need for large, carefully labeled datasets for every target category. In domains like rare disease diagnosis, ecological monitoring, or low-resource language understanding, labeled data is scarce by definition. ZSL offers a principled path toward models that generalize the way humans do — drawing on structured prior knowledge to reason confidently about the unfamiliar.

Related

Related

Zero-shot Capability
Zero-shot Capability

An AI model's ability to perform unseen tasks without task-specific training examples.

Generality: 650
One-Shot Learning
One-Shot Learning

A learning paradigm where models generalize effectively from a single training example per class.

Generality: 694
FSL (Few-Shot Learning)
FSL (Few-Shot Learning)

Training models to generalize accurately from only a handful of labeled examples.

Generality: 710
Few-Shot Learning
Few-Shot Learning

Training ML models to generalize accurately from only a handful of labeled examples.

Generality: 759
SSL (Self-Supervised Learning)
SSL (Self-Supervised Learning)

A learning paradigm where models generate their own supervisory signal from unlabeled data.

Generality: 820
Similarity Learning
Similarity Learning

Training models to measure meaningful similarity between data points for comparison tasks.

Generality: 694