Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Generalization

Generalization

A model's ability to perform accurately on new, previously unseen data.

Year: 1984Generality: 913
Back to Vocab

Generalization is the central objective of machine learning: a model that generalizes well has learned the true underlying structure of a problem rather than memorizing the specific examples it was trained on. When a model generalizes poorly, it may achieve near-perfect accuracy on training data while failing badly on new inputs — a condition known as overfitting. The opposite failure, underfitting, occurs when a model is too simple to capture meaningful patterns even in training data. Striking the right balance is the core challenge of supervised learning.

The theoretical foundation for understanding generalization comes largely from statistical learning theory, particularly the Vapnik-Chervonenkis (VC) framework developed in the 1970s. VC dimension quantifies a model's capacity — roughly, how complex a set of patterns it can represent — and bounds the gap between training and test performance. Related tools include PAC (Probably Approximately Correct) learning theory and Rademacher complexity, all of which formalize the intuition that models with higher capacity require more training data to generalize reliably. Regularization techniques such as L1/L2 penalties, dropout, and early stopping are practical interventions designed to improve generalization by constraining model complexity.

Generalization has taken on renewed importance in the deep learning era, where models with millions or billions of parameters routinely generalize well despite having far more capacity than classical theory would suggest should be safe. Phenomena like double descent — where test error improves again after an initial rise as model size grows — have challenged older theoretical frameworks and spurred new research into implicit regularization, the geometry of loss landscapes, and the role of optimization algorithms in shaping what a model learns. Understanding and reliably achieving generalization remains one of the most active and consequential research areas in machine learning.

Related

Related

Overfitting
Overfitting

When a model memorizes training data noise instead of learning generalizable patterns.

Generality: 875
Overparameterization Regime
Overparameterization Regime

When a model has more parameters than training samples, yet still generalizes well.

Generality: 520
Regularization
Regularization

A technique that penalizes model complexity to prevent overfitting and improve generalization.

Generality: 876
Learnability
Learnability

Whether and how efficiently a model class can generalize from finite training data.

Generality: 794
Hierarchy of Generalizations
Hierarchy of Generalizations

A layered framework where neural networks learn increasingly abstract data representations.

Generality: 695
Overparameterized
Overparameterized

A model with more parameters than available training data points.

Generality: 590