A model's ability to perform accurately on new, previously unseen data.
Generalization is the central objective of machine learning: a model that generalizes well has learned the true underlying structure of a problem rather than memorizing the specific examples it was trained on. When a model generalizes poorly, it may achieve near-perfect accuracy on training data while failing badly on new inputs — a condition known as overfitting. The opposite failure, underfitting, occurs when a model is too simple to capture meaningful patterns even in training data. Striking the right balance is the core challenge of supervised learning.
The theoretical foundation for understanding generalization comes largely from statistical learning theory, particularly the Vapnik-Chervonenkis (VC) framework developed in the 1970s. VC dimension quantifies a model's capacity — roughly, how complex a set of patterns it can represent — and bounds the gap between training and test performance. Related tools include PAC (Probably Approximately Correct) learning theory and Rademacher complexity, all of which formalize the intuition that models with higher capacity require more training data to generalize reliably. Regularization techniques such as L1/L2 penalties, dropout, and early stopping are practical interventions designed to improve generalization by constraining model complexity.
Generalization has taken on renewed importance in the deep learning era, where models with millions or billions of parameters routinely generalize well despite having far more capacity than classical theory would suggest should be safe. Phenomena like double descent — where test error improves again after an initial rise as model size grows — have challenged older theoretical frameworks and spurred new research into implicit regularization, the geometry of loss landscapes, and the role of optimization algorithms in shaping what a model learns. Understanding and reliably achieving generalization remains one of the most active and consequential research areas in machine learning.