Test error drops, rises, then drops again as model complexity increases.
Double descent describes a counterintuitive pattern in how machine learning models behave as their complexity grows. Classical statistical learning theory predicted a U-shaped bias-variance tradeoff: as model capacity increases, test error first falls as the model learns meaningful structure, then rises as overfitting sets in. Double descent reveals that this picture is incomplete. Beyond the overfitting peak — particularly near the interpolation threshold, where the model has just enough parameters to fit the training data exactly — test error can rise sharply, but then falls again as capacity continues to grow, eventually reaching lower error than before.
The mechanism behind this second descent involves how overparameterized models navigate the space of possible solutions. When a model has far more parameters than training examples, gradient-based optimization tends to find solutions that interpolate the training data while remaining as simple as possible in some implicit sense — a property related to the inductive biases of the optimizer and architecture. These "minimum-norm" solutions often generalize surprisingly well, even though they memorize training labels perfectly. This behavior has been observed across linear models, kernel methods, random forests, and deep neural networks, suggesting it reflects something fundamental about learning rather than a quirk of any particular architecture.
The phenomenon was formally characterized and named in a 2019 paper by Belkin and colleagues, though related observations had appeared in earlier empirical work on neural networks and in classical results on interpolating estimators. The finding prompted a significant reassessment of when and why regularization is necessary, and helped explain why modern deep learning models — which are massively overparameterized by classical standards — generalize as well as they do in practice.
Double descent has practical implications for model selection and training. It suggests that stopping at intermediate model sizes can sometimes yield worse generalization than going much larger, and that the conventional wisdom of "simpler is better" does not always hold. Researchers have since documented epoch-wise double descent, where the same phenomenon appears over training time rather than model size, further enriching the theoretical picture of how overparameterized models learn.