A plot showing how model complexity affects the balance between bias and variance.
The bias-variance curve is a diagnostic visualization that captures one of the most fundamental tensions in supervised machine learning: the trade-off between a model's ability to fit training data and its ability to generalize to unseen examples. As model complexity increases — whether through added parameters, deeper architectures, or reduced regularization — the curve tracks how two competing sources of prediction error evolve in opposite directions. Bias, the error introduced by overly simplistic assumptions, decreases as complexity grows. Variance, the error arising from excessive sensitivity to the specific training sample, increases. The curve makes this dynamic legible at a glance.
In practice, the curve plots total expected prediction error (or a proxy like test loss) against a complexity axis — which might represent polynomial degree, tree depth, number of hidden units, or regularization strength. At low complexity, the model underfits: both training and test error are high due to strong bias. As complexity rises, training error falls while test error initially follows, then diverges upward as variance dominates. The point of minimum test error marks the sweet spot where bias and variance are jointly minimized — the model is expressive enough to capture real structure without memorizing noise.
The bias-variance curve is closely related to the classic learning curve and is a cornerstone of model selection methodology. It provides intuition for why cross-validation is necessary, why regularization helps, and why simply minimizing training loss is insufficient. The curve also underpins ensemble methods like bagging, which explicitly target variance reduction, and boosting, which targets bias. More recently, the discovery of double descent — where test error decreases again after a second peak at very high model complexity — has extended and complicated the traditional picture, revealing that modern overparameterized models like deep neural networks do not always follow the expected U-shaped test error curve.
Understanding the bias-variance curve is essential for any practitioner diagnosing model behavior, tuning hyperparameters, or choosing between model families. It transforms abstract statistical concepts into actionable guidance, making it one of the most pedagogically and practically valuable tools in the machine learning toolkit.