Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Underfitting

Underfitting

When a model is too simple to capture meaningful patterns in data.

Year: 1992Generality: 720
Back to Vocab

Underfitting occurs when a machine learning model is insufficiently complex to capture the underlying structure of the data it is trained on. The result is a model with high bias — one whose assumptions are too rigid or simplistic to represent the true relationships in the data. Unlike overfitting, where a model memorizes noise and spurious patterns, an underfit model fails to learn even the genuine signal, producing poor performance on both the training set and any new data it encounters.

The root causes of underfitting typically include using a model with too few parameters, training for too few iterations, or applying excessive regularization that penalizes complexity so heavily that the model cannot adapt to the data. For example, fitting a straight line to data with a clearly nonlinear relationship will almost always underfit, regardless of how much training data is available. The model's capacity — its ability to represent a wide range of functions — is simply too limited for the task.

Underfitting is best understood within the bias-variance tradeoff, a foundational concept in statistical learning theory. High bias corresponds to underfitting: the model makes strong, often incorrect assumptions that cause it to systematically miss the true pattern. Reducing underfitting generally requires increasing model capacity — switching to a more expressive architecture, engineering more informative features, reducing regularization strength, or training longer. The goal is to find the sweet spot where the model is complex enough to learn real patterns but not so complex that it overfits.

Recognizing underfitting in practice is relatively straightforward: if training error remains high after reasonable optimization, the model is likely underfit. This distinguishes it from overfitting, where training error is low but validation or test error is high. Monitoring both training and validation performance throughout the learning process is the standard diagnostic approach, and tools like learning curves make it easy to identify whether a model suffers from excessive bias or variance.

Related

Related

Overfitting
Overfitting

When a model memorizes training data noise instead of learning generalizable patterns.

Generality: 875
Bias-Variance Trade-off
Bias-Variance Trade-off

The fundamental tension between model complexity and generalization that governs prediction error.

Generality: 875
Bias-Variance Dilemma
Bias-Variance Dilemma

The fundamental trade-off between model simplicity and sensitivity to training data.

Generality: 838
Simplicity Bias
Simplicity Bias

The tendency of ML models to favor simpler patterns or hypotheses over complex ones.

Generality: 520
Bias-Variance Curve
Bias-Variance Curve

A plot showing how model complexity affects the balance between bias and variance.

Generality: 694
Coverage Bias
Coverage Bias

A dataset imbalance where underrepresented groups cause skewed model performance.

Generality: 520