Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Loss Function

Loss Function

A mathematical measure of error that guides model training toward better predictions.

Year: 1986Generality: 909
Back to Vocab

A loss function is a mathematical function that quantifies how far a model's predictions deviate from the true target values. During training, the model produces outputs for a given set of inputs, and the loss function computes a scalar score — often called the loss or cost — that summarizes the magnitude of prediction errors across a batch or dataset. Common examples include mean squared error (MSE) for regression tasks, which penalizes large deviations quadratically, and cross-entropy loss for classification tasks, which measures the divergence between predicted probability distributions and true class labels. The specific choice of loss function encodes assumptions about the problem structure and directly shapes what the model learns to optimize.

The loss function sits at the center of the training loop. Optimization algorithms such as stochastic gradient descent (SGD) compute the gradient of the loss with respect to each model parameter — a process made tractable in neural networks by the backpropagation algorithm — and update parameters in the direction that reduces the loss. This iterative process continues until the loss converges to a minimum or a stopping criterion is met. Because the loss landscape can be highly non-convex in deep networks, the choice of loss function interacts critically with optimizer design, learning rate schedules, and regularization strategies.

Beyond standard supervised learning, loss functions have been extended and customized for a wide range of settings. Contrastive losses and triplet losses power metric learning systems; adversarial losses define the training dynamics of generative adversarial networks; and reinforcement learning uses reward signals that function analogously to negative loss. Researchers also design task-specific losses — such as focal loss for class-imbalanced detection problems or perceptual loss for image synthesis — to encode domain knowledge directly into the optimization objective.

The loss function is arguably the most consequential design decision in building a machine learning system. A poorly chosen loss can cause a model to optimize for a proxy objective that diverges from real-world goals, a phenomenon sometimes called Goodhart's Law in ML contexts. Conversely, a well-designed loss function can dramatically accelerate convergence, improve generalization, and align model behavior with intended outcomes, making it a primary lever for both researchers and practitioners.

Related

Related

Loss Optimization
Loss Optimization

Iteratively adjusting model parameters to minimize prediction error measured by a loss function.

Generality: 875
Objective Function
Objective Function

A mathematical function that quantifies what a machine learning model is optimizing.

Generality: 908
Cross-Entropy Loss
Cross-Entropy Loss

A loss function measuring divergence between predicted probability distributions and true labels.

Generality: 838
Training Objective
Training Objective

The criterion a machine learning model optimizes to learn from data.

Generality: 820
Mean Squared Error
Mean Squared Error

A loss function measuring average squared differences between predicted and actual values.

Generality: 871
Prediction Error
Prediction Error

The gap between a model's predicted values and the actual observed outcomes.

Generality: 875