Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Model Drift

Model Drift

When shifting real-world data patterns cause a deployed ML model's performance to degrade.

Year: 2016Generality: 694
Back to Vocab

Model drift refers to the gradual degradation in a machine learning model's predictive performance caused by changes in the statistical relationships between input features and the target variable over time. Because models are trained on historical data that reflects a particular snapshot of the world, they can become misaligned with reality as the underlying data-generating process evolves. This misalignment manifests as declining accuracy, precision, recall, or other performance metrics — often subtly at first, making it easy to miss without active monitoring.

Two primary forms of drift are commonly distinguished. Data drift (or covariate shift) occurs when the distribution of input features changes — for example, a customer base shifting demographically. Concept drift is more fundamental: the relationship between inputs and outputs itself changes, such as when fraud patterns evolve in response to new detection methods. Both forms can occur gradually or abruptly, and both require different detection and remediation strategies.

Detecting model drift typically involves continuously monitoring model outputs and comparing incoming data distributions against training baselines using statistical tests such as the Kolmogorov-Smirnov test, Population Stability Index (PSI), or Jensen-Shannon divergence. In production systems, shadow models, champion-challenger frameworks, and automated retraining pipelines are common engineering responses. The challenge is compounded in settings where ground-truth labels arrive with significant delay — such as credit default prediction — making it hard to confirm performance degradation quickly.

Model drift is particularly consequential in high-stakes, fast-moving domains like fraud detection, financial forecasting, recommendation systems, and healthcare diagnostics, where stale models can cause real harm or significant business loss. As MLOps practices have matured, drift detection has become a core component of model lifecycle management, with dedicated tooling from platforms like Evidently AI, Arize, and Fiddler. Addressing drift is not a one-time fix but an ongoing operational discipline that distinguishes robust production ML systems from brittle ones.

Related

Related

Model Drift Minimization
Model Drift Minimization

Techniques that keep ML models accurate as real-world data distributions shift over time.

Generality: 694
Criteria Drift
Criteria Drift

When evaluation metrics for a ML model shift over time, degrading measured performance.

Generality: 337
Performance Degradation
Performance Degradation

The decline in an AI model's accuracy or reliability over time or under new conditions.

Generality: 702
Training-Serving Skew
Training-Serving Skew

A mismatch between data distributions seen during training versus real-world inference.

Generality: 620
Model Management
Model Management

Systematic practices for governing ML models across their entire operational lifecycle.

Generality: 710
Model Stability
Model Stability

A model's ability to produce consistent, reliable outputs across varying inputs and data conditions.

Generality: 708