When shifting real-world data patterns cause a deployed ML model's performance to degrade.
Model drift refers to the gradual degradation in a machine learning model's predictive performance caused by changes in the statistical relationships between input features and the target variable over time. Because models are trained on historical data that reflects a particular snapshot of the world, they can become misaligned with reality as the underlying data-generating process evolves. This misalignment manifests as declining accuracy, precision, recall, or other performance metrics — often subtly at first, making it easy to miss without active monitoring.
Two primary forms of drift are commonly distinguished. Data drift (or covariate shift) occurs when the distribution of input features changes — for example, a customer base shifting demographically. Concept drift is more fundamental: the relationship between inputs and outputs itself changes, such as when fraud patterns evolve in response to new detection methods. Both forms can occur gradually or abruptly, and both require different detection and remediation strategies.
Detecting model drift typically involves continuously monitoring model outputs and comparing incoming data distributions against training baselines using statistical tests such as the Kolmogorov-Smirnov test, Population Stability Index (PSI), or Jensen-Shannon divergence. In production systems, shadow models, champion-challenger frameworks, and automated retraining pipelines are common engineering responses. The challenge is compounded in settings where ground-truth labels arrive with significant delay — such as credit default prediction — making it hard to confirm performance degradation quickly.
Model drift is particularly consequential in high-stakes, fast-moving domains like fraud detection, financial forecasting, recommendation systems, and healthcare diagnostics, where stale models can cause real harm or significant business loss. As MLOps practices have matured, drift detection has become a core component of model lifecycle management, with dedicated tooling from platforms like Evidently AI, Arize, and Fiddler. Addressing drift is not a one-time fix but an ongoing operational discipline that distinguishes robust production ML systems from brittle ones.