By clicking "Accept", you agree to the storing of cookies on your device to enhance site navigation, analyze site usage, and assist in our marketing efforts. See our Privacy Policy for more information
Glossary
Model Drift
AI DEFINITION

Model Drift

Machine learning models are not static. Over time, they decay — a process known as model drift. The accuracy you achieved on day one might look impressive, but as reality shifts, the model’s predictions begin to lose their edge.

Think of recommendation systems: a movie recommender trained on last year’s viewing patterns will miss the sudden popularity of a new genre or director. The same holds for medical AI: changes in patient demographics or treatment protocols can render previously “state-of-the-art” models dangerously outdated.

Types of drift

  • Feature distribution drift: input data statistics change.
  • Label distribution drift: the output itself evolves.
  • Concept drift: the mapping between features and labels shifts.

Why it matters

Unchecked drift leads to silent failures: fraud goes undetected, diagnoses become inaccurate, user experiences deteriorate. This is why modern MLOps emphasizes monitoring, continuous validation, and retraining pipelines.

Big tech companies treat drift not as an exception but as the norm. For example, Google constantly re-trains its spam filters, while Netflix refreshes its recommender algorithms daily.

Model drift is often categorized into two forms: data drift (when the statistical properties of input data change) and concept drift (when the relationship between inputs and outputs shifts). Both erode predictive performance, but concept drift is particularly challenging because it implies that the model’s very assumptions about the world are no longer valid.

Organizations tackling drift usually adopt continuous evaluation pipelines, often integrated with MLOps platforms. These pipelines can automate retraining, but full automation can be risky: retraining on noisy or biased new data may propagate errors. Thus, many companies use a hybrid strategy—combining automation with human-in-the-loop validation.

Beyond technical fixes, drift management is also about governance and agility. Business teams, domain experts, and engineers must collaborate to decide when retraining is justified, how new data is curated, and how to balance costs with accuracy. In high-stakes domains like healthcare or finance, regulatory compliance also mandates transparent monitoring of drift and corrective actions.

🔗 References:

  • Sculley et al., Hidden Technical Debt in Machine Learning Systems (NIPS 2015)