ML model drift is not a single phenomenon — it is a family of related degradation patterns that compound silently. Understanding the taxonomy is the prerequisite for building detection that actually works. There are three primary drift types: feature drift (also called data drift or covariate shift), concept drift (the relationship between features and labels changes), and bias drift (systematic errors in model outputs that compound over time).
Feature drift is the most common and most detectable. The distribution of input features changes after training. If your fraud model was trained on pre-pandemic transaction patterns, the feature distributions of 2025 transactions may have drifted significantly enough to invalidate the model's learned feature importances. Detection requires statistical comparison between training distributions and production distributions.
Concept drift is harder to detect because it requires labeled production data — which arrives slowly or not at all for many ML applications. A churn prediction model trained in Q1 may encounter a fundamentally different relationship between usage patterns and churn in Q3 if a competitive product launched mid-year. Without labels, you can only infer concept drift from proxy signals: unexpected changes in output distributions, anomalous feature importance rankings, or sudden degradation in business metrics.
Bias drift accumulates from feedback loops. A recommendation model that slightly over-recommends one category will cause users to interact more with that category, which biases the next training dataset toward that category, amplifying the original bias. Each retraining cycle tightens the feedback loop. Detection requires tracking model output distributions over time and comparing against balanced baselines.
The ML Reliability Score (MRS) provides a composite health signal that aggregates these drift signals into a single actionable metric. The scoring formula weights feature drift (40%), prediction confidence (30%), output distribution stability (20%), and data freshness (10%). An MRS above 0.90 is healthy; 0.75-0.90 warrants investigation; below 0.75 triggers automatic model review.
Z-score based drift detection computes statistical distances between current feature distributions and rolling baselines. For each feature, we compute the z-score of the current mean against the baseline distribution of means across the last N=5 pipeline runs. A z-score above 3.0 on any feature triggers a drift alert. This approach is robust to normal variation while sensitive to genuine distributional shifts.
Practical implementation requires three infrastructure components: a baseline store (DynamoDB with TTL-based retention works well for rolling window baselines), a drift computation layer (integrated into the Gold pipeline validation step), and an alerting path (SENTINEL circuit breakers that hold model refreshes when MRS drops below threshold).
The circuit breaker pattern is critical. When drift is detected, you have two options: immediately retrain (expensive, slow, may introduce worse drift if training data is insufficient) or hold the model in safe mode (serve the previous version, alert the team, collect additional labeled data). The circuit breaker implements the hold-and-alert path automatically, buying time for the team to investigate before committing to retraining.
Observed results from production deployments: drift detection with z-score baselines catches distributional changes an average of 11 days before they surface in business metrics. For a recommendation engine, that is 11 days of protected revenue. For a fraud detection model, it is 11 days of prevented false negatives. The investment in a reliable MRS computation pipeline typically pays for itself on the first caught drift event.
Key Takeaways
- Three distinct drift types require different detection strategies: feature drift (statistical), concept drift (proxy signals), bias drift (output distribution tracking)
- The ML Reliability Score (MRS) aggregates drift signals into a single actionable metric with a 0.90/0.75 threshold ladder
- Z-score based detection with N=5 rolling baselines catches distributional changes 11 days before they surface in business metrics
- Circuit breakers hold model refreshes automatically during drift events — buy time to investigate before committing to retraining