Model Drift
Model drift is the gradual decline in a machine learning model's accuracy and usefulness that occurs after the model has been deployed to production. It happens because the real-world data the model encounters changes over time, diverging from the data the model was originally trained on. This degradation can cause a model to make increasingly unreliable predictions without any change to the model itself.
Model drift refers to the degradation of a deployed machine learning model's predictive power resulting from changes in the statistical distribution of input data, output variables, or the relationships between input and output variables in the production environment. It is observed during inference, when incoming data deviates from the distribution present in the training dataset. Drift is typically measured by monitoring changes in the distributions of model inputs, outputs, and ground-truth actuals over time. It is a runtime and post-deployment phenomenon and cannot be detected through static code analysis or pre-deployment testing alone, as it manifests only through exposure to live or evolving production data.
Why it matters
Machine learning models are trained on historical data that represents conditions at a specific point in time. Once deployed, those conditions evolve: user behavior shifts, market dynamics change, fraud patterns mutate, and real-world distributions diverge from what the model was built to handle. Because the model itself has not changed, this degradation is silent by default. Without active monitoring, teams may not detect that a model's predictions have become unreliable until business outcomes are already affected.
Who it's relevant to
Inside Model Drift
Common questions
Answers to the questions practitioners most commonly ask about Model Drift.