AI Governance Institute logo
AI Governance Institute

Practical Governance for Enterprise AI

Monitoring & Drift
MON · Monitoring & DriftMON-002Medium effort

Model Drift Detection

Monitor production AI systems for data drift, concept drift, and output distribution shifts that indicate degraded or changed model behavior.

Objective

Detect when AI model behavior has meaningfully changed from its deployment baseline, triggering investigation and remediation before harm occurs.

Maturity Levels

1

Initial

No drift monitoring exists; model degradation is discovered reactively.

2

Developing

Some output metrics are monitored but input distribution and concept drift are not assessed.

3

Defined

Automated drift detection monitors input distributions, output distributions, and key performance metrics with defined alert thresholds.

4

Managed

Drift alerts are triaged and resolved within defined SLAs; patterns are analyzed to identify root causes.

5

Optimizing

Drift detection models are themselves evaluated for accuracy; thresholds are tuned based on false positive/negative rates.

Evidence Requirements

What an auditor or assessor would expect to see for this control.

  • Drift detection configuration documenting methods, metrics monitored, and alert thresholds for each system
  • Periodic drift reports showing statistical distribution comparisons against the reference baseline
  • Alert records for drift events including detection date, severity, and response timeline
  • Root cause analysis records for significant drift events, with remediation actions and outcomes
  • Retraining or recalibration records triggered by confirmed drift, including test results before redeployment

Implementation Notes

Key steps

  • Monitor three types of drift separately: data drift (input distribution changes), concept drift (the relationship between inputs and outputs changes), and prediction drift (output distribution changes).
  • For LLM-based systems, traditional statistical drift detection does not apply directly — monitor proxy metrics: output length distributions, topic distributions, refusal rates, and human feedback signals.
  • Set drift alerts that notify the model owner, not just the monitoring dashboard — unactioned alerts are as bad as no alerts.
  • Document what actions are taken in response to drift alerts: investigation, retraining, rollback, or escalation.

Example Implementation

Product recommendation model experiencing seasonal purchasing pattern shifts

Drift Monitoring Configuration — Product Recommendation Engine

Monitored drift types and methods:

Drift TypeMethodMetricAlert ThresholdCadence
Data drift (input)Population Stability Index on top 20 featuresPSI> 0.2 (any feature)Daily
Prediction driftKS test on score distributionKS statisticp < 0.01Daily
Business outcome driftCTR on recommended productsClick-through rate> 15% drop from 7-day rolling avgHourly
Latency driftp95 inference timems> 200msReal-time

Alert routing: All alerts → #ml-monitoring Slack channel + model owner email

Response SLAs:

  • Business outcome alert: investigate within 2 hours; escalate if not resolved in 4
  • Data/prediction drift: investigate within 1 business day; present findings in weekly model review

Seasonal adjustment: Baseline updated each quarter to account for known seasonal patterns (documented in monitoring runbook)

Control Details

Control ID
MON-002
Typical owner
AI Engineering / MLOps
Implementation effort
Medium effort
Agent-relevant
No

Tags

model driftdata driftconcept driftmonitoring