AI Governance Institute logo
AI Governance Institute

Practical Governance for Enterprise AI

Monitoring & Drift
MON · Monitoring & DriftMON-003High effort

AI Bias and Fairness Monitoring

Continuously monitor AI system outputs for discriminatory patterns across protected demographic attributes in production.

Objective

Detect and remediate disparate impact in AI decision-making before it causes regulatory exposure or harm to affected individuals.

Maturity Levels

1

Initial

No bias monitoring exists; disparate impact is not measured.

2

Developing

Bias was assessed pre-deployment but no ongoing monitoring is in place.

3

Defined

Automated monitoring tracks outcome distributions across protected attributes with defined alert thresholds.

4

Managed

Monitoring results are reviewed by a designated fairness owner; disparate impact findings trigger investigation.

5

Optimizing

Fairness metrics are included in model deployment gates; monitoring methodology is externally validated.

Evidence Requirements

What an auditor or assessor would expect to see for this control.

  • Fairness metrics documentation defining which demographic groups are monitored and which metrics are used (e.g., demographic parity, equalized odds)
  • Periodic bias monitoring reports showing metric values by group over a defined period
  • Disparity alert records for any group difference exceeding defined thresholds, with investigation and resolution
  • Remediation records for confirmed bias findings, including actions taken and post-remediation metric results
  • Legal or compliance review records confirming the monitoring approach satisfies applicable anti-discrimination obligations

Implementation Notes

Key steps

  • Define which fairness metrics apply to your use case before deploying — demographic parity, equalized odds, and individual fairness make different trade-offs and are not simultaneously achievable.
  • Identify which protected attributes are relevant to each AI use case and ensure you have sufficient data to detect disparate impact across those groups.
  • Establish a threshold for 'material disparity' in advance — without a pre-defined threshold, remediation decisions become subjective and difficult to defend.
  • Report bias monitoring results to governance bodies regularly; bias that is known and unaddressed creates significant regulatory and legal risk.

Example Implementation

Consumer lending company monitoring an AI model that influences credit limit decisions

Bias and Fairness Monitoring Dashboard — Credit Limit Model

Protected attributes monitored: Gender, race/ethnicity (proxied via zip code), age group (18–25, 26–45, 46–65, 65+)

Fairness metric: Demographic parity — difference in approval rate between groups

Monthly monitoring report (sample):

GroupApproval RateVs. Reference GroupMaterial Disparity?
Male63.2%Reference
Female61.8%-1.4 ppNo (threshold: > 5 pp)
Age 18–2554.1%-9.1 ppYes — investigate
Age 26–4566.3%+3.1 ppNo
Age 46–6564.2%+1.0 ppNo

Materiality threshold: 5 percentage point difference from reference group

Action taken for Age 18–25 flag: Opened investigation to determine if disparity is explained by legitimate credit risk factors or represents proxy discrimination — Legal and Compliance notified per policy

Governance reporting: Results presented to AI Risk Committee monthly

Control Details

Control ID
MON-003
Typical owner
AI Governance Team / Legal / Compliance
Implementation effort
High effort
Agent-relevant
No

Tags

bias monitoringfairnessdisparate impactalgorithmic accountability