machine-learning statistics

Definition

Ethical AI

Ethical AI is the study of systematic bias, transparency, and accountability in automated decision-making systems. The primary objective is to ensure that machine learning models do not perpetuate discrimination against individuals based on protected attributes (e.g., race, gender) through various formal fairness criteria.

Statistical Fairness Criteria

Group Fairness: Requires specific performance metrics to be equal across protected groups. Common measures include:

  • Demographic Parity: .
  • Calibration: .
  • Error Rate Balance: Equalising False Positive and False Negative rates across groups.

Individual Fairness: Posits that similar individuals should be treated similarly. Formally, for a distance metric and model : .

Fair Regression

In numerical prediction tasks, fairness is often achieved by identifying and adjusting for the influence of protected features.

Mitigation via Substitution: A common strategy involves training a standard regression model (where is the protected attribute) and then replacing with a constant during inference. Setting to the average value of the privileged group ensures that predictions are debiased relative to the protected characteristic while preserving the influence of non-protected features.

Theoretical Constraints

Impossibility Theorem: Established by Kleinberg et al. (2017), it is mathematically impossible to satisfy calibration, false positive parity, and false negative parity simultaneously, except in cases of perfect prediction or identical group base rates. This necessitates explicit policy decisions regarding the distribution of error types.

Transparency and Accountability

Model Cards: Technical documentation detailing model architecture, training data characteristics, and performance disparities to ensure transparency and identify potential for disparate impact.