Effects of Matching on Evaluation of Accuracy, Fairness, and Fairness Impossibility in AI-ML Systems

Authors

  • Phillip Honenberger Morgan State University
  • Omolade Ola Morgan State University
  • William Mapp
  • Pilhwa Lee

DOI:

https://doi.org/10.32473/flairs.37.1.135585

Keywords:

matching, fairness, fairness impossibility, algorithmic fairness, statistical methods, AI ethics, COMPAS

Abstract

“Matching” procedures in statistics involve construction of datasets with similar covariates between compared groups. Matching has recently been proposed as a means of addressing fairness impossibility (i.e. inconsistency of fairness metrics) in AI and ML systems: Beigang argues on conceptual grounds that, when matched rather than unmatched datasets are analyzed, the tradeoff between the fairness metrics equalized odds (EO) and positive predictive value (PPV) will be reduced. Here we evaluate matching as a practical rather than merely conceptual approach to reducing fairness impossibility. As a case study we conduct pre-match and post-match analyses on the well-known COMPAS dataset from Broward Co., Florida, 2013-2014. We then reflect on what these results suggest about effects of matching on (a) accuracy estimates, (b) fairness estimates, and (c) difference between fairness estimates – that is, the extent to which matching reduces “fairness impossibility” in practice. We conclude that matching is a promising tool for improving evaluations on all three fronts, but faces problems due to potential biases introduced by matching procedures themselves, as well as limited power under conditions extremely common to ML evaluation contexts such as non-independent variables and relevance of hidden variables.

Downloads

Published

12-05-2024

How to Cite

Honenberger, P., Ola, O., Mapp, W., & Lee, P. (2024). Effects of Matching on Evaluation of Accuracy, Fairness, and Fairness Impossibility in AI-ML Systems. The International FLAIRS Conference Proceedings, 37(1). https://doi.org/10.32473/flairs.37.1.135585

Issue

Section

Special Track: Explainable, Fair, and Trustworthy AI