Effects of Matching on Evaluation of Accuracy, Fairness, and Fairness Impossibility in AI-ML Systems
DOI :
https://doi.org/10.32473/flairs.37.1.135585Mots-clés :
matching, fairness, fairness impossibility, algorithmic fairness, statistical methods, AI ethics, COMPASRésumé
“Matching” procedures in statistics involve construction of datasets with similar covariates between compared groups. Matching has recently been proposed as a means of addressing fairness impossibility (i.e. inconsistency of fairness metrics) in AI and ML systems: Beigang argues on conceptual grounds that, when matched rather than unmatched datasets are analyzed, the tradeoff between the fairness metrics equalized odds (EO) and positive predictive value (PPV) will be reduced. Here we evaluate matching as a practical rather than merely conceptual approach to reducing fairness impossibility. As a case study we conduct pre-match and post-match analyses on the well-known COMPAS dataset from Broward Co., Florida, 2013-2014. We then reflect on what these results suggest about effects of matching on (a) accuracy estimates, (b) fairness estimates, and (c) difference between fairness estimates – that is, the extent to which matching reduces “fairness impossibility” in practice. We conclude that matching is a promising tool for improving evaluations on all three fronts, but faces problems due to potential biases introduced by matching procedures themselves, as well as limited power under conditions extremely common to ML evaluation contexts such as non-independent variables and relevance of hidden variables.
Téléchargements
Publié-e
Comment citer
Numéro
Rubrique
Licence
© Phillip Honenberger, Omolade Ola, William Mapp, Pilhwa Lee 2024
Cette œuvre est sous licence Creative Commons Attribution - Pas d'Utilisation Commerciale 4.0 International.