Adjusting Machine Learning Decisions for Equal Opportunity and Counterfactual Fairness


Journal article


Yixin Wang, Dhanya Sridhar, D. Blei
Trans. Mach. Learn. Res., 2023

Semantic Scholar DBLP
Cite

Cite

APA   Click to copy
Wang, Y., Sridhar, D., & Blei, D. (2023). Adjusting Machine Learning Decisions for Equal Opportunity and Counterfactual Fairness. Trans. Mach. Learn. Res.


Chicago/Turabian   Click to copy
Wang, Yixin, Dhanya Sridhar, and D. Blei. “Adjusting Machine Learning Decisions for Equal Opportunity and Counterfactual Fairness.” Trans. Mach. Learn. Res. (2023).


MLA   Click to copy
Wang, Yixin, et al. “Adjusting Machine Learning Decisions for Equal Opportunity and Counterfactual Fairness.” Trans. Mach. Learn. Res., 2023.


BibTeX   Click to copy

@article{yixin2023a,
  title = {Adjusting Machine Learning Decisions for Equal Opportunity and Counterfactual Fairness},
  year = {2023},
  journal = {Trans. Mach. Learn. Res.},
  author = {Wang, Yixin and Sridhar, Dhanya and Blei, D.}
}

Abstract

Machine learning ( ml ) methods have the potential to automate high-stakes decisions, such as bail admissions or credit lending, by analyzing and learning from historical data. But these algorithmic decisions may be unfair: in learning from historical data, they may replicate discriminatory practices from the past. In this paper, we propose two algorithms that adjust fitted ML predictors to produce decisions that are fair. Our methods provide post-hoc adjustments to the predictors, without requiring that they be retrained. We consider a causal model of the ML decisions, define fairness through counterfactual decisions within the model, and then form algorithmic decisions that capture the historical data as well as possible, but are provably fair. In particular, we consider two definitions of fairness. The first is “equal counterfactual opportunity,” where the counterfactual distribution of the decision is the same regardless of the protected attribute; the second is counterfactual fairness. We evaluate the algorithms, and the trade-o � between accuracy and fairness, on datasets about admissions, income, credit, and recidivism.