skip to main content
10.1145/3292500.3330664acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
research-article
Open Access

Mathematical Notions vs. Human Perception of Fairness: A Descriptive Approach to Fairness for Machine Learning

Published:25 July 2019Publication History

ABSTRACT

Fairness for Machine Learning has received considerable attention, recently. Various mathematical formulations of fairness have been proposed, and it has been shown that it is impossible to satisfy all of them simultaneously. The literature so far has dealt with these impossibility results by quantifying the tradeoffs between different formulations of fairness. Our work takes a different perspective on this issue. Rather than requiring all notions of fairness to (partially) hold at the same time, we ask which one of them is the most appropriate given the societal domain in which the decision-making model is to be deployed. We take a descriptive approach and set out to identify the notion of fairness that best captures lay people's perception of fairness. We run adaptive experiments designed to pinpoint the most compatible notion of fairness with each participant's choices through a small number of tests. Perhaps surprisingly, we find that the most simplistic mathematical definition of fairness---namely, demographic parity---most closely matches people's idea of fairness in two distinct application scenarios. This conclusion remains intact even when we explicitly tell the participants about the alternative, more complicated definitions of fairness, and we reduce the cognitive burden of evaluating those notions for them. Our findings have important implications for the Fair ML literature and the discourse on formalizing algorithmic fairness.

References

  1. Julia Angwin, Jeff Larson, Surya Mattu, and Lauren Kirchner. 2016. Machine Bias. Propublica (2016).Google ScholarGoogle Scholar
  2. Edmond Awad, Sohan Dsouza, Richard Kim, Jonathan Schulz, Joseph Henrich, Azim Shariff, Jean-Francc ois Bonnefon, and Iyad Rahwan. 2018. The moral machine experiment. Nature , Vol. 563, 7729 (2018), 59.Google ScholarGoogle Scholar
  3. Anna Barry-Jester, Ben Casselman, and Dana Goldstein. 2015. The New Science of Sentencing. The Marshall Project (August 2015).Google ScholarGoogle Scholar
  4. Reuben Binns, Max Van Kleek, Michael Veale, Ulrik Lyngs, Jun Zhao, and Nigel Shadbolt. 2018. 'It's Reducing a Human Being to a Percentage': Perceptions of Justice in Algorithmic Decisions. In CHI. ACM, 377. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Joy Buolamwini and Timnit Gebru. 2018. Gender shades: Intersectional accuracy disparities in commercial gender classification. In In Proceedings of the 1st Conference on Fairness, Accountability and Transparency (FAT*). 77--91.Google ScholarGoogle Scholar
  6. Alexandra Chouldechova. 2017. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. arXiv preprint arXiv:1703.00056 (2017).Google ScholarGoogle Scholar
  7. Sam Corbett-Davies, Emma Pierson, Avi Feller, Sharad Goel, and Aziz Huq. 2017. Algorithmic decision making and the cost of fairness. In KDD. ACM, 797--806. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Rahul C. Deo. 2015. Machine learning in medicine. Circulation , Vol. 132, 20 (2015), 1920--1930.Google ScholarGoogle ScholarCross RefCross Ref
  9. Cynthia Dwork, Moritz Hardt, Toniann Pitassi, Omer Reingold, and Richard Zemel. 2012. Fairness through awareness. In Proceedings of the Innovations in Theoretical Computer Science Conference (ITCS). ACM, 214--226. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Michael Feldman, Sorelle A. Friedler, John Moeller, Carlos Scheidegger, and Suresh Venkatasubramanian. 2015. Certifying and removing disparate impact. In KDD. ACM, 259--268. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Pratik Gajane and Mykola Pechenizkiy. 2017. On formalizing fairness in prediction with machine learning. arXiv preprint arXiv:1710.03184 (2017).Google ScholarGoogle Scholar
  12. Daniel Golovin, Andreas Krause, and Debajyoti Ray. 2010. Near-optimal bayesian active learning with noisy observations. In NIPS . 766--774. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Nina Grgic-Hlaca, Elissa M. Redmiles, Krishna P Gummadi, and Adrian Weller. 2018. Human perceptions of fairness in algorithmic decision making: A case study of criminal risk prediction. In WWW. 903--912. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Moritz Hardt, Eric Price, and Nati Srebro. 2016. Equality of opportunity in supervised learning. In NIPS. 3315--3323. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Robert David Hart. 2017. If you're not a white male, artificial intelligence's use in healthcare could be dangerous. Quartz (July 2017).Google ScholarGoogle Scholar
  16. Hoda Heidari, Michele Loi, Krishna P. Gummadi, and Andreas Krause. 2019. A Moral Framework for Understanding of Fair ML through Economic Models of Equality of Opportunity. In FAT* . Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Kenneth Holstein, Jennifer Wortman Vaughan, Hal Daumé III, Miro Dud'ik, and Hanna Wallach. 2018. Improving fairness in machine learning systems: What do industry practitioners need?. In CHI . Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Matthew Joseph, Michael Kearns, Jamie Morgenstern, and Aaron Roth. 2016. Fairness in learning: Classic and contextual bandits. In NIPS. 325--333. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Jon Kleinberg, Sendhil Mullainathan, and Manish Raghavan. 2017. Inherent trade-offs in the fair determination of risk scores. In ITCS .Google ScholarGoogle Scholar
  20. Min Kyung Lee and Su Baykal. 2017. Algorithmic Mediation in Group Decisions: Fairness Perceptions of Algorithmically Mediated vs. Discussion-Based Social Division.. In CSCW. 1035--1048. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. Min Kyung Lee, Daniel Kusbit, Anson Kahng, Ji Tae Kim, Xinran Yuan, Allissa Chan, Ritesh Noothigattu, Daniel See, Siheon Lee, and Christos-Alexandros Psomas. 2018. WeBuildAI: Participatory Framework for Fair and Efficient Algorithmic Governance.Google ScholarGoogle Scholar
  22. Ritesh Noothigattu, Snehalkumar 'Neil' S. Gaikwad, Edmond Awad, Sohan Dsouza, Iyad Rahwan, Pradeep Ravikumar, and Ariel D. Procaccia. 2018. A voting-based system for ethical decision making. In AAAI .Google ScholarGoogle Scholar
  23. Kevin Petrasic, Benjamin Saul, James Greig, and Matthew Bornfreund. 2017. Algorithms and bias: What lenders need to know. White & Case (2017).Google ScholarGoogle Scholar
  24. Debajyoti Ray, Daniel Golovin, Andreas Krause, and Colin Camerer. 2012. Bayesian rapid optimal adaptive design (broad): Method and application distinguishing models of risky choice. California Institute of Technology working paper (2012).Google ScholarGoogle Scholar
  25. Cynthia Rudin. 2013. Predictive Policing Using Machine Learning to Detect Patterns of Crime. Wired Magazine (August 2013). Retrieved 4/28/2016.Google ScholarGoogle Scholar
  26. Nripsuta Saxena, Karen Huang, Evan DeFilippis, Goran Radanovic, David Parkes, and Yang Liu. 2018. How Do Fairness Definitions Fare? Examining Public Attitudes Towards Algorithmic Definitions of Fairness. arXiv preprint arXiv:1811.03654 (2018). Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. Till Speicher, Hoda Heidari, Nina Grgic-Hlaca, Krishna P. Gummadi, Adish Singla, Adrian Weller, and Muhammad Bilal Zafar. 2018. A Unified Approach to Quantifying Algorithmic Unfairness: Measuring Individual and Group Unfairness via Inequality Indices. In KDD . Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Latanya Sweeney. 2013. Discrimination in online ad delivery. Queue , Vol. 11, 3 (2013), 10. Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. Anne Taylor and Jackson Wright. 2005. Importance of race/ethnicity in clinical trials. Circulation , Vol. 112, 23 (2005), 3654--3660.Google ScholarGoogle ScholarCross RefCross Ref
  30. Michael Veale, Max Van Kleek, and Reuben Binns. 2018. Fairness and Accountability Design Needs for Algorithmic Support in High-Stakes Public Sector Decision-Making. In CHI. ACM, 440. Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. Allison Woodruff, Sarah E. Fox, Steven Rousso-Schindler, and Jeffrey Warshaw. 2018. A Qualitative Exploration of Perceptions of Algorithmic Fairness. In CHI. ACM, 656. Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez Rodriguez, and Krishna P. Gummadi. 2017. Fairness beyond disparate treatment & disparate impact: Learning classification without disparate mistreatment. In WWW. 1171--1180. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Mathematical Notions vs. Human Perception of Fairness: A Descriptive Approach to Fairness for Machine Learning

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader