skip to main content
10.1145/3308558.3313443acmotherconferencesArticle/Chapter ViewAbstractPublication PageswwwConference Proceedingsconference-collections
research-article

FARE: Diagnostics for Fair Ranking using Pairwise Error Metrics

Published:13 May 2019Publication History

ABSTRACT

Ranking, used extensively online and as a critical tool for decision making across many domains, may embed unfair bias. Tools to measure and correct for discriminatory bias are required to ensure that ranking models do not perpetuate unfair practices. Recently, a number of error-based criteria have been proposed to assess fairness with regard to the treatment of protected groups (as determined by sensitive data attributes, e.g., race, gender, or age). However this has largely been limited to classification tasks, and error metrics used in these approaches are not applicable for ranking. Therefore, in this work we propose to broaden the scope of fairness assessment to include error-based fairness criteria for rankings. Our approach supports three criteria: Rank Equality, Rank Calibration, and Rank Parity, which cover a broad spectrum of fairness considerations from proportional group representation to error rate similarity. The underlying error metrics are formulated to be rank-appropriate, using pairwise discordance to measure prediction error in a model-agnostic fashion. Based on this foundation, we then design a fair auditing mechanism which captures group treatment throughout the entire ranking, generating in-depth yet nuanced diagnostics. We demonstrate the efficacy of our error metrics using real-world scenarios, exposing trade-offs among fairness criteria and providing guidance in the selection of fair-ranking algorithms.

References

  1. Ifeoma Ajunwa, Sorelle Friedler, Carlos E Scheidegger, and Suresh Venkatasubramanian. 2016. Hiring by algorithm: predicting and preventing disparate impact. (2016).Google ScholarGoogle Scholar
  2. Julia Angwin, Jeff Larson, Surya Mattu, and Lauren Kirchner. 2016. Machine bias. Pro Publica (2016).Google ScholarGoogle Scholar
  3. Solon Barocas and Andrew D Selbst. 2016. Big data's disparate impact. Cal. L. Rev. 104(2016), 671.Google ScholarGoogle Scholar
  4. Rich Caruana and Alexandru Niculescu-Mizil. 2004. Data mining in metric space: an empirical analysis of supervised learning performance criteria. In Proceedings of the tenth ACM SIGKDD international conference on Knowledge discovery and data mining. ACM, 69-78. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. L Elisa Celis, Damian Straszak, and Nisheeth K Vishnoi. 2017. Ranking with fairness constraints. arXiv preprint arXiv:1704.06840(2017).Google ScholarGoogle Scholar
  6. Timothy M Chan and Mihai Patrascu. 2010. Counting inversions, offline orthogonal range counting, and related problems. In Proceedings of the twenty-first annual ACM-SIAM symposium on Discrete Algorithms. Society for Industrial and Applied Mathematics, 161-173. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Alexandra Chouldechova. 2017. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big data 5, 2 (2017), 153-163.Google ScholarGoogle Scholar
  8. Sam Corbett-Davies, Emma Pierson, Avi Feller, Sharad Goel, and Aziz Huq. 2017. Algorithmic decision making and the cost of fairness. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM, 797-806. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Corinna Cortes and Mehryar Mohri. 2004. AUC optimization vs. error rate minimization. In Advances in neural information processing systems. 313-320. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Jeffrey Dastin. 2018. Amazon scraps secret AI recruiting tool that showed bias against women. Reuters (2018). https://www.reuters.com/article/us-amazon-com-jobs-automation-insight/amazon-scraps-secret-ai-recruiting-tool-that-showed-bias-against-women-idUSKCN1MK08GGoogle ScholarGoogle Scholar
  11. Persi Diaconis. 1988. Group representations in probability and statistics. Lecture Notes-Monograph Series 11 (1988), i-192.Google ScholarGoogle ScholarCross RefCross Ref
  12. Persi Diaconis and Ronald L Graham. 1977. Spearman's footrule as a measure of disarray. Journal of the Royal Statistical Society. Series B (Methodological) (1977), 262-268.Google ScholarGoogle Scholar
  13. Cynthia Dwork, Nicole Immorlica, Adam Tauman Kalai, and Mark DM Leiserson. 2018. Decoupled Classifiers for Group-Fair and Efficient Machine Learning. In Conference on Fairness, Accountability and Transparency. 119-133.Google ScholarGoogle Scholar
  14. Cynthia Dwork, Ravi Kumar, Moni Naor, and Dandapani Sivakumar. 2001. Rank aggregation methods for the web. In Proceedings of the 10th International Conference on World Wide Web. ACM, 613-622. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Michael Feldman, Sorelle A Friedler, John Moeller, Carlos Scheidegger, and Suresh Venkatasubramanian. 2015. Certifying and removing disparate impact. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM, 259-268. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Malcolm Gladwell. 2011. The order of things. The New Yorker 87, 1 (2011), 68-75.Google ScholarGoogle Scholar
  17. Moritz Hardt, Eric Price, Nati Srebro, 2016. Equality of opportunity in supervised learning. In Advances in Neural Information Processing Systems. 3315-3323. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Ralf Herbrich, Thore Graepel, and Klaus Obermayer. 1999. Support vector learning for ordinal regression. (1999).Google ScholarGoogle Scholar
  19. PDH Hofmann. 1994. Statlog (German Credit Data) Data Set. UCI Repository of Machine Learning Databases (1994).Google ScholarGoogle Scholar
  20. Mikella Hurley and Julius Adebayo. 2016. Credit scoring in the era of big data. Yale JL & Tech. 18(2016), 148.Google ScholarGoogle Scholar
  21. Kalervo Järvelin and Jaana Kekäläinen. 2002. Cumulated gain-based evaluation of IR techniques. ACM Transactions on Information Systems (TOIS) 20, 4 (2002), 422-446. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Maurice G Kendall. 1938. A new measure of rank correlation. Biometrika 30, 1/2 (1938), 81-93.Google ScholarGoogle ScholarCross RefCross Ref
  23. Jon M. Kleinberg, Sendhil Mullainathan, and Manish Raghavan. 2017. Inherent Trade-Offs in the Fair Determination of Risk Scores. In Proceedings of the 8th Conference on Innovation in Theoretical Computer Science.Google ScholarGoogle Scholar
  24. Ravi Kumar and Sergei Vassilvitskii. 2010. Generalized distances between rankings. In Proceedings of the 19th International Conference on World Wide Web. ACM, 571-580. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Tie-Yan Liu. 2009. Learning to rank for information retrieval. Foundations and Trends® in Information Retrieval 3, 3(2009), 225-331. Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. Cathy O'Neil. 2017. Weapons of math destruction: How big data increases inequality and threatens democracy. Broadway Books. Google ScholarGoogle Scholar
  27. Geoff Pleiss, Manish Raghavan, Felix Wu, Jon Kleinberg, and Kilian Q Weinberger. 2017. On fairness and calibration. In Advances in Neural Information Processing Systems. 5680-5689. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Ashudeep Singh and Thorsten Joachims. 2018. Fairness of Exposure in Rankings. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM, 2219-2228. Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. Annie Waldman and SiSi Wei. 2015. Colleges Flush With Cash Saddle Poorest Students With Debt. Pro Publica (2015).Google ScholarGoogle Scholar
  30. Ke Yang and Julia Stoyanovich. 2017. Measuring fairness in ranked outputs. In Proceedings of the 29th International Conference on Scientific and Statistical Database Management. ACM, 22. Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez Rodriguez, and Krishna P Gummadi. 2017. Fairness beyond disparate treatment & disparate impact: Learning classification without disparate mistreatment. In Proceedings of the 26th International Conference on World Wide Web. ACM, 1171-1180. Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. Meike Zehlike, Francesco Bonchi, Carlos Castillo, Sara Hajian, Mohamed Megahed, and Ricardo Baeza-Yates. 2017. Fa* ir: A fair top-k ranking algorithm. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management. ACM, 1569-1578. Google ScholarGoogle ScholarDigital LibraryDigital Library

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Other conferences
    WWW '19: The World Wide Web Conference
    May 2019
    3620 pages
    ISBN:9781450366748
    DOI:10.1145/3308558

    Copyright © 2019 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 13 May 2019

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • research-article
    • Research
    • Refereed limited

    Acceptance Rates

    Overall Acceptance Rate1,899of8,196submissions,23%

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format .

View HTML Format