skip to main content
10.1145/3368089.3409697acmconferencesArticle/Chapter ViewAbstractPublication PagesfseConference Proceedingsconference-collections
research-article

Fairway: a way to build fair ML software

Published: 08 November 2020 Publication History
  • Get Citation Alerts
  • Abstract

    Machine learning software is increasingly being used to make decisions that affect people's lives. But sometimes, the core part of this software (the learned model), behaves in a biased manner that gives undue advantages to a specific group of people (where those groups are determined by sex, race, etc.). This "algorithmic discrimination" in the AI software systems has become a matter of serious concern in the machine learning and software engineering community. There have been works done to find "algorithmic bias" or "ethical bias" in the software system. Once the bias is detected in the AI software system, the mitigation of bias is extremely important. In this work, we a)explain how ground-truth bias in training data affects machine learning model fairness and how to find that bias in AI software,b)propose a method Fairway which combines pre-processing and in-processing approach to remove ethical bias from training data and trained model. Our results show that we can find bias and mitigate bias in a learned model, without much damaging the predictive performance of that model. We propose that (1) testing for bias and (2) bias mitigation should be a routine part of the machine learning software development life cycle. Fairway offers much support for these two purposes.

    Supplementary Material

    Auxiliary Teaser Video (fse20main-p232-p-teaser.mp4)
    fse20main-p232-p-teaser.mp4 is the teaser video and fse20main-p232-p-main.mp4 video is the whole talk.
    Auxiliary Presentation Video (fse20main-p232-p-video.mp4)
    fse20main-p232-p-teaser.mp4 is the teaser video and fse20main-p232-p-main.mp4 video is the whole talk.

    References

    [1]
    “Health care start-up says a.i. can diagnose patients better than humans can, doctors call that 'dubious',” CNBC, June 2018. [Online]. Available: https://www.cnbc.com/ 2018 /06/28/babylon-claims-its-ai-can-diagnosepatients-better-than-doctors.html
    [2]
    E. Strickland, “ Doc bot preps for the o.r.” IEEE Spectrum, vol. 53, no. 6, pp. 32-60, June 2016.
    [3]
    “ On orbitz, mac users steered to pricier hotels,” 2012. [Online]. Available: https: //www.wsj.com/articles/SB10001424052702304458604577488822667325882
    [4]
    “The algorithm that beats your bank manager,” 2011. [Online]. Available: https://www.forbes.com/sites/parmyolson/2011/03/15/the-algorithm-thatbeats-your-bank-manager/#15da2651ae99
    [5]
    “ Machine bias: There's software used across the country to predict future criminals. and it's biased against blacks,” 2016. [Online]. Available: https://www. propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing
    [6]
    “Can you program ethics into a self-driving car ?” 2016. [Online]. Available: https://spectrum.ieee.org/transportation/self-driving/ can-you-programethics-into-a-selfdriving-car
    [7]
    R. Angell, B. Johnson, Y. Brun, and A. Meliou, “Themis: Automatically testing software for discrimination,” ser. ESEC/FSE 18.
    [8]
    Y. Brun and A. Meliou, “Software fairness,” in ESEC/FSE 2018, 2018.
    [9]
    “Acm conference on fairness, accountability, and transparency (acm fat* ).” [Online]. Available: https://fatconference.org/
    [10]
    “Explain 2019.” [Online]. Available: https://2019.ase-conferences.org/home/ explain-2019
    [11]
    A. Aggarwal, P. Lohia, S. Nagar, K. Dey, and D. Saha, “ Black box fairness testing of machine learning models,” in Proceedings of the 2019 27th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering, ser. ESEC/FSE 2019. New York, NY, USA: ACM, 2019, pp. 625-635. [Online]. Available: http://doi.acm.org/10.1145/3338906.3338937
    [12]
    “White men account for 72% of corporate leadership at 16 of the fortune 500 companies,” 2017. [Online]. Available: https://fortune.com/ 2017 /06/09/whitemen-senior-executives-fortune-500-companies-diversity-data/
    [13]
    I. Chen, F. D. Johansson, and D. Sontag, “Why is my classifier discriminatory?” 2018.
    [14]
    R. Berk, H. Heidari, S. Jabbari, M. Joseph, M. Kearns, J. Morgenstern, S. Neel, and A. Roth, “ A convex framework for fair regression,” 2017.
    [15]
    J. Chakraborty, “Fairway,” 6 2020. [Online]. Available: https://figshare.com/ articles/software/Fairway/12521408
    [16]
    “Google's sentiment analyzer thinks being gay is bad,” Motherboard, Oct 2017. [Online]. Available: https://bit.ly/2yMax8V
    [17]
    “ Google apologizes for mis-tagging photos of african americans, ” July 2015. [Online]. Available: https://cbsn.ws/2LBYbdy
    [18]
    A. Caliskan, J. J. Bryson, and A. Narayanan, “ Semantics derived automatically from language corpora contain human-like biases,” Science, vol. 356, no. 6334, pp. 183-186, 2017. [Online]. Available: https://science.sciencemag.org/content/356/ 6334/183
    [19]
    R. Tatman, “ Gender and dialect bias in YouTube's automatic captions,” in Proceedings of the First ACL Workshop on Ethics in Natural Language Processing. Valencia, Spain: Association for Computational Linguistics, Apr. 2017, pp. 53-59. [Online]. Available: https://www.aclweb.org/anthology/W17-1606
    [20]
    “ Study finds gender and skin-type bias in commercial artificial-intelligence systems,” 2018. [Online]. Available: http://news.mit.edu/2018/study-finds-genderskin-type-bias-artificial-intelligence-systems-0212
    [21]
    “Machine bias,” www.propublica.org, May 2016. [Online]. Available: https://www. propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing
    [22]
    “ Amazon scraps secret ai recruiting tool that showed bias against women, ” Oct 2018. [Online]. Available: https://www.reuters.com/article/us-amazoncom-jobs-automation-insight/amazon-scraps-secret-ai-recruiting-tool-thatshowed-bias-against-women-idUSKCN1MK08G
    [23]
    “ Ethically-aligned design: A vision for prioritizing human well-begin with autonomous and intelligence systems.” 2019.
    [24]
    “Ethics guidelines for trustworthy artificial intelligence.” 2018. [Online]. Available: https://ec.europa.eu/digital-single-market/en/news/ethics-guidelinestrustworthy-ai
    [25]
    “ Microsoft ai principles. 2019.” 2019. [Online]. Available: https://www.microsoft. com/en-us/ai/our-approach-to-ai
    [26]
    “ Ai fairness 360 : An extensible toolkit for detecting, understanding, and mitigating unwanted algorithmic bias,” 10 2018. [Online]. Available: https: //github.com/IBM/AIF360
    [27]
    “Fate: Fairness, accountability, transparency, and ethics in ai,” 2018. [Online]. Available: https://www.microsoft.com/en-us/research/group/fate/
    [28]
    “ Facebook says it has a tool to detect bias in its artificial intelligence,” 2018. [Online]. Available: https://qz.com/1268520/facebook-says-it-has-a-toolto-detect-bias-in-its-artificial-intelligence/
    [29]
    F. Tramer, V. Atlidakis, R. Geambasu, D. Hsu, J.-P. Hubaux, M. Humbert, A. Juels, and H. Lin, “Fairtest: Discovering unwarranted associations in data-driven applications,” EuroS&P17, Apr.
    [30]
    S. Galhotra, Y. Brun, and A. Meliou, “ Fairness testing: testing software for discrimination, ” Proceedings of the 2017 11th Joint Meeting on Foundations of Software Engineering-ESEC/FSE 2017, 2017. [Online]. Available: http://dx.doi.org/10.1145/3106237.3106277
    [31]
    S. Udeshi, P. Arora, and S. Chattopadhyay, “ Automated directed fairness testing,” Proceedings of the 33rd ACM/IEEE International Conference on Automated Software Engineering-ASE 2018, 2018. [Online]. Available: http://dx.doi.org/10.1145/3238147.3238165
    [32]
    L. Zhang, Y. Wu, and X. Wu, “ Situation testing-based discrimination discovery: A causal inference approach,” in Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, ser. IJCAI'16. AAAI Press, 2016, p. 2718-2724.
    [33]
    F. Kamiran and T. Calders, “ Data preprocessing techniques for classification without discrimination, ” Knowledge and Information Systems, vol. 33, no. 1, pp. 1-33, Oct 2012. [Online]. Available: https://doi.org/10.1007/s10115-011-0463-8
    [34]
    F. Calmon, D. Wei, B. Vinzamuri, K. Natesan Ramamurthy, and K. R. Varshney, “ Optimized pre-processing for discrimination prevention,” in Advances in Neural Information Processing Systems 30, I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, Eds. Curran Associates, Inc., 2017, pp. 3992-4001. [Online]. Available: http://papers.nips.cc/paper/6988-optimizedpre-processing-for-discrimination-prevention.pdf
    [35]
    B. H. Zhang, B. Lemoine, and M. Mitchell, “ Mitigating unwanted biases with adversarial learning,” 2018.
    [36]
    T. Kamishima, S. Akaho, H. Asoh, and J. Sakuma, “ Fairness-aware classifier with prejudice remover regularizer,” in Machine Learning and Knowledge Discovery in Databases, P. A. Flach, T. De Bie, and N. Cristianini, Eds. Berlin, Heidelberg: Springer Berlin Heidelberg, 2012, pp. 35-50.
    [37]
    F. Kamiran, S. Mansha, A. Karim, and X. Zhang, “ Exploiting reject option in classification for social discrimination control,” Inf. Sci., 2018.
    [38]
    G. Pleiss, M. Raghavan, F. Wu, J. Kleinberg, and K. Q. Weinberger, “ On fairness and calibration,” 2017.
    [39]
    M. Hardt, E. Price, and N. Srebro, “ Equality of opportunity in supervised learning,” 2016.
    [40]
    J. Martin, “ Bias or systematic error (validity ),” 2010. [Online]. Available: https://www.ctspedia.org/do/view/CTSpedia/BiasDefinition
    [41]
    R. Bellamy, K. Dey, M. Hind, S. C. Hofman, S. Houde, K. Kannan, P. Lohia, J. Martino, S. Mehta, A. Mojsilovic, S. Nagar, K. Natesan Ramamurthy, J. Richards, D. Saha, P. Sattigeri, M. Singh, R. Kush, and Y. Zhang, “ Ai fairness 360 : An extensible toolkit for detecting, understanding, and mitigating unwanted algorithmic bias,” 10 2018.
    [42]
    S. Corbett-Davies, E. Pierson, A. Feller, S. Goel, and A. Huq, “ Algorithmic decision making and the cost of fairness,” 2017.
    [43]
    M. Feldman, S. Friedler, J. Moeller, C. Scheidegger, and S. Venkatasubramanian, “ Certifying and removing disparate impact,” 2014.
    [44]
    A. Chouldechova, “ Fair prediction with disparate impact: A study of bias in recidivism prediction instruments,” 2016.
    [45]
    J. Kleinberg, S. Mullainathan, and M. Raghavan, “ Inherent trade-ofs in the fair determination of risk scores,” 2016.
    [46]
    A. Beutel, J. Chen, T. Doshi, H. Qian, A. Woodruf, C. Luu, P. Kreitmann, J. Bischof, and E. H. Chi, “ Putting fairness principles into practice: Challenges, metrics, and improvements,” 2019.
    [47]
    J. Chakraborty, T. Xia, F. M. Fahid, and T. Menzies, “ Software engineering for fairness: A case study with hyperparameter optimization,” 2019.
    [48]
    “ Uci:adult data set,” 1994. [Online]. Available: http://mlr.cs.umass.edu/ml/ datasets/Adult
    [49]
    “propublica/compas-analysis,” 2015. [Online]. Available: https://github.com/ propublica/compas-analysis
    [50]
    “ Uci:statlog (german credit data) data set,” 2000. [Online]. Available: https: //archive.ics.uci.edu/ml/datasets/Statlog+(German+Credit+Data)
    [51]
    “ Uci:default of credit card clients data set,” 2016. [Online]. Available: https://archive.ics.uci.edu/ml/datasets/default+of+credit+card+clients
    [52]
    “ Uci:heart disease data set,” 2001. [Online]. Available: https://archive.ics.uci.edu/ ml/datasets/Heart+Disease
    [53]
    “Fairware 2018 :international workshop on software fairness.” [Online]. Available: http://fairware.cs.umass.edu/
    [54]
    “ Amazon just showed us that 'unbiased' algorithms can be inadvertently racist.” [Online]. Available: https://www.businessinsider. com/how-algorithms-can-beracist-2016-4
    [55]
    “ Why split data in the ratio 70 :30?” 2012. [Online]. Available: http://informationgain.blogspot.com/
    [56]
    V. Nair, Z. Yu, T. Menzies, N. Siegmund, and S. Apel, “ Finding faster configurations using flash, ” TSE, pp. 1-1, 2018.
    [57]
    V. Nair, T. Menzies, N. Siegmund, and S. Apel, “ Using bad learners to find good configurations,” in Proceedings of the 2017 11th Joint Meeting on Foundations of Software Engineering, 2017, pp. 257-267.
    [58]
    R. Storn and K. V. Price, “ Diferential evolution-a simple and eficient heuristic for global optimization over continuous spaces, ” Journal of Global Optimization, vol. 11, pp. 341-359, 1997.
    [59]
    K. Deb, A. Pratap, S. Agarwal, and T. Meyarivan, “ A fast and elitist multiobjective genetic algorithm: Nsga-ii,” IEEE Transactions on Evolutionary Computation, vol. 6, no. 2, pp. 182-197, April 2002.
    [60]
    K. Holstein, J. Wortman Vaughan, H. Daumé, M. Dudik, and H. Wallach, “ Improving fairness in machine learning systems, ” Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems-CHI '19, 2019. [Online]. Available: http://dx.doi.org/10.1145/3290605.3300830
    [61]
    S. M. Lundberg and S.-I. Lee, “ A unified approach to interpreting model predictions,” in Advances in Neural Information Processing Systems 30, I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, Eds. Curran Associates, Inc., 2017, pp. 4765-4774. [Online]. Available: http://papers.nips.cc/paper/7062-a-unified-approach-to-interpretingmodel-predictions.pdf

    Cited By

    View all
    • (2024)Mitigating Biases in Training Data: Technical and Legal Challenges for Sub-Saharan AfricaInternational Journal of Applied Research in Business and Management10.51137/ijarbm.2024.5.1.105:1(209-224)Online publication date: 24-May-2024
    • (2024)Assessing and Mitigating Bias in Artificial Intelligence: A ReviewRecent Advances in Computer Science and Communications10.2174/266625581666623052311442517:1Online publication date: Jan-2024
    • (2024)Predicting Fairness of ML Software ConfigurationsProceedings of the 20th International Conference on Predictive Models and Data Analytics in Software Engineering10.1145/3663533.3664040(56-65)Online publication date: 10-Jul-2024
    • Show More Cited By

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    ESEC/FSE 2020: Proceedings of the 28th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering
    November 2020
    1703 pages
    ISBN:9781450370431
    DOI:10.1145/3368089
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 08 November 2020

    Permissions

    Request permissions for this article.

    Check for updates

    Badges

    Author Tags

    1. Bias Mitigation
    2. Fairness Metrics
    3. Software Fairness

    Qualifiers

    • Research-article

    Conference

    ESEC/FSE '20
    Sponsor:

    Acceptance Rates

    Overall Acceptance Rate 112 of 543 submissions, 21%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)204
    • Downloads (Last 6 weeks)18
    Reflects downloads up to 14 Aug 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Mitigating Biases in Training Data: Technical and Legal Challenges for Sub-Saharan AfricaInternational Journal of Applied Research in Business and Management10.51137/ijarbm.2024.5.1.105:1(209-224)Online publication date: 24-May-2024
    • (2024)Assessing and Mitigating Bias in Artificial Intelligence: A ReviewRecent Advances in Computer Science and Communications10.2174/266625581666623052311442517:1Online publication date: Jan-2024
    • (2024)Predicting Fairness of ML Software ConfigurationsProceedings of the 20th International Conference on Predictive Models and Data Analytics in Software Engineering10.1145/3663533.3664040(56-65)Online publication date: 10-Jul-2024
    • (2024)MirrorFair: Fixing Fairness Bugs in Machine Learning Software via Counterfactual PredictionsProceedings of the ACM on Software Engineering10.1145/36608011:FSE(2121-2143)Online publication date: 12-Jul-2024
    • (2024)Fairness Testing: A Comprehensive Survey and Analysis of TrendsACM Transactions on Software Engineering and Methodology10.1145/365215533:5(1-59)Online publication date: 4-Jun-2024
    • (2024)How do Hugging Face Models Document Datasets, Bias, and Licenses? An Empirical StudyProceedings of the 32nd IEEE/ACM International Conference on Program Comprehension10.1145/3643916.3644412(370-381)Online publication date: 15-Apr-2024
    • (2024)A Post-training Framework for Improving the Performance of Deep Learning Models via Model TransformationACM Transactions on Software Engineering and Methodology10.1145/363001133:3(1-41)Online publication date: 15-Mar-2024
    • (2024)Fairness Improvement with Multiple Protected Attributes: How Far Are We?Proceedings of the IEEE/ACM 46th International Conference on Software Engineering10.1145/3597503.3639083(1-13)Online publication date: 20-May-2024
    • (2024)An Empirical Study on Correlations Between Deep Neural Network Fairness and Neuron Coverage CriteriaIEEE Transactions on Software Engineering10.1109/TSE.2023.334900150:3(391-412)Online publication date: Mar-2024
    • (2024)To be forgotten or to be fair: unveiling fairness implications of machine unlearning methodsAI and Ethics10.1007/s43681-023-00398-y4:1(83-93)Online publication date: 3-Jan-2024
    • Show More Cited By

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media