Pelegrina, G.D., Siraj, S. orcid.org/0000-0002-7962-9930, Duarte, L.T. et al. (1 more author) (2024) Explaining contributions of features towards unfairness in classifiers: A novel threshold-dependent Shapley value-based approach. Engineering Applications of Artificial Intelligence, 138 (Part B). 109427. ISSN 0952-1976
Abstract
A number of approaches has been proposed to investigate and mitigate unfairness in machine learning algorithms. However, as the definition and understanding of fairness may vary in different situations, the study of ethical disparities remains an open area of research. Besides the importance of analyzing ethical disparities, explainability in machine learning is also a relevant issue in Trustworthy Artificial Intelligence. Usually, both fairness and explainability analysis are based on a fixed decision threshold, which differentiates the positive cases from the negative ones according to the predicted probabilities. In this paper, we investigate how changes in this threshold can impact the fairness of predictions between protected and other groups and how features contribute towards such a measure. We propose a novel Shapley value-based approach as a tool to investigate how changes in the threshold values change the contribution of each feature towards unfairness. This gives us an ability to evaluate how fairness measures vary for different threshold values and which features have the higher (or lower) impact on creating ethical disparities. We demonstrate this using three different case studies that are carefully chosen to highlight different unfairness scenarios and features contributions. We also applied our proposal as a feature selection strategy, which contributed to decrease unfair results substantially.
Metadata
Item Type: | Article |
---|---|
Authors/Creators: |
|
Copyright, Publisher and Additional Information: | © 2024, Elsevier. This manuscript version is made available under the CC-BY-NC-ND 4.0 license http://creativecommons.org/licenses/by-nc-nd/4.0/. This is an author produced version of an article published in Engineering Applications of Artificial Intelligence. Uploaded in accordance with the publisher's self-archiving policy. |
Keywords: | Interpretable machine learning; Shapley value; Fairness; Feature contribution |
Dates: |
|
Institution: | The University of Leeds |
Academic Units: | The University of Leeds > Faculty of Business (Leeds) > Analytics, Technology & Ops Department |
Depositing User: | Symplectic Publications |
Date Deposited: | 07 Oct 2024 15:54 |
Last Modified: | 18 Oct 2024 14:59 |
Published Version: | https://www.sciencedirect.com/science/article/pii/... |
Status: | Published |
Publisher: | Elsevier |
Identification Number: | 10.1016/j.engappai.2024.109427 |
Open Archives Initiative ID (OAI ID): | oai:eprints.whiterose.ac.uk:217962 |
Download
