Ameliorating Algorithmic Bias, or Why Explainable AI Needs Feminist Philosophy

Feminist Philosophy Quarterly 8 (3) (2022)
  Copy   BIBTEX

Abstract

Artificial intelligence (AI) systems are increasingly adopted to make decisions in domains such as business, education, health care, and criminal justice. However, such algorithmic decision systems can have prevalent biases against marginalized social groups and undermine social justice. Explainable artificial intelligence (XAI) is a recent development aiming to make an AI system’s decision processes less opaque and to expose its problematic biases. This paper argues against technical XAI, according to which the detection and interpretation of algorithmic bias can be handled more or less independently by technical experts who specialize in XAI methods. Drawing on resources from feminist epistemology, we show why technical XAI is mistaken. Specifically, we demonstrate that the proper detection of algorithmic bias requires relevant interpretive resources, which can only be made available, in practice, by actively involving a diverse group of stakeholders. Finally, we suggest how feminist theories can help shape integrated XAI: an inclusive social-epistemic process that facilitates the amelioration of algorithmic bias.

Author Profiles

Linus Huang
Hong Kong University of Science and Technology
Hsiang-Yun Chen
Academia Sinica
Ying-Tung Lin
National Yang Ming Chiao Tung University
1 more

Analytics

Added to PP
2022-10-09

Downloads
1,084 (#16,184)

6 months
285 (#6,129)

Historical graph of downloads since first upload
This graph includes both downloads from PhilArchive and clicks on external links on PhilPapers.
How can I increase my downloads?