Reasons, Values, Stakeholders: A Philosophical Framework for Explainable Artificial Intelligence Atoosa Kasirzadeh
[email protected] University of Toronto (Toronto, Canada) & Australian National University (Canberra, Australia) Abstract The societal and ethical implications of the use of opaque artificial intelligence systems for consequential decisions, such as welfare allocation and criminal justice, have generated a lively debate among multiple stakeholder groups, in- cluding computer scientists, ethicists, social scientists, policy makers, and end users. However, the lack of a common language or a multi-dimensional frame- work to appropriately bridge the technical, epistemic, and normative aspects of this debate prevents the discussion from being as productive as it could be. Drawing on the philosophical literature on the nature and value of explana- tions, this paper offers a multi-faceted framework that brings more conceptual precision to the present debate by (1) identifying the types of explanations that are most pertinent to artificial intelligence predictions, (2) recognizing the rel- evance and importance of social and ethical values for the evaluation of these explanations, and (3) demonstrating the importance of these explanations for incorporating a diversified approach to improving the design of truthful al- arXiv:2103.00752v1 [cs.CY] 1 Mar 2021 gorithmic ecosystems. The proposed philosophical framework thus lays the groundwork for establishing a pertinent connection between the technical and ethical aspects of artificial intelligence systems. Keywords: Ethical Algorithm, Explainable AI, Philosophy of Explanation, Ethics of Artificial Intelligence Preprint submitted to Elsevier 1. Preliminaries Governments and private actors are seeking to leverage recent develop- ments in artificial intelligence (AI) and machine learning (ML) algorithms for use in high-stakes decision making.