Back to All Events

A philosophical theory of AI explanations presentation, UC Berkeley

A philosophical theory of AI explanations

Atoosa Kasirzadeh

Public online presentation to an academic audience, 11am 6 July 2020 [Pacific Time; 4am 7 July AEST]

The social and ethical implications of prediction-based decision systems in sensitive contexts have generated lively debates among multiple stakeholders in- cluding computer scientists, ethicists, social scientists, policy makers, and end users. Yet, the lack of a common language and a multi-dimensional framework for an appropriate bridging of the technical, ethical, and legal aspects of the debate prevents the discussion to be as effective as it can be. Drawing on philosophy, this paper offers a multi-faceted unifying theory for the varieties of data and non-data analytical explanations as to why a prediction-based decision is obtained. The theory identifies the existence and significance of dependencies between different kinds of AI explanations as well as the role of normative and pragmatic values in making sense of these explanations. This framework lays the groundwork for establishing the relevant connection between technical, moral, and legal aspects of artificially-intelligent decision making.

This paper has been presented at the Interpretable Machine Learning workshop, Simons Institute for the theory of computing, University of California Berkeley.