Hostname: page-component-cd9895bd7-jkksz Total loading time: 0 Render date: 2024-12-23T06:19:43.294Z Has data issue: false hasContentIssue false

Explanation and the Right to Explanation

Published online by Cambridge University Press:  29 August 2023

ELANOR TAYLOR*
Affiliation:
JOHNS HOPKINS UNIVERSITY [email protected]
Rights & Permissions [Opens in a new window]

Abstract

In response to widespread use of automated decision-making technology, some have considered a right to explanation. In this article, I draw on insights from philosophical work on explanation to present a series of challenges to this idea, showing that the normative motivations for access to such explanations ask for something difficult, if not impossible, to extract from automated systems. I consider an alternative, outcomes-focused approach to the normative evaluation of automated decision making and recommend it as a way to pursue the goods originally associated with explainability.

Type
Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
Copyright © The Author(s), 2023. Published by Cambridge University Press on behalf of the American Philosophical Association

1. Automated Decision Making and the Right to Explanation

Increasingly, an enormous number of decisions about our lives are made by automated systems, including decisions about mortgages, credit lines, health care, entertainment recommendations, policing, sentencing, and personal matchmaking. Automated decision making is more readily accessible than ever before, but its widespread use has raised public concern. For example, automated decision making guides central parts of the criminal justice system in the United States, including parole decisions, and many have argued that this practice entrenches established patterns of judicial injustice (Angwin et al. Reference Angwin, Larson, Mattu and Kirchner2016; Hao Reference Hao2019; Metz and Satariano Reference Metz and Adam Satariano2020). In 2020, automated grade calculations replaced in-person final examinations for many UK school students during the COVID-19 pandemic. The results met with public outcry as calculations took expected performance from schools into account, resulting in lower grades for students from underperforming schools (Ascher-Shapiro Reference Ascher-Shapiro2020; Broussard Reference Broussard2020; Hern Reference Hern2020; Zimmerman Reference Zimmerman2020). Even apparently benign applications of automated decision making have come under scrutiny. For example, the algorithm that governs recommendations on YouTube has been implicated in an increase in political polarization and the spread of conspiracy theories, such as QAnon and the denial of the COVID-19 pandemic (Ovide Reference Ovide2020; Roose Reference Roose2020).

Automated decision-making systems are often opaque: the bases of their decisions are inaccessible to the average person. Opacity comes in different forms. Some systems are opaque simply in that the decision making procedure is kept secret by the organizations that create and use them. Sometimes opacity is more extreme, as in cases of algorithmic opacity where the system's decision-making procedures are unavailable even to its designers because its decision-making capacity emerges from interactions between its code and complex training data (Paudyal and Wong Reference Paudyal and Wong2018: 193). Such systems effectively train themselves to deliver results, and typically even their designers do not, and cannot, understand how they operate.

The widespread use and opacity of automated decision making is worrying and has led some to consider a right to explanation. On this line of thought, if an automated system makes a decision about us, then we have a right to an explanation of that decision. Some take the European Union's 2018 General Data Protection Regulation (known widely as the GDPR) to enshrine a right to explanation, though this is a matter of some legal controversy (Selbst and Powles Reference Selbst and Powles2017; Wachter, Mittelstadt, and Floridi Reference Wachter, Mittselstadt and Floridi2017b). According to Article 13 of the GDPR if a decision is made about a person through an automated process then that person should have access to ‘meaningful information about the logic involved’ in the decision (European Union 2018: Article 13, 2f).Footnote 1 The GDPR applies only in the European Union, but similar legislation has been pursued elsewhere, such as New York City's LL 144, which regulates automated decision making in hiring and human resources (Hunton Andrews Kurth LLP 2023). Further, some have argued that independent of explicit legal regulation there are good ethical, political, and social reasons to embrace a more transparent approach to automated decision making, a central motivation for developments in explainable artificial intelligence (often abbreviated as XAI) (Paudyal and Wong Reference Paudyal and Wong2018; Wachter, Mittelstadt, and Floridi Reference Wachter, Mittelstadt and Floridi2017a).

There has been a flurry of discussion about the challenges raised by the prospect of explainable AI and a right to explanation. For example, some argue that the GDPR does not legally guarantee a right to explanation, while others hold that explainable AI is impossible to engineer (see Wachter, Mittelstadt, and Floridi Reference Wachter, Mittelstadt and Floridi2017a). Indeed, representatives of Google Research referred to machine learning as ‘alchemy’ in a speech given in 2017 (see Rahimi and Recht Reference Rahimi and Recht2017). However, at the heart of this issue is a philosophical question: what is the right to explanation a right to? What does it take to explain an automated decision? There is little consensus on this matter. Article 13 of the GDPR uses the language of ‘meaningful information’, while some appeal to the idea that an explanation is an answer to a why-question (EU 2018: Article 13, 2f; Burrel Reference Burrel2016). But these answers are too broad to be of much use. For example, the fact that a decision was made by an algorithmically opaque system is meaningful information and offers an answer (of a sort) to the question of why the decision was made. But this is clearly an inadequate explanation for most purposes, and so more insight is necessary.

The normative motivations for access to explanations offer a valuable source of insight. By reflecting on why access to explanation might be a good thing, we can clarify what an explanation must be such that it can satisfy those motivations. However, the normative motivations for explainable AI appear in many cases to demand explanations that are difficult, if not impossible, to obtain from automated decision-making systems, including explanations that give information about reasons. This threatens the claim that there is a right to explanation. Opacity undermines the kind of evaluation that involves seeking reasons and explanations for decisions, and I discuss some broader implications of this insight for normative exchange and deliberation in opaque contexts. I sketch an alternative approach to automated decision making focused on outcomes rather than explanations, as a way to pursue the goods originally associated with explainable AI. Many organizations are currently moving away from pursuing explainability toward an emphasis on outcomes, and this discussion can be understood as offering a philosophical basis for such strategies. For example, see Microsoft's guide to impact assessment (Microsoft 2022); NYC L44, which focuses in part on outcome-focused audits (Hunton Andrews Kurth LLP 2023); and the European Union's proposed Artificial Intelligence Act (EU 2021).

2. Normative Motivations for Explainable AI

Public conversation about explainable AI is dominated by the language of rights, in particular the phrase right to explanation. But showing that there is a right to explanation is different from showing that there is a normative case for access to explanations. The case for a right to φ must show that the normative case for φ is strong enough to justify the burden of providing φ. If that burden is considerable, the case for the right may be undermined. I follow Kate Vredenburgh (Reference Vredenburgh2021) in adopting this framing of the right to explanation. In doing so I treat the right to explanation as a positive rather than a negative right and leave open questions about which more fundamental rights it may derive from. I exploit this distinction between making a normative case and establishing a right, and so I treat the case for explainable AI and the case for a right to explanation as distinct.

One central motivation is to promote transparency. This is mentioned in the GDPR recitals: ‘It should be transparent to natural persons that personal data concerning them are collected, used, consulted or otherwise processed and to what extent the personal data are or will be processed. The principle of transparency requires that any information and communication relating to the processing of those personal data be easily accessible and easy to understand, and that clear and plain language be used’ (EU GDPR, Recital 39).

Recently, some commentators have explored this motivation. For example, Seth Lazar argues that automated decision making intensifies existing political power relations and as such is subject to a publicity requirement, that it should be possible for those who authorize the power's use to determine that it is being used legitimately and with proper authority. Lazar holds that explainability is necessary to meet this publicity requirement and so that it motivates the demand for explanation (Lazar Reference Lazar2022: 8). Mario Günther and Atoosa Kasirzadeh go further, arguing that automated systems should be held to an even higher standard of transparency than human beings (Günther and Kasirzadeh Reference Günther and Kasirzadeh2022).

However, there is more to this issue beyond transparency or publicity, as there are further goods that transparency can promote. The GDPR recitals mention that access to an explanation of an automated decision will help in evaluating that decision and the decision-making procedure for fairness (EU 2018: Recital 39). This requires that the explanation provide information not only about the decision in an individual case but also about how equitably the decision-making criteria are applied across populations. For example, all things considered, applying a higher salary requirement to one candidate over another for the same credit decision is a violation of fairness. This issue is central to public outcry about automated decision-making technology, much of which appeared in response to apparently unfair applications of decision making, such as automated exam grade calculations and parole decisions.

A further good that transparency promotes is the protection of individual autonomy with respect to decisions made by automated systems. Access to an explanation can help us protect our autonomy because it permits us to understand the bases of the automated decisions and, where appropriate and possible, exert control over our circumstances to achieve different results. If my mortgage application is refused, for example, it is reasonable to expect an explanation that tells me what I can change to raise the likelihood that a future application will be accepted. It is not always possible to change the outcome of a decision, as when a health insurance request is rejected because of a preexisting condition, but my access to that information still supports my autonomy by enabling me to understand that fact and base further actions upon it, such as seeking different health insurance or voting for changes to the healthcare system. Similarly, if a decision is made through a random process, as in a lottery, then it is clear that I can take no further steps to change the outcome. Much discussion of the right to explanation has focused on the significance of autonomy (Vredenburgh Reference Vredenburgh2021; Wachter, Mittselstadt, and Russell Reference Wachter, Mittelstadt and Russell2018). Some may disagree that protecting the pursuit of certain goals, as in these cases, is properly called protecting autonomy. If so, we can adopt different language and say instead that access to explanations may be valuable in part because it facilitates goal seeking.

Access to an explanation may be valuable because it allows us to evaluate the use of decision-making systems for its capacity to generate broader social harm. Such harm, which goes beyond instances of individual unfairness or violations of individual autonomy, has been the subject of much of the public conversation about automated decision making. For example, as mentioned earlier, some commentators argue that the YouTube selection algorithm, which nudges viewers toward content similar to content they have already viewed, is implicated in online radicalization, the spread of conspiracy theories, and the undermining of public trust in mainstream news sources and scientific institutions (Ovide Reference Ovide2020; Roose Reference Roose2020). Facebook has been subject to similar scrutiny for news and content recommendations that trap users in ‘filter bubbles’ and so hamper the exchange of ideas essential to the democratic process, and facilitate the spread of misinformation and conspiracy theories (Adee Reference Adee2016; Gatehouse Reference Gatehouse2021Reference Gatehouse2022). From these examples it is evident that social harm can often be unintentional and hard to predict, but transparency about the bases of the decisions made by automated systems will allow us at least to evaluate the use of such systems for the likelihood of generating social harm. A central characteristic of social harm is that it is structural, affecting society at the level of systems and institutions, and as such is not equivalent to individual-level injustice, though it inevitably generates harm to individuals.

This list of motivations is not exhaustive, but it offers some sense of why explainable AI might be a good thing. Access to explanations of automated decisions will promote transparency, which is valuable in turn because it allows us to evaluate those decisions for fairness, protect our autonomy with respect to the outcomes of the decisions, and evaluate the use of the technology for its potential to cause social harm. These goals are general, in that widespread access to explanations will promote them in general, but not every attempt to explain a decision made by an automated system must in itself meet all of these demands. For instance, an explanation that permits an individual to protect their autonomy, such as an explanation that identifies the factors relevant to a credit decision, may not be rich enough to facilitate evaluation of the decision-making system for its capacity to generate social harm. This does not undermine the value of the explanation, as it still permits the individual to protect their autonomy and thereby satisfies one of the normative motivations for explainable AI. More broadly, however, even if each attempt at explanation need only meet the needs of those seeking it, in general people should have access to explanations that satisfy all of the motivations described here.

3. Normatively Satisfying Explanations

There is little consensus about the nature of explanation. Philosophers disagree on almost every aspect of the subject, including whether or not explanation must be causal, whether or not it must be factive, and what kind of relation it is, if it is a relation at all (Achinstein Reference Achinstein1983; Craver Reference Craver, Kaiser, Scholz and Plenge2014; Lange Reference Lange2016; Lewis Reference Lewis1986; Reutlinger Reference Reutlinger2017; Ruben Reference Ruben1990; Salmon Reference Salmon1989; Skow Reference Skow2014; Van Fraassen Reference Van Fraassen1980). Indeed, some have argued that it is misguided to attempt to give a general account of explanation (Díez, Khalifa, and Leuridan Reference Díez, Khalifa and Leuridan2013; Nickel Reference Nickel2010). If we turn to philosophy for insight into the right to explanation, then this level of disagreement might seem frustrating. But thankfully we do not need a full theory of explanation to illuminate this issue. Instead, I focus on two claims about explanation that, while not universally endorsed, are fairly uncontroversial. The first is that explanation has a contextual aspect, and the second is that we can at least partially explain events by identifying difference makers.

The idea that explanation has a contextual aspect is simple to motivate. Imagine that I ask you to explain a particular event: the theft of a car. You would naturally consider features of our context before offering the explanation. If we are engineers and are interested in the car theft as a physical event, then you would be likely to offer a physical, but not micro-physical, explanation of the theft, giving information about the forces exerted upon the car, how the glass was broken, and so on. Alternatively, if I am your friend and am asking what possessed you to steal this car, then a different sort of explanation will be appropriate. In this situation, I want an explanation that tells me about your motivations and what inspired you to steal the car. This contextual aspect of explanation is commonplace, and some accounts rest more significance on it than others. For example, some hold that what counts as an explanation is determined at least partly by context (Achinstein Reference Achinstein1983; Van Fraassen Reference Van Fraassen1980). On other views, explanation itself is context-independent, but context affects how successfully the explanation may be received or understood (Lewis Reference Lewis1986). It is not necessary to decide on this issue here; it is enough to acknowledge that explanation has some contextual aspect, as follows:

Context: Successful explanations are contextually appropriate.

The second claim, that we can explain events by identifying difference makers, is more controversial though still widely endorsed (Woodward Reference Woodward2003). The rough idea is that we can explain an event by identifying the factors that made a difference to the occurrence of the event, picked out by the counterfactual test: had those factors not been in place, the event would not have occurred. For example, if it is the case that had I not thrown the rock, the window would not have broken, then the information that I threw the rock will (at least in part) explain the event of the window breaking. The difference-making approach to explanation has this idea at its heart and captures a number of intuitively plausible claims about explanation such as that explanation has a counterfactual aspect and that there is a close connection between explanation and intervention. The full difference-making approach to explanation is detailed and includes a formalism for representing causal, and hence explanatory, relations in structural equation models (Pearl Reference Pearl2000; Woodward Reference Woodward2003). However, most of that detail is irrelevant to this conversation. All we need here is the rough idea that we can (at least partially) explain events by identifying difference makers:

Difference: We can (at least partially) explain an event by identifying factors that made a difference to that event.

There are good reasons to adopt difference as a frame for this issue. First, many normative motivations for access to explanations of automated decisions appear to ask for information about difference makers, such as the case in which the information that a lack of long-term loan history made a difference to a credit denial satisfied the requirements of transparency and of protecting autonomy. Second, the output of an automated decision-making system is an event. Despite the lack of consensus about explanation in general, difference making is the most commonly accepted approach to explanation of events in philosophy of science, which makes this a reasonable place to start (Reutlinger Reference Reutlinger2017). Third, although the literature on what explanation means in right to explanation is in a developmental stage, the proposal that explanations must identify difference makers has been the subject of some discussion (Wachter, Mittselstadt, and Russell Reference Wachter, Mittelstadt and Russell2018). Fourth, the proposal as it stands is not that all explanation, or all causal explanation, proceeds by identifying difference makers but is, instead, the weaker claim that one can, at least in part, explain events by identifying difference makers. Accordingly, one need not endorse a full difference-making approach to explanation to endorse this proposal.

Some may worry that a difference-making approach is inapplicable to decisions made by automated systems because such explanations must be noncausal. I remain neutral on whether explanations of automated decisions are causal or noncausal, but note that although difference making was developed as a model of causal explanation, it has been extended to noncausal explanation (Reutlinger Reference Reutlinger, Reutlinger and Saatsi2018; Schaffer Reference Schaffer2016; Wilson Reference Wilson2018; for challenges, see Kasirzadeh Reference Kasirzadeh2023). Others may worry that difference-making explanations are impossible to extract from many automated decision-making systems, particularly black-box systems (Grimsley Reference GrimsleyMS). I return to this issue below, where I suggest that if they are impossible to extract, then this reveals a tension at the heart of the explainable AI project.

Putting context and difference together gives a clearer picture of the target of the normative motivations for explainable AI. To explain a decision made by an automated system we should identify difference makers—factors such that, had they not been in place, the decision would have gone differently—in a contextually appropriate way. The standards for contextual appropriateness are found in the normative motivations discussed above. These explanations must promote transparency and must enable us to evaluate the decisions for fairness, permit the inquirer to protect their autonomy, or permit evaluation for the capacity to generate broader social harm, depending on the goals of the inquirer:

Target: Normatively satisfactory explanations of automated decisions identify factors that made a difference to the decision in a manner that promotes transparency, permits evaluation of the decision and decision-making system for fairness or capacity to generate social harm, or enables the inquirer to protect their autonomy.

As before, the normative motivations for explainable AI do not require that every explanation should meet each of the individual criteria. Depending on the interests of the inquirer, in some cases an explanation that permits evaluation for fairness but is not rich enough to support evaluation for social harm will be sufficient. But overall, the normative case for explainable AI supports widespread access to explanations that meet all of these criteria.

What does it take for an explanation to meet these standards? One obvious criterion is that to play any of these roles an explanation must be understandable to its audience. This follows straightforwardly from the transparency requirement. A second criterion, also obvious, is that the explanation must give information about factors that made a difference to the decision. It may be tempting to stop here, holding that all that is required is merely to identify some factor that made a difference to the decision and to do so in an understandable way. But this is the point at which the contextual aspect of explanation becomes salient. In many contexts merely identifying some difference maker will not satisfy the normative motivations for explainable AI. Instead, target specifies that we need information about difference makers that is particularly suited to the inquirer's goals.

For insight, consider the range of reasons why you may want an explanation of an automated decision. In some cases, you may simply want to protect your own interests. When my credit application is denied, for example, I want to know what action I can take to improve the likelihood of a positive decision next time. For this I need information about a difference maker pitched at a level that gives me control over future outcomes, where possible. The information that the system is designed to avoid credit risk and regarded my application as risky fails to provide this, while identifying a particular aspect of my case, such as my lack of loan history, is more successful. Sometimes, however, our needs are more complex, as when we are evaluating decisions and decision-making systems for their capacity to generate social harm. To perform these kinds of tasks we sometimes need information about a difference maker pitched at a very particular level of grain.

Consider some examples. Imagine that my application for a mortgage is turned down and I am told that my lack of long-term loan history made a difference to the decision. This is a difference maker, so if merely identifying a difference maker is sufficient for explanation, then the task is complete. But I might reasonably wonder why long-term loan history made a difference. Perhaps it is important because it is independently motivated as an indicator of potential for responsible credit use, say. But alternatively, perhaps long-term loan history is important simply because it is correlated with affluence. These differences are significant when it comes to evaluating the decision making for fairness and for social harm, as the latter basis for credit decisions is arguably less fair than the former and has potential to entrench existing inequality. This example indicates that sometimes we need to identify difference makers at different levels of grain, where the appropriate level of grain is set by what the inquirer wants from the explanation.

Consider an alternative case. Imagine that I want to know why YouTube selected a particular video outlining a conspiracy theory for me to watch next. A difference-making explanation might tell me that this video is similar to those I have watched before, and so the difference maker is just that this content is similar to content I have selected and watched in the past. The algorithm is designed to keep viewers watching, and similarity is a factor that will keep viewers watching. That seems fairly benign, though it might lead to some overconsumption. But if we keep asking questions, a new picture may emerge. Imagine that the next video was selected because the content is similar. Furthermore, the content I watched previously is about conspiracy theories, and people who watch that kind of content are more likely than others to binge-watch YouTube videos. This is a finer-grained instance of the coarser difference maker similarity of content. These reasons are more exploitative, and this politically relevant information is not available from the more general claim that this video is similar to content I watched before. In each case similarity of content is a difference maker. But the reason why similarity of content is selected as a difference maker is different in each case. In the first, it was selected simply because a viewer is more likely to watch similar content. In the second, it was selected because the viewer of a particular, politically salient type of content is more likely to watch similar content. If we are interested in evaluating decisions and decision-making systems for fairness, autonomy, and social harm, such differences make a difference.

As these examples illustrate, the explanations needed to satisfy the normative motivations for explainable AI must get to a particular level of detail, set by the interests of the inquirer. And in some cases, the correct level of detail is for the explanation to identify a reason. A reason explanation is an explanation of an agent's action that gives information about the agent's own motivations for their action. If I explain why I donated to a particular charity by pointing out that I took it to be the morally right thing to do, then I am giving you a reason explanation for my action. And in many cases, to satisfy the normative case for explainable AI we need to get to the reasons behind a decision, rather than merely to identify any difference maker. (On a standard view of reasons as causes, reasons are difference makers. However, the view of reasons as causes has been challenged. See the discussions in Queloz [Reference Queloz2018] and Sehon [Reference Sehon2000].) For instance, in the mortgage application case, we are interested in the reason why the loan history made a difference, and in the YouTube case we want to know the reason why similarity of content was selected as a difference maker. In each case, we want the justification and motivation for that factor making a difference, not just that it did happen to make a difference. Without that information, we cannot meet our normative goals.

4. Revisiting the Right to Explanation

The normative case for explainable AI is a case for widespread access to contextually appropriate difference-making explanations of the decisions made by automated systems. Contextual appropriateness varies from case to case, depending on the inquirer's interests, and the factors that determine contextual appropriateness include the capacity of the explanation to promote transparency, evaluation for fairness, individual autonomy, and evaluation for social harm. To meet these standards in some cases the explanation must provide reasons for the decisions and reasons why certain difference makers were selected. I now return to the implications of this proposal for the claim that there is a right to explanation.

As mentioned above, establishing a right to φ is different from making a normative case for φ, as the former must show that the normative case for φ outweighs or justifies the burden of providing φ. This exploration of normative motivations for explainable AI and the kinds of explanations that meet them reveals serious challenges to the claim that there is a right to explanation.

The first problem is simple: it is not clear that difference-making explanations of the decisions made by automated systems are reliably available. In simple cases, as when systems operate as decision trees, or operate with a small number of predictable parameters, there may be no in-principle barrier to identifying difference makers. But in more complex cases opacity may arise from a range of sources, such as the complexity and high dimensionality of the decision-making systems and the inscrutability of the mathematical methods the systems use to develop decision-making procedures (Lazar Reference Lazar2022: 2–4). Each source of opacity generates different normative and epistemic challenges, and each presents barriers to widespread access to difference-making explanations. The nature and depth of the opacity of automated decision-making systems is a lively topic in computer science, so it is an open empirical question and a matter of some controversy whether viable difference-making explanations of the decisions made by automated systems are or will soon be available (Grimsley Reference GrimsleyMS; Hamon et al. Reference Hamon, Junklewitz, Malgieri, De Hert, Beslay and Sanchez2021). But insofar as features of the systems themselves or of the organizations that use them preclude access to difference-making explanations, the normative motivations for explainable AI cannot be met.

In response, one might consider an idealized standard for explainability along the following lines: that some explanation for the decision must be available in principle, even if no actual person can currently access it. This standard would avoid problems of intractability. However, the normative motivations for access to explanations are based on the idea that it is a social good for actual people to have access to these explanations for their own purposes, not that an idealized agent could in principle access the explanation. Even though an idealized standard for explanation may avoid this challenge, it also fails to meet the normative motivations for explainable AI.

Second, in some cases the normative motivations for explainable AI are best satisfied by a reason explanation, but it is not clear that reason explanations of automated decisions are available, or that the prospect of a reason explanation of a decision made by an automated system is even coherent. This is in part because of the engineering problems involved in identifying difference makers in opaque systems, as above. But it is also about the very idea of asking for reasons from an automated system. This is a strange demand, and it leaves us searching for candidate sources of reasons in the decisions of the engineer, the organization conducting the decision making, or even society. It is natural to find such a request jarring, as reason explanation is the kind of explanation that we use to explain specifically human affairs. Mechanistic and other kinds of causal explanation may be satisfactory when it comes to events in natural and artificial worlds, but when we want to understand human action and decision, we turn to reason explanation. These are the explanations that we ask for when we want to understand one another, and to hold one another accountable. Insofar as the normative case for explainable AI asks for reason explanations, it asks us to explain how automated decisions are made in a specifically human way.

Third, putting aside these concerns about opacity and the apparent absence of reasons, the resulting picture of the right to explanation is radical. The case for access to explanations of automated decision-making requires a significant level of information about corporate and institutional decision making. If we take that case seriously, then it seems that we should demand a level of scrutiny of institutions using automated decision making far beyond what was ever applied to actuaries, bank managers, and the like, regardless of the automation or otherwise of their decisions. To countenance a right to explanation is therefore to countenance an overhaul of industries and services, including banking, computing, social media, education, medicine, government, and policing. If the normative case for access to such explanations is significant then this overhaul is required, and the fact that it may be challenging to implement is not in itself an objection. But it does put an extra burden on the positive case for a right to explanation, which must be very strong to overcome these practical challenges.

Each consideration generates a problem for the case for a right to explanation. The first two may show that a right to explanation is literally impossible because in many cases the case for explainable AI appears to ask for something that cannot be provided. Putting aside those worries, we are left facing the third concern: providing such explanations demands an unprecedented level of transparency from any industry or organization using automated decision making. This presents a significant burden, making it harder to show that the normative case for access to such explanations outweighs the cost of providing them. It may be, of course, that the normative case for explainable AI meets this standard or that a high level of transparency is motivated by other considerations. But overall, these are serious challenges to the claim that there is a right to explanation.

5. Responses and Reflections: Opacity and the Shift to Outcomes

One response to these challenges is to return to my reasoning through the normative motivations for explainable AI and reject some of that framing, arguing that an alternative picture offers a more promising outlook for a right to explanation. Perhaps a non-difference-making form of explanation is available from algorithmically opaque systems, or perhaps information about reasons is never required to meet the normative motivations for explainable AI. I do not consider such strategies here as I suspect that most explanations that can meet the standard of target will generate similar problems, but there is logical space to attempt to show otherwise. A more pessimistic alternative is to abandon the normative evaluation of automated decision making and with it the idea that there is a right to explanation. On this line of thought, the normative motivations for explainable AI ask for something that can never be delivered, so attempts to normatively evaluate automated decision making are incoherent.

An in-between strategy involves seeking explanations of some automated decisions but not others. For example, Alex London (Reference London2019) points out that many clinical recommendations with a strong evidence base are not well explained, even without the influence of automated decision making. Because opacity is a standard feature of medical decision making, London argues, it is misguided to demand explanations of all automated clinical decisions. To take this line of reasoning is not to give up on normative evaluation of automated decision making, but instead to acknowledge that some contexts, such as medicine, have epistemic features that can preclude explainability, which undermines a general demand for explanation.

While each of these strategies has its merits, there are good reasons for moving away from explainability. London (Reference London2019) is correct that it is misguided to seek explanations of decision making in opaque contexts and the reasons why reflect deeper philosophical considerations about what it is to explain a decision. Seeking explanations of decisions in order to evaluate them normatively is to engage with those decisions in a broadly reason-giving way. In opaque contexts, however, the standard bases for reason-giving evaluation are often unavailable. A viable alternative is to switch the focus of evaluation to the outcomes of decisions, rather than the reasons for them. In the case of automated decision making, this entails a switch from evaluating decisions by explaining them to evaluating decisions by examining their outcomes. Given that many organizations and institutions are abandoning explainability in favor of the evaluation of outcomes for expediency, this discussion can be understood as an attempt to offer a more robust normative basis for this shift and to recommend it as an alternative way to pursue the goods originally associated with explainable AI.

To seek an explanation of a decision, especially an explanation that articulates reasons, is to treat the source of the decision as an entity like the agent seeking the explanation: as a rational being. In asking for these explanations we seek reasons for, justifications of, and motivations for the decisions, and thereby focus our evaluation on the rationale for the decision, rather than its outcomes. As such, this is a traditionally deontological mode of normative evaluation (though some approaches to deontology are more reason-focused than others). However, in the context of automated decision making, the standard sources of such evaluation, like sentience or rationality, are not in play. The systems we seek these explanations from do not have the features typically taken to ground reason-giving interactions and evaluations. In the same way that it is wrongheaded to blame a tree for dropping a branch onto my car or to ask the cloud to explain why it rained on me, it is wrongheaded to ask an automated system to explain its actions—and especially to give an account of its reasons.

Automated decision making systems may have these features by proxy. While an algorithm in itself may not have reasons, its decision making may reflect the reasons of the company or institution using it. However, as we have seen, when automated decision making is used, often those reasons are also kept opaque. The automated system may display its own kind of opacity, and the company or institution using the system often imposes further layers of opacity. Even if we take the automated system to have reasons by proxy, we typically still do not have access to those reasons.

The normative case for explainable AI is based on the idea that access to explanations of automated decisions will promote transparency, and will help us to ensure that we are treated fairly, can protect our autonomy, and can protect ourselves from broader social harms. As shown above, considerations about opacity, intractability, and the absence of reasons threaten to undermine this focus on explanation and warrant turning toward a more tractable place: the outcomes of automated decision making. This shift of focus is in the spirit of a political philosophy that recognizes that harms are easily generated by well-intentioned systems and takes unjust outcomes as sufficient to warrant political response without requiring evidence of unjust intent. (Recent work on algorithmic bias adopts a similar strategy, such as Johnson (Reference Johnson2021). If automated decision making consistently, say, privileges the rich, promotes racist judicial policies, hampers social mobility, and breaks down information channels essential to democratic communication, then, on this line of thought, that is enough to be working with, without having to also scrutinize an opaque system for the source of the decisions that generated such harms.

Consider a comparison with similar reasoning in a very different context: feminist theories of misogyny. Traditionally misogyny was understood in psychological terms, such that the misogynist, by definition, hates women. In contrast, Kate Manne recommends understanding misogyny as a social phenomenon. On this approach, misogyny is something that women experience and face, not something that misogynists think (Manne Reference Manne2017: 59–67). In adopting this view, Manne recommends a shift from a mode of evaluation and engagement that focuses on the misogynist's reasons for their actions to a mode of evaluation that focuses on the outcomes of those actions in women's lives.

The details of Manne's view and its place in broader discourse about misogyny are not relevant to this discussion. But the structure of this shift—from regarding the misogynist as a source of reasons to merely a source of outcomes—mirrors the shift I recommend in the case of automated decision making. This is a move from considering the reasons why automated decisions are made to considering what the decisions do. This parallel is not coincidental, because both contexts display forms of opacity that undermine reason-giving evaluation. In political discourse under gender inequality (and similar conditions such as racism), the norms of reason giving, explaining, justifying, and so on can often break down. The person who shifts their focus from reasons to outcomes in such situations treats the source of harm as opaque.

Structural harm such as misogyny is not required to justify this change in mode of evaluation, because other sources of opacity suffice. Different sorts and sources of opacity may generate different results here. For example, consider the opacity of decision making in a company that keeps its procedures secret. One might reasonably adopt a consequences-forward approach to their decisions on the basis of the belief that, as long as the bases of their decisions are unavailable, it makes no sense to recognize them as a proper subject of reason-giving evaluation. If the company were to change policy in favor of a more transparent approach, one could then reasonably ask for explanations and information about their reasons and motivations. Alternatively, the kind of opacity evident in algorithmically opaque systems makes the search for explanations less tractable, and the mere fact that the decision maker is an automated system undermines treating it as an agent with reasons in the way that full deontological evaluation (of this reason-focused sort) demands.

These considerations indicate that in opaque contexts, where the bases of a decision maker's decisions and their decision-making procedures are unavailable for scrutiny, it is reasonable, and perhaps necessary, to adopt an outcomes-based model of normative evaluation. Accordingly, for those who find the normative case for explainable AI compelling and are worried about opacity and intractability, a shift in focus to the outcomes of automated decision making offers a way to pursue the goods originally associated with access to explanations.

Precisely how well a focus on outcomes can serve the various normative motivations for explainable AI is unclear, and a developed proposal will have to show how information about outcomes can promote fairness, protection of individual autonomy, and so on. But overall, abandoning the focus on explanation as the core of normative concerns about automated decision making and shifting instead toward evaluating outcomes offers a way to pursue the benefits originally associated with access to explanations, without the associated problems of intractability and implementation.

Footnotes

Thanks to Ramón Alvarado, Michael Brent, Nina Emery, James Goodrich, Neil Manson, Elizabeth Miller, Dean Moyar, Meghan Page, Kyle Rawlins, Erica Shumener, John Symons, and Michael Titelbaum for helpful discussion and feedback. Particular thanks to Tal Linzen for an introduction to the topic and for discussion and comments on an early draft. Thanks to audiences at ‘Computational Methods and the Future of Science’ at the University of Kansas, a philosophy department colloquium at the University of Mississippi and the 2021 meeting of the Philosophy of Science Association. Thanks to two online groups for workshopping the article: the WiM Research Network, and the Pre-Tenure Women's online working group. Finally, thanks to two anonymous referees for this journal for their generous feedback.

1 The GDPR, like other EU laws, is organized into articles that constitute the legal requirements and recitals that offer contextualization and interpretation of the articles.

References

Achinstein, Peter. (1983) The Nature of Explanation. New York: Oxford University Press.Google Scholar
Adee, Sally. (2016) ‘How Can Facebook and Its Users Burst the “Filter Bubble?”’ New Scientist, November 18, 2016. https://www.newscientist.com/article/2113246-how-can-facebook-and-its-users-burst-the-filter-bubble/.Google Scholar
Angwin, Julia, Larson, Jeff, Mattu, Surya, and Kirchner, Lauren. (2016) ‘Machine Bias’. Propublica, May 23, 2016. https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing.Google Scholar
Ascher-Shapiro, Avi. (2020) ‘Exam Grading Algorithms Amid Coronavirus: What's the Row About?’ Reuters, August 12, 2020. https://www.reuters.com/article/global-tech-education-idUSL8N2FD0EI.Google Scholar
Broussard, Meredith. (2020) ‘When Algorithms Give Real Students Imaginary Grades’. New York Times, September 8, 2020. https://www.nytimes.com/2020/09/08/opinion/international-baccalaureate-algorithm-grades.html.Google Scholar
Burrel, J. (2016) ‘How The Machine “Thinks”: Understanding Opacity in Machine Learning Algorithms’. Big Data & Society, 3. https://doi.org/10.1177/205395171562251.Google Scholar
Craver, Carl F. (2014) ‘The Ontic Account of Scientific Explanation’. In Kaiser, Marie I., Scholz, Oliver R., Plenge, Daniel, and Andreas Hüttemann (eds.), Explanation in the Special Sciences: The Case of Biology and History (Dordrecht: Springer), 2752.10.1007/978-94-007-7563-3_2CrossRefGoogle Scholar
Díez, José, Khalifa, Kareem, Leuridan, Bert. (2013) ‘General Theories of Explanation: Buyer Beware’. Synthese, 190, 379–96.10.1007/s11229-011-0020-8CrossRefGoogle Scholar
European Union [EU]. (2018) General Data Protection Regulation. https://gdpr-info.eu.Google Scholar
European Union [EU]. (2021) Proposal: The Artificial Intelligence Act https://artificialintelligenceact.eu/ .Google Scholar
Gatehouse, Gabriel. (2021–2022) The Coming Storm. BBC Radio 4, Podcast. https://www.bbc.co.uk/programmes/m001324r .Google Scholar
Grimsley, Christopher. (manuscript) ‘Causal and Non-causal Explanations of Artificial Intelligence’. http://philsci-archive.pitt.edu/17359/1/ExplanationAttention.pdf.Google Scholar
Günther, Mario, and Kasirzadeh, Atoosa. (2022) ‘Algorithmic and Human Decision Making: For a Double Standard of Transparency’. AI and Society, 37, 375–81.10.1007/s00146-021-01200-5CrossRefGoogle Scholar
Hamon, Ronan, Junklewitz, Henrik, Malgieri, Gianclaudio, De Hert, Paul, Beslay, Laurent, and Sanchez, Ignacio. (2021) ‘Impossible Explanations? Beyond Explainable AI in the GDPR from a COVID-19 Use Case Scenario’. In FAccT ’21: Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency (New York: Association for Computing Machinery) 549–59. https://dl.acm.org/doi/10.1145/3442188.3445917CrossRefGoogle Scholar
Hao, Karen. (2019) ‘AI Is Sending People to Jail—and Getting It Wrong’. MIT Technology Review, January 21, 2019. https://www.technologyreview.com/2019/01/21/137783/algorithms-criminal-justice-ai/ .Google Scholar
Hern, Alex. (2020) ‘Ofqual's A-Level Algorithm: Why Did It Fail to Make the Grade?’ The Guardian, August 21, 2020. https://www.theguardian.com/education/2020/aug/21/ofqual-exams-algorithm-why-did-it-fail-make-grade-a-levels .Google Scholar
Hunton Andrews Kurth LLP (2023) ‘NYC DCWP Proposes Rules to Implement New Law Governing Automated Employment Decision Tools’. National Law Review, 12. https://www.natlawreview.com/article/nyc-dcwp-proposes-rules-to-implement-new-law-governing-automated-employment-decision .Google Scholar
Johnson, Gabbrielle M. (2021) ‘Algorithmic Bias: On the Implicit Biases of Social Technology’. Synthese, 198, 9941–61.10.1007/s11229-020-02696-yCrossRefGoogle Scholar
Kasirzadeh, Atoosa. (2023) ‘Counter Countermathematical Explanations’. Erkenntnis 88, 2537–60.10.1007/s10670-021-00466-xCrossRefGoogle Scholar
Lange, Marc. (2016) Because without Cause. New York: Oxford University Press.10.1093/acprof:oso/9780190269487.001.0001CrossRefGoogle Scholar
Lazar, Seth. (2022) ‘Legitimacy, Authority, and the Political Value of Explanations’. Paper presented at Oxford Studies in Political Philosophy workshop, Tucson AZ, October 2022. https://arxiv.org/ftp/arxiv/papers/2208/2208.08628.pdf.Google Scholar
Lewis, David. (1986) ‘Causal Explanation’. In Philosophical Papers Volume II (New York: Oxford University Press), 214–40.Google Scholar
London, Alex John. (2019) ‘Artificial Intelligence and Black-Box Medical Decisions: Accuracy versus Explainability’. Hastings Center Report, 49, 1521.10.1002/hast.973CrossRefGoogle ScholarPubMed
Manne, Kate. (2017) Down Girl: The Logic of Misogyny. New York: Oxford University Press.10.1093/oso/9780190604981.001.0001CrossRefGoogle Scholar
Metz, Cade, and Adam Satariano, A. (2020) ‘An Algorithm That Grants Freedom, or Takes it Away’. New York Times, February 6. 2020. https://www.nytimes.com/2020/02/06/technology/predictive-algorithms-crime.html .Google Scholar
Nickel, Bernhard. (2010) ‘How General Do Theories of Explanation Need to Be?’ Noûs, 44, 305–28.10.1111/j.1468-0068.2010.00741.xCrossRefGoogle Scholar
Ovide, Shira. (2020) ‘Take YouTube's Dangers Seriously’. New York Times, October 15, 2020. https://www.nytimes.com/2020/04/20/technology/youtube-conspiracy-theories.html .Google Scholar
Paudyal, Pragya, and Wong, William. (2018) ‘Algorithmic Opacity: Making Algorithmic Processes Transparent through Abstraction Hierarchy’. Proceedings of the Human Factors and Ergonomics Society 2018 Annual Meeting, 62, 192–96.10.1177/1541931218621046CrossRefGoogle Scholar
Pearl, Judea. (2000) Causality: Models, Reasoning, and Inference. Cambridge: Cambridge University Press.Google Scholar
Queloz, Matthieu. (2018) ‘Davidsonian Causalism and Wittgensteinian Anti-causalism: A Rapprochement’. Ergo, 5, 153–72.Google Scholar
Rahimi, Ali, and Recht, Ben. (2017) ‘Reflections on Random Kitchen Sinks’. Arg Min Blog, December 5, 2017. http://www.argmin.net/2017/12/05/kitchen-sinks/ .Google Scholar
Reutlinger, Alexander. (2017) ‘Explanation beyond Causation? New Directions in the Philosophy of Scientific Explanation’. Philosophy Compass, 12, e12395. https://doi.org/10.1111/phc3.12395.CrossRefGoogle Scholar
Reutlinger, Alexander. (2018) ‘Extending the Counterfactual Theory of Explanation’. In Reutlinger, Alexander and Saatsi, Juha (eds.), Explanation beyond Causation: Philosophical Perspectives on Non-causal Explanations (New York: Oxford University Press), 7495.10.1093/oso/9780198777946.001.0001CrossRefGoogle Scholar
Roose, Kevin. (2020) Rabbit Hole. New York Times, Podcast. https://www.nytimes.com/column/rabbit-hole .Google Scholar
Ruben, David-Hillel. (1990) Explaining Explanation. Abingdon: Routledge.Google Scholar
Salmon, Wesley C. (1989) Four Decades of Scientific Explanation. Pittsburgh: University of Pittsburgh Press.Google Scholar
Schaffer, Jonathan. (2016) ‘Grounding in the Image of Causation’. Philosophical Studies, 173, 49100.10.1007/s11098-014-0438-1CrossRefGoogle Scholar
Sehon, Scott R. (2000) ‘An Argument against the Causal Theory of Action Explanation’. Philosophy and Phenomenological Research, 60, 6785.10.2307/2653428CrossRefGoogle Scholar
Selbst, Andrew D., and Powles, Julia. (2017) ‘Meaningful Information and the Right to Explanation’. International Data Privacy Law, 7, 233–42.10.1093/idpl/ipx022CrossRefGoogle Scholar
Skow, Bradford. (2014) ‘Are There Non-causal Explanations (of Particular Events)?’ British Journal for the Philosophy of Science, 65, 445–67.10.1093/bjps/axs047CrossRefGoogle Scholar
Van Fraassen, Bas C. (1980) The Scientific Image. Oxford: Clarendon Press.10.1093/0198244274.001.0001CrossRefGoogle Scholar
Vredenburgh, Kate. (2021) ‘The Right to Explanation’. Journal of Political Philosophy, 30, 209–29.10.1111/jopp.12262CrossRefGoogle Scholar
Wachter, Sandra, Mittelstadt, Brent, and Floridi, Luciano. (2017a) ‘Transparent, Explainable, and Accountable AI for Robotics’. Science: Robotics, 2. https://www.science.org/doi/10.1126/scirobotics.aan6080.Google ScholarPubMed
Wachter, Sanda, Mittelstadt, Brent, and Russell, Chris. (2018) ‘Counterfactual Explanations without Opening the Black Box: Automated Decisions and the GDPR’. Harvard Journal of Law and Technology, 31, 841–87.Google Scholar
Wachter, Sandra, Mittselstadt, Brent, and Floridi, Luciano. (2017b) ‘Why a Right to Explanation of Automated Decision-Making Does Not Exist in the General Data Protection Regulation’. International Data Privacy Law, 7, 7699.10.1093/idpl/ipx005CrossRefGoogle Scholar
Wilson, Alastair. (2018) ‘Metaphysical Causation’. Noûs, 52, 723–51.10.1111/nous.12190CrossRefGoogle Scholar
Woodward, James. (2003) Making Things Happen: A Theory of Causal Explanation. New York: Oxford University Press.Google Scholar
Zimmerman, Annette. (2020) ‘The A-level Results Injustice Shows Why Algorithms Are Never Neutral’. New Statesman, October 27, 2022. https://www.newstatesman.com/politics/2020/08/the-a-level-results-injustice-shows-why-algorithms-are-never-neutral.Google Scholar