Hostname: page-component-586b7cd67f-gb8f7 Total loading time: 0 Render date: 2024-11-25T06:06:06.523Z Has data issue: false hasContentIssue false

Purely Probabilistic Measures of Explanatory Power: A Critique

Published online by Cambridge University Press:  10 February 2022

William Roche*
Affiliation:
Department of Philosophy, Texas Christian University, Fort Worth, TX, US
Elliott Sober
Affiliation:
Department of Philosophy, University of Wisconsin at Madison, Madison, WI, US
*
*Corresponding author. Email: [email protected]
Rights & Permissions [Opens in a new window]

Abstract

All extant purely probabilistic measures of explanatory power satisfy the following technical condition: if Pr(E | H1) > Pr(E | H2) and Pr(E | ∼H1) < Pr(E | ∼H2), then H1’s explanatory power with respect to E is greater than H2’s explanatory power with respect to E. We argue that any measure satisfying this condition faces three serious problems—the Problem of Temporal Shallowness, the Problem of Negative Causal Interactions, and the Problem of Nonexplanations. We further argue that many such measures face a fourth problem—the Problem of Explanatory Irrelevance.

Type
Article
Copyright
© The Author(s), 2022. Published by Cambridge University Press on behalf of the Philosophy of Science Association

1. Introduction

There are clear cases in which two hypotheses each explain an observation, but one of them does so with greater explanatory power (or strength). Here’s one example:

Suppose that a randomized control trial is performed to see how the incidence of heart disease is affected by eliminating fried foods from one’s diet, and that the result is that the reduction is 10%. Now consider two causal models that might explain why this is the outcome. The first says that the reduction should have been between 9% and 11% while the second says that the reduction should have been between 2% and 20%. Suppose that both models are theoretically well-grounded, and that each is explanatorily relevant to the observed reduction. We think that there is a clear sense in which the first model has greater explanatory power than the second with respect to the observed reduction. Each model explains the observed reduction to some extent, but the explanation provided by the first is stronger than the one provided by the second.

Here’s another:

Suppose Jones smoked cigarettes and inhaled asbestos, and that both of these causally contributed to her getting lung cancer, where a person’s risk of lung cancer is greater if she inhales both carcinogens than if she inhales just one. Jones’s getting lung cancer is explained by her smoking cigarettes, by her inhaling asbestos, and by her doing both, but the smoking + asbestos explanation is more powerful than the smoking explanation is and it’s also more powerful than the asbestos explanation is.

We’ll consider other examples in what follows, but these two suffice to set the stage for our main question: Is there a plausible mathematical measure of the power with which a hypothesis H explains an explanadum E?

That the answer is affirmative has prima facie plausibility. Each of the previous examples can be fleshed out so that the different explanans statements have different probabilities and confer different probabilities on their explananda. Perhaps these probabilistic differences ground differences in explanatory power, and perhaps more generally explanatory power can be measured in terms of probabilities.

The idea of a mathematical measure of explanatory power carries baggage that more modest ideas about explanatory power do not shoulder. Here is a less demanding question:

For any H1, H2, and O, if H1 and H2 each explains O, is there a plausible criterion that determines whether H1’s explanation of O is more powerful than H2’s?

This is the question of whether there is a criterion for comparative explanatory power, where both hypotheses address the same observation. No numbers are involved. A measure of explanatory power does much more. For example, it would describe whether the power of H1’s explanation of O1 exceeds the power of H2’s explanation of O2, and by how much. That may sound like a paradigmatic example of apples and oranges, but that’s what a mathematical measure delivers. If you think that the power of an explanation has multiple orthogonal dimensions, with the result that H1 may be a more powerful explanation of O than H2 is on some dimensions, while the reverse is true on others (as do Ylikoski and Kuorikoski Reference Ylikoski and Kuorikoski2010), you may think that the quest for such a measure is quixotic.

That said, several able philosophers have dared to plunge ahead. Are they tilting at windmills? One such attempt is Schupbach and Sprenger’s (Reference Schupbach and Sprenger2011). They set out several conditions of adequacy on measures of explanatory power, which they use to derive the following measure:

$$\[{\rm{EXP}}{{\rm{O}}_{{\rm{SS}}}}\left( {{\rm{E}},\;{\rm{H}}} \right) = \frac{{\Pr ({\rm{H}}\;{\rm{|E}}) - \Pr ({\rm{H}}\;{\rm{|}} \sim {\rm{E}})}}{{\Pr ({\rm{H}}\;{\rm{|E}}) + \Pr ({\rm{H}}\;{\rm{|}} \sim {\rm{E}})}}\]$$

EXPOSS is a purely probabilistic measure of explanatory power (a PPM, for short) in that the value of EXPOSS(E, H) is fully determined by probabilities.Footnote 1, Footnote 2

Schupbach and Sprenger’s measure is not the only game in town. There are many—indeed, very many—purely probabilistic measures of explanatory power in logical space in addition to EXPOSS. Here are four:

$${\rm{EXP}}{{\rm{O}}_{{\rm{CT}}}}\left( {{\rm{E}},\;{\rm{H}}} \right) = \!\left\{ {\matrix{ {{{\Pr ({\rm{E}}\;{\rm{|H}}) - {\rm{Pr}}\left( {\rm{E}} \right)} \over {1 - {\rm{Pr}}\left( {\rm{E}} \right)}}\;{\rm{if}}\Pr ({\rm{E}}\;{\rm{|H}})\; \ge {\rm{Pr}}\left( {\rm{E}} \right)} \cr {{{\Pr ({\rm{E}}\;{\rm{|H}}) - {\rm{Pr}}\left( {\rm{E}} \right)} \over {{\rm{Pr}}\left( {\rm{E}} \right)}}{\rm{\ if}}\Pr ({\rm{E}}\;{\rm{|H}}) \lt Pr\left( E \right)} \cr } } \right.$$
$${\rm{EXP}}{{\rm{O}}_{\rm{G}}}\left( {{\rm{E}},{\rm{\;H}}} \right) = {\rm{log}}\left[ {{{\Pr {\rm{(E\;|H}})} \over {{\rm{Pr}}\left( {\rm{E}} \right)}}} \right]$$
$$\rm EXP{O_C}\left( {E,{\rm{ }}H} \right){\rm{ }} = {\rm{ }}Pr\left( {E{\rm{ }}|{\rm{ }}H} \right){\rm{ }}-{\rm{ }}Pr\left( {E{\rm{ }}|{\rm{ }}\sim H} \right)$$
$$\rm EXP{O_D}\left( {E,{\rm{ }}H} \right){\rm{ }} = {\rm{ }}Pr\left( {E{\rm{ }}|{\rm{ }}H} \right){\rm{ }}-{\rm{ }}Pr\left( E \right)$$

EXPOCT is defended in Crupi and Tentori (Reference Crupi and Tentori2012). EXPOG is defended in Good (Reference Good1960).Footnote 3 EXPOC and EXPOD are mentioned but not defended in Schupbach (Reference Schupbach2011). In what follows, we’ll use “The Five” to refer to these five measures.Footnote 4

No two of The Five are ordinally equivalent, but they all satisfy the following two conditions, each of which involves two likelihood inequalities:

TLI-1. If (a) Pr(E | H1) > Pr(E | H2) and (b) Pr(E | ∼H1) < Pr(E | ∼H2), then EXPO(E, H1) > EXPO(E, H2).

TLI-2. If (a) Pr(E | H1) ≥ Pr(E | H2) and (b) Pr(E | ∼H1) ≤ Pr(E | ∼H2), then EXPO(E, H1) ≥ EXPO(E, H2).

EXPOSS, EXPOCT, EXPOG, and EXPOD meet these conditions because they meet the following stronger conditions, each of which involves one likelihood inequality:

OLI-1. If Pr(E | H1) > Pr(E | H2), then EXPO(E, H1) > EXPO(E, H2).

OLI-2. If Pr(E | H1) ≥ Pr(E | H2), then EXPO(E, H1) ≥ EXPO(E, H2).

Neither OLI-1 nor OLI-2 is satisfied by EXPOC because this measure says that H’s explanatory power with respect to E depends not just on Pr(E | H) but also on Pr(E | ∼H).Footnote 5

Our focus in what follows will be on “TLI” measures, defined as purely probabilistic measures of explanatory power that meet both TLI-1 and TLI-2. These include each of The Five along with many additional measures. Consider, for example, the following:

$${\rm{EXP}}{{\rm{O}}_{{\rm{LR}}}}\left( {{\rm{E}},{\rm{\;H}}} \right) = {\rm{log}}\left[ {{{\Pr {\rm{(E\;|H}})} \over {{\rm{Pr(E\;|}}\sim{\rm{H}})}}} \right]$$

This measure and measures ordinally equivalent to it have been discussed in the context of confirmation (as a measure of the degree to which E incrementally confirms H),Footnote 6 but not, as far as we are aware, in the context of explanatory power. But, as EXPOLR meets both TLI-1 and TLI-2, it is included in our target class of purely probabilistic measures of explanatory power. The question is whether any TLI measure is adequate.

We want to emphasize that the issue here is not about inference. For example, in the cancer case described at the outset, we aren’t claiming that the smoking + asbestos explanation is more probable than the smoking explanation. It isn’t. The claim, rather, is that the smoking + asbestos explanation is stronger than the smoking explanation in that it sheds more light on why Jones has lung cancer. Both explanations shed light on why Jones has lung cancer, but the first does this to a greater extent.

The remainder of this article is organized as follows. In section 2, we present the Problem of Temporal Shallowness. In section 3, we introduce the Problem of Negative Causal Interactions. In section 4, we consider a reply that might be made to the objections we presented in the previous two sections. In section 5, we describe the Problem of Nonexplanations. In section 6, we consider a possible reply to what we’ve said about that problem. In section 7, we discuss the Problem of Explanatory Irrelevance. In section 8, we conclude.

2. The problem of temporal shallowness

Consider a causal chain from a distal cause Cd, to a proximate cause Cp, and then to an effect E. For example, you throw a baseball, which causes the baseball to hit a window, which causes the window to break. Someone then asks: Why did the window break? Consider two answers:

  1. (2.1) A baseball hit the window and was going 65 mph.

  2. (2.2) You threw a baseball, which caused the baseball to hit the window at 65 mph.

Notice that proposition (2.2) entails proposition (2.1) but not conversely. Explanation (2.1) is temporally shallow, in that it describes only the more proximate cause. Explanation (2.2) has more temporal depth, in that it traces the effect back to a proximate cause, and traces that proximate cause back to your dastardly deed. This is a case, we suggest, in which greater temporal depth makes for greater explanatory power: Explanation (2.2) is stronger than explanation (2.1).

The baseball example is mundane, but the point of the example applies to scientific examples as well. Evolutionary biologists and philosophers of biology routinely distinguish “proximate” from “ultimate” explanations. Ernst Mayr (Reference Mayr1961) used this terminology to distinguish two ways of explaining the phenotypes of organisms. For example, why do sunflowers turn toward the sun? A proximate explanation of that behavior would describe the mechanisms inside individual sunflowers that cause phototropism. An ultimate explanation would describe the evolutionary processes, including natural selection, that caused that trait to increase in frequency and remain at high frequency in the lineage leading to present day sunflowers. As Mayr (Reference Mayr1961, 1503) says, “[t]here is always a proximate set of causes and an ultimate set of causes; both have to be explained and interpreted for a complete understanding of the given phenomenon.” Mayr opposes temporal shallowness, but he also rejects the idea that only evolutionary considerations should be mentioned in explaining the present phenotypes of organisms.Footnote 7

To put the sunflower example into the baseball format, consider these two explanations of why sunflowers now turn toward the sun:

  1. (2.3) Sunflowers now have mechanism M inside their cells (where M has features F1, F2, …, Fn).

  2. (2.4) The ancestors of present day sunflowers were subject to selection for phototropism with the result that genes (G1, G2, …, Gm) that cause sunflowers to develop mechanism M (where M has features F1, F2, …, Fn) increased in frequency. These genes were passed down the generations, and present-day sunflowers inherited them, with the result that they have mechanism M.

What goes for the broken window also goes for the sunflowers. Propositions (2.3) and (2.4) are both explanatory, but we suggest that (2.4) has more explanatory power than (2.3).

Some readers may balk at this last, absolute, pronouncement. Perhaps they believe that what counts as the more powerful explanation depends on your interests. Sometimes all you want is the temporally shallow cause; at other times you want more. Neither is better than the other unconditionally. We sympathize with that pluralistic claim, but we do feel the tug of the idea that temporally more inclusive explanations are often more explanatorily powerful than explanations that are temporally less inclusive (Sober Reference Sober1999).

How do TLI measures address the question of temporal shallowness? Are some or all of them pluralistic? Consider two types of cases in which there is a causal chain from X to Y to Z (or, strictly speaking, from the events described by X to the events described by Y to the events described by Z):

Type 1 (i) X causes Y, (ii) Y causes Z, (iii) X and Y are logically independent of each other, (iv) Y and Z are logically independent of each other, (v) X increases the probability of Y, (vi) Y increases the probability of Z, and (vii) X is screened-off from Z both by Y and by ∼Y in that Pr(Z | Y&X) = Pr(Z | Y) and Pr(Z | ∼Y&X) = Pr(Z |∼Y).

Type 2 (i) X causes Y, (ii) Y causes Z, (iii) X entails and increases the probability of Y, (iv) Y increases the probability of Z, and (v) X is screened-off from Z by Y in that Pr(Z | Y&X) = Pr(Z | Y).

We show in Appendix 1 that TLI measures are heavily biased in favor of temporally shallow explanations in the following sense:

Theorem 1 In cases of Type 1, TLI measures entail that EXPO(Z, Y) > EXPO(Z, X) and EXPO(Z, Y) ≥ EXPO(Z, X&Y).

Theorem 2 In cases of Type 2, TLI measures entail that EXPO(Z, Y) ≥ EXPO(Z, X) and EXPO(Z, Y) ≥ EXPO(Z, X&Y).Footnote 8

Surely, however, an adequate measure of explanatory power should allow EXPO(Z, Y) to be less than EXPO(Z, X) or less than EXPO(Z, X&Y) in causal chains of either kind.Footnote 9 In such causal chains, TLI measures say that explanatory power is never increased by adding information about distal causes. Ernst Mayr is not the only scientist who would scoff at that suggestion.

It might be objected that our argument here depends on a wholesale commitment to the thesis that a more complete explanation of E is always more powerful than a less complete explanation of E. Philosophers often argue against this thesis by pointing out that more complete explanations are sometimes harder to understand and work with, and that less complete explanations sometimes apply to more explananda (see, e.g., Ylikoski and Kuorikoski [Reference Ylikoski and Kuorikoski2010, 216–17]). We agree with both these points, but they aren’t relevant to the thesis we want to endorse. We’re interested in whether more complete explanations ever have greater explanatory power than less complete explanations of the same explanandum when both are easy to grasp. In addition, L is not shown to be a better explanation of E1 than M is by the fact that L explains E1, E2, …, En whereas M explains only E1.Footnote 10 In any event, the examples we have discussed in this section do not fall in the cracks that Ylikoski and Kuorikoski’s observations reveal. Neither of the explanations of window breaking that we described strains our cognitive capacities to the breaking point, and the same point holds of the two explanations we described of sunflower phototropism.

3. The problem of negative causal interactions

At the start of this article, we considered an example in which asbestos exposure and smoking cigarettes both causally promote lung cancer. We assumed that lung cancer goes up in probability if an individual smokes, if an individual inhales asbestos, and if an individual does both, with lung cancer being more probable if both causes are present than it would be if there were just one.

Here we tweak the example, as shown in Table 1. Suppose that c, a, and i are positive, and that i > c and i > a. Now there is a causal interaction between smoking and asbestos; the model is not additive (which is the condition that obtains precisely when the interaction term i=0). In particular, suppose the interaction is negative. Although smoking cigarettes increases the probability of lung cancer if Jones doesn’t inhale asbestos, and inhaling asbestos increases the probability of lung cancer if Jones doesn’t smoke, smoking cigarettes decreases the probability of lung cancer if Jones inhales asbestos, and inhaling asbestos decreases the probability of lung cancer if Jones smokes. Somehow if Jones inhales cigarette smoke and asbestos particles, each inhibits the other’s carcinogenic propensities.Footnote 11

Table 1. The probability of Jones’s having lung cancer conditional on four conjunctions

Pr(Jones has lung cancer | –)
Jones inhales asbestos Jones does not inhale asbestos
Jones smokes cigarettes x + c + ai x + a
Jones does not smoke cigarettes x + c x

Now suppose Jones has lung cancer (this is the explanandum E) and each of the following three hypotheses is true:

H1: Jones smoked cigarettes.

H2: Jones inhaled asbestos.

H1&H2: Jones smoked cigarettes and Jones inhaled asbestos.

Jones’s cancer is explained by each of these hypotheses, but H1 and H2 provide censored pictures of the causal facts whereas H1&H2’s explanation is more causally complete. H1 and H2 each resemble the one-sided story presented by the prosecutor in a trial who neglects to mention details that would cast doubt on the defendant’s guilt. Doing so may be rhetorically powerful, but it is evidentially inferior to a fuller accounting that mentions both the bits of evidence that favor guilt and the bits that favor innocence. So it is with explanation—powerful explanations should describe both the causal promoters and the causal inhibitors that impacted the explanandum event. Removing all the latter reduces explanatory power.Footnote 12

How do TLI measures handle cases like that of Jones? Consider, more generally, cases of the following type:

Type 3 (i) Y causes X, (ii) Z causes X, and (iii) there’s a negative causal interaction in that (a) Pr(X | Y) is greater than Pr(X), (b) Pr(X | Z) is greater than Pr(X), (c) Pr(X | ∼Y&∼Z) is less than Pr(X | Y&Z), and (d) the latter is less than each of Pr(X | Y&∼Z) and Pr(X | ∼Y&Z).

We show in Appendix 2 that TLI measures are heavily biased in favor of censored explanations in that:

Theorem 3 In cases of Type 3, TLI measures entail that EXPO(X, Y) > EXPO(X, Y&Z) and EXPO(X, Z) > EXPO(X, Y&Z).

The upshot is that each of The Five, and all other TLI measures, entail an erroneous interpretation of negative causal interactions of the kind we’ve described. They all say that a hypothesis that describes both positive and negative causal factors has less explanatory power than a hypothesis that describes the positive and ignores the negative.Footnote 13, Footnote 14

We note that this point about negative causal interactions involves no commitment to the general thesis that more complete explanations are always more powerful. What we said at the end of the previous section about window-breaking and phototropism also applies to the present example of cigarettes and asbestos.

4. A possible reply to the temporal shallowness and the negative causal interactions problems

The arguments offered in the previous two sections concerning temporal shallowness and negative causal interactions could be defused if there were an important sense of explanatory power that demands the verdicts that TLI measures deliver on those topics. A possible defusion of this type can be found in the motivation that Schupbach and Sprenger (Reference Schupbach and Sprenger2011) offer for their proposed measure of explanatory power. They begin with the idea that there’s a sense of explanatory power on which explanatory power is a matter of surprise reduction. They further hold that surprise reduction is psychological and so should be understood in terms of actual or rational degrees of belief. They defend this idea by arguing that their interpretation of explanatory power “finds precedence in many classical discussions of explanation”:

Perhaps its clearest historical expression occurs when Peirce (Reference Peirce, Hartshorne and Weiss1931–35, 5.189) identifies the explanatoriness of a hypothesis with its ability to render an otherwise “surprising fact” as “a matter of course.” This sense of explanatory power may also be seen as underlying many of the most popular accounts of explanation. Most obviously, Deductive-Nomological and Inductive-Statistical accounts (Hempel Reference Hempel1965) and necessity accounts (Glymour Reference Glymour1980) explicitly analyze explanation in such a way that a theory that is judged to be explanatory of some explanandum will necessarily increase the degree to which we expect that explanandum. (Schupbach and Sprenger Reference Schupbach and Sprenger2011, 108)

We find this dubious, especially when it comes to Hempel’s D-N and I-S models. First, it’s clear from passages such as the following that Hempel sees his D-N and I-S models in terms of a nonpsychological, or objective, conception of explanation:

Scientific research seeks to give an account—both descriptive and explanatory—of empirical phenomena which is objective in the sense that its implications and its evidential support do not depend essentially on the individuals who happen to apply or to test them. This ideal suggests the problem of constructing a nonpragmatic conception of scientific explanation—a conception that requires reference to puzzled individuals no more than does the concept of mathematical proof. And it is this nonpragmatic conception of explanation with which the two covering-law models are concerned. (Hempel Reference Hempel and Fetzer2001, 82)

Second, in Hempel’s D-N model of explanation, what matters (roughly) is that the explanans propositions entail the explanandum proposition. Your degrees of belief (normative or actual) play no role, much less the idea that learning the explanans must change your degrees of belief in the explanandum. Even if it’s true that you increase your degree of belief in E when you use Hempel’s D-N model and conclude that H explains E, it doesn’t follow that Hempel’s model asserts that explanatory power is a matter of increase in degree of belief. Compare: Even if it’s true that you become excited when you use Hempel’s D-N model and conclude that H explains E, it doesn’t follow that Hempel’s model asserts that explanatory power is a matter of becoming excited.Footnote 15, Footnote 16

What about Peirce (Reference Peirce, Hartshorne and Weiss1931–35, Reference Peirce and Burks1958)? It’s true that the passage from Peirce that Schupbach and Sprenger cite includes the expressions “surprising fact” and “a matter of course.” However, the main subject of that passage is abduction, not the question of what an explanation is. Peirce thinks that when H is abductively inferred (or conjectured) from E, H explains E and reduces E’s surprisingness, but it doesn’t follow that he takes this to be essential to explanation. In fact, there are many passages where Peirce characterizes explanation with no appeal to surprise reduction or to an agent’s degrees of belief. One such passage is cited by Schupbach and Sprenger (Reference Schupbach and Sprenger2011, 108, n. 5). They quote Peirce as saying that “to explain a fact is to show that it is a necessary or, at least, a probable result from another fact, known or supposed.” This sounds like Hempel but without all the formal niceties.Footnote 17

In addition to appealing to the precedents of Hempel and Peirce, Schupbach and Sprenger cite Glymour (Reference Glymour1980) as a precursor of the idea of explanatory power as surprise reduction. Glymour’s basic idea is that when an explanans is deduced from necessary truths involving property identities, the resulting explanans shows that the explanandum is necessary rather than contingent. This is a contingency-reducing explanation, which Glymour doesn’t think of psychologistically.

Schupbach and Sprenger (Reference Schupbach and Sprenger2011, 108) also appeal to statistical practice, saying that the surprise-reduction concept of explanatory power “dominates statistical reasoning in which scientists are ‘explaining away’ surprise in the data by means of assuming a specific statistical model (e.g., in the omnipresent linear regression procedures).” Our response is that regression procedures are part of the frequentist toolkit, and frequentists have no truck with degrees of belief.

In summary, the historical precedents that Schupbach and Sprenger (Reference Schupbach and Sprenger2011) note do not lend credence to the idea that there is an important sense of “explanatory power” according to which explanatory power should be understood in terms of surprise reduction where the latter concept is understood in terms of the changes in credence that an agent makes or ought to make. This leaves it open that their starting idea can be motivated in some other way. We now turn to our third argument against TLI measures of explanatory power.

5. The problem of nonexplanations

Take some case of Type 1 (as described in section 2)—that is, an example in which X causes and increases the probability of Y, Y causes and increases the probability of Z, X and Y are logically independent of each other, Y and Z are logically independent of each other, and X is screened-off from Z both by Y and by ∼Y. Is X’s explanatory power with respect to Y greater than Z’s explanatory power with respect to Y? The answer is clearly yes. Because X is a cause of Y whereas Z is an effect of Y, X explains Y at least to some extent, whereas Z explains Y not at all.

This simple example in which an explanation is contrasted with a nonexplanation suggests a condition of adequacy on measures of explanatory power:

Nonexplanations Worse. If (i) X is an explanation of E in that X explains E at least to some extent and (ii) Y is a nonexplanation with respect to E in that Y explains E not at all, then EXPO(E, X) > EXPO(E, Y).Footnote 18

Nonexplanations have no explanatory power. Explanations have some explanatory power. Hence nonexplanations have less explanatory power than explanations.

TLI measures fail to satisfy Nonexplanations Worse. It’s straightforward to construct cases of Type 1, for example, in which:

(5.1) $${\rm{Pr(Y}}\,{\rm{|}}\,{\rm{Z)}} \gt {\rm{Pr(Y|}}\,\,{\rm{X)}}$$
(5.2) $${\rm{Pr(Y}}|\sim{\rm{Z}}) \lt {\rm{Pr}}({\rm{Y}}|\sim{\rm{X}})$$

We describe an example of this sort in Appendix 3. Because X is an explanation of Y whereas Z is not an explanation of Y, Nonexplanations Worse implies that EXPO(Y, X) > EXPO(Y, Z). However, given (5.1) and (5.2), TLI measures deliver the result that EXPO(Y, X) < EXPO(Y, Z).Footnote 19

Consider, for comparison, the following condition of adequacy on probabilistic measures of confirmation in the sense of increase in probability:

Nonconfirmers Worse. If (i) E confirms X in that E increases the probability of X at least to some extent and (ii) E is a nonconfirmer with respect to Y in that E decreases or has no impact on the probability of Y, then CONF(X, E) > CONF(Y, E).

This condition, qua condition of adequacy, is uncontroversial in Bayesian confirmation theory. The situation with respect to explanatory power seems similar: just as confirmers confirm to a greater degree than nonconfirmers, explanations explain more powerfully than nonexplanations.

6. A possible reply to the problem of nonexplanations

We have been understanding PPMs as unrestricted in the sense that the EXPO(E, H) is well defined even if H is not an explanation of E. The reason why is straightforward: Some of our beliefs about explanatory power concern nonexplanations, and we want to know how well PPMs do in terms of all our considered judgments about explanatory power. In contrast, Schupbach and Sprenger (Reference Schupbach and Sprenger2011) restrict EXPOSS, along with rival PPMs, as follows:

Potential-Explanation Restriction. EXPO(E, H) is defined precisely when H is a potential explanation of E in that H would explain E if H were true.Footnote 20

Given this way of understanding PPMs, TLI measures do not deliver the result that EXPO(Y, X) < EXPO(Y, Z) in our example of a causal chain from X to Y to Z because Z isn’t a potential explanation of Y, and so EXPO(Y, Z) isn’t defined.Footnote 21 The restriction allows TLI measures to avoid generating false claims about explanatory power. But these theories still fail to satisfy Nonexplanations Worse because they fail to deliver some manifestly true propositions about explanatory power.Footnote 22 The result is that the measures are incomplete, not that they entail falsehoods.

We argued in the previous section that it is a condition of adequacy on a quantitative measure of explanatory power that it judge nonexplanations as having less explanatory power than genuine explanations. This condition of adequacy of course obviously clashes with the restriction that Schupbach and Sprenger embrace. We hope the following comments will help adjudicate.

It is one thing for a mathematical criterion to fail to generate a result, as when the ratio x/y fails to be defined when y=0. It is another thing for a mathematical criterion to entail a result, which friends of the criterion wish to disown. Schupbach and Sprenger will understandably dislike what their mathematical criterion says about the causal chain from X to Y to Z discussed in section 5, but to stipulate that their criterion does not apply to such cases is not a legitimate solution. Where do such stipulations stop? Will friends of TLI measures of explanatory power urge that their theories do not apply to cases involving temporal shallowness or to cases involving negative causal factors? We are not saying that they are inclined to do so. Our question concerns the ground rules for using restrictive stipulations to avoid apparent counterexamples.

An anonymous referee for this journal suggests that friends of TLI measures of explanatory power could set aside the restriction strategy and instead simply stipulate that if H is not an explanation of E, then EXPO(E, H) has the lowest possible value. Consider, once more, the example of Schupbach and Sprenger’s proposed measure. EXPOSS(E, H) might be understood in terms of a function from inputs to outputs such that:

  1. (i) If the input consists of (a) explanandum E and (b) hypothesis H where H is not an explanation of E, then the output is –1.

  2. (ii) If the input consists of (a) explanandum E, (b) hypothesis H where H is an explanation of E, and (c) probability function Pr(–), then the output is a number equal to ${{\Pr ({\rm{H}}\;{\rm{|E}}) - \Pr ({\rm{H}}\;{\rm{|}}\sim{\rm{E}})} \over {\Pr ({\rm{H}}\;{\rm{|E}}) + \Pr ({\rm{H}}\;{\rm{|}}\sim{\rm{E}})}}$ .

The same referee further notes that this suggestion is similar to the fact that the likelihood ratio measure of confirmation—CONF(H, E) = ${{\Pr ({\rm{E}}\;{\rm{|H}})} \over {\Pr ({\rm{E}}\;{\rm{|}}\sim{\rm{H}})}}$ —is standardly understood so that CONF(H, E) is maximal at ∞ when Pr(E | H)>0 and Pr(E | ∼H)=0, even though, strictly speaking, division by 0 is undefined. There is an important difference here, however. Take some value for Pr(E | H) greater than Pr(E | ∼H). The limit of ${{\Pr ({\rm{E}}\;{\rm{|H}})} \over {\Pr ({\rm{E}}\;{\rm{|}}\sim{\rm{H}})}}$ approaches positive infinity if Pr(E | H) is held fixed and Pr(E | ∼H) is made smaller and smaller. This is the rationale behind the standard way of understanding the likelihood ratio measure. However, nothing about surprise reduction or ${{\Pr ({\rm{H}}\;{\rm{|E}}) - \Pr ({\rm{H}}\;{\rm{|}}\sim{\rm{E}})} \over {\Pr ({\rm{H}}\;{\rm{|E}}) + \Pr ({\rm{H}}\;{\rm{|}}\sim{\rm{E}})}}$ justifies the stipulation put forward in (i). Rather, what motivates (i) is the idea that nonexplanations have no explanatory power. We of course agree with that idea, but we think it’s reasonable to expect that a mathematical measure should capture that idea without recourse to a special stipulation.

We think our criticism of the referee’s suggested stipulation can be further motivated by the following “continuity argument.” Consider a sequence of hypotheses H1, H2, etc. where each entry has less explanatory power than the one before, relative to a fixed explanandum proposition E. According to Schupbach and Sprenger’s measure, these hypotheses can be ranked for their explanatory power just by looking at the value of Pr(E|Hi)—the bigger this is, the more explanatory it is, with the result that Pr(E|BestPossible)=1, and Pr(E|worst possible) = 0. The point of relevance here is that numerous nonexplanations are nowhere “close” to Pr(E|WorstPossible) = 0. Consider, for instance, some hypothesis H about the present that deductively retrodicts E. Although H fails to explain E (assuming that the past isn’t explained by the present), the fact remains that Pr(E | H)=1, which is also the value of Pr(E|BestPossible). Here the reduction in E’s surprisingness is maximal, and yet the referee’s suggested stipulation puts H in the same camp as hypotheses that entail ∼E.

In summary, Schupbach and Sprenger (Reference Schupbach and Sprenger2011) stipulate that their measure applies only to potential explanations. The referee mentioned suggests that if this stipulation fails to satisfy, proponents of TLI measures can stipulate instead that a nonexplanation has the minimum possible degree of explanatory power. These are different stipulations, but there is a common flaw, or so we have suggested. The use of stipulations in either of these two ways reminds us of Russell’s (Reference Russell1919, 71) famous quip that “[t]he method of ‘postulating’ what we want has many advantages; they are the same as the advantages of theft over honest toil.”

7. The problem of explanatory irrelevance

In the previous two sections we discussed the dichotomy between an explanation of E and a nonexplanation of E. We now want to consider something that is a matter of degree. If H is a genuine explanation of E, and you conjoin explanatorily irrelevant proposition I to H, does the resulting conjunction, H&I, provide a less powerful explanation of E than H provides?

A related question has a distinguished history in philosophical discussions of explanation. Consider two influential examples:

This sample of table salt dissolves in water, for it had a dissolving spell cast upon it, and all samples of table salt that have had dissolving spells cast on them dissolve in water. (Kyburg Reference Kyburg1965, 147)

John Jones avoided becoming pregnant during the past year, for he has taken his wife’s birth control pills regularly, and every man who regularly takes birth control pills avoids pregnancy. (Salmon Reference Salmon and Salmon1971, 34)

Kyburg and Salmon present these examples to illustrate how a genuine explanation can be turned into a nonexplanation by adding information that is explanatorily irrelevant. They seem to be committed to the general thesis that if H contains even one detail that is explanatorily irrelevant to E, then H is not an explanation of E. We think this may go too far.

Shifting from the question of whether a hypothesis explains to the question of how powerfully it does so, our modest thesis about irrelevant additions is this:

Adding Irrelevance Sometimes Hurts. EXPO(E, H) is sometimes greater than EXPO(E, H&I) when (a) I is explanatorily irrelevant to E and (b) I is probabilistically independent of E, H, and E&H.

Salmon and Kyburg appear to be purists; for them adding even a teaspoon of irrelevancy to a sea of explanatorily relevant information is enough to turn the sea into a nonexplanation. We aren’t purists about such additions when the question is degree of explanatory power. That’s why we say sometimes. Here we have in mind examples in which a brief, telling H is conjoined with a giant, complicated, and doubly irrelevant proposition I.

How do TLI measures fair in terms of the Adding Irrelevance Sometimes Hurts condition? The situation is mixed, in that some TLI measures meet Adding Irrelevance Sometimes Hurts whereas others do not. We show in Appendix 4 that Adding Irrelevance Sometimes Hurts is met by EXPOC and EXPOLR but not by EXPOSS, EXPOCT, EXPOG, or EXPOD.Footnote 23

TLI measures would satisfy the Adding Irrelevance Sometimes Hurts condition if they were understood so that H&I fails to explain E when I is explanatorily irrelevant to E, and if nonexplanations are stipulated to have the minimum possible degree of explanatory power. However, this way of understanding those measures leads back to our earlier discussion of the role of stipulation in the formulation of PPMs of explanatory power.

8. Concluding comments

We have focused on TLI measures of explanatory power, which include all extant purely probabilistic measures of explanatory power. All TLI measures face three problems—the Problem of Temporal Shallowness, the Problem of Negative Causal Interactions, and the Problem of Nonexplanations. Some TLI measures face a fourth problem—the Problem of Explanatory Irrelevance. These problems together raise the suspicion that an adequate measure of explanatory power can’t be purely probabilistic. However, the door has not slammed shut absolutely. It is left open that an adequate PPM of explanatory power can be found; the place to search, we suggest, is in measures that violate the TLI constraints.

Acknowledgments

We thank the three anonymous referees for the journal, and also the journal’s editorial team, for helpful comments on prior versions of the paper.

Appendix 1

Proof of Theorem 1

The aim is to show:

Theorem 1 In cases of Type 1, TLI measures entail that EXPO(Z, Y) > EXPO(Z, X) and EXPO(Z, Y) ≥ EXPO(Z, X&Y).

Consider a case of Type 1 in which (i) X causes Y, (ii) Y causes Z, (iii) X and Y are logically independent of each other, (iv) Y and Z are logically independent of each other, (v) X increases the probability of Y, (vi) Y increases the probability of Z, and (vii) X is screened-off from Z both by Y and by ∼Y in that Pr(Z | Y&X) = Pr(Z | Y) and Pr(Z | ∼Y&X) = Pr(Z | ∼Y). It follows from condition (vii) that (see Shogenji Reference Shogenji2003):

(A1.1) $$\Pr ({\rm{Z}}\;{\rm{|X}}) - \Pr \left( {\rm{Z}} \right) = \;\left[ {{{\Pr ({\rm{Y}}\;{\rm{|X}}) - {\rm{Pr}}\left( {\rm{Y}} \right)} \over {1 - {\rm{Pr}}\left( {\rm{Y}} \right)}}} \right]\left[ {\Pr ({\rm{Z}}\;{\rm{|Y}}) - {\rm{Pr}}\left( {\rm{Z}} \right)} \right]$$

The fact that Pr(Z | ∼Y&X) = Pr(Z | ∼Y) implies that Pr(∼Y&X)>0, and so Pr(Y | X)<1. This, in turn, implies:

(A1.2) $${{\Pr ({\rm{Y}}\;{\rm{|X}}) - {\rm{Pr}}\left( {\rm{Y}} \right)} \over {1 - {\rm{Pr}}\left( {\rm{Y}} \right)}} \lt 1$$

(A1.1) and (A1.2) together imply:

(A1.3) $$\rm Pr\left( {Z{\rm{ }}|{\rm{ }}X} \right){\rm{ }} \lt {\rm{ }}Pr\left( {Z{\rm{ }}|{\rm{ }}Y} \right)$$

Next, note that condition (vii) implies:

(A1.4) $$\eqalign{\Pr {\rm{(Z\;|}}\sim {\rm{X}}) = \Pr {\rm{(Y\;|}}\sim {\rm{X}})\Pr {\rm{(Z\;|Y\;}}& \sim X) + \;\Pr {\rm{(}}\sim {\rm{Y\;|}}\sim {\rm{X}})\Pr {\rm{(Z\;|}}\sim {\rm{Y\;}}& \sim X) \cr = \Pr {\rm{(Y\;|}}\sim {\rm{X}})\Pr {\rm{(Z\;|Y}}) + {\rm{\;}}\Pr {\rm{(}}\sim {\rm{Y\;|}}\sim {\rm{X}})\Pr {\rm{(Z\;|}}\sim {\rm{Y}})} $$

The sum after the second equality sign in (A1.4) equals:

(A1.5) $$\Pr {\rm{(Y\;|}}\sim {\rm{X}})\left[ {\Pr {\rm{(Z\;|Y}}) - \Pr {\rm{(Z\;|}}\sim {\rm{Y}})} \right] + \Pr {\rm{(Z\;|}}\sim {\rm{Y}})$$

Because Pr(Z | Y) > Pr(Z), it follows that Pr(Z | Y) > Pr(Z | ∼Y). Given this, and given that Pr(Y | ∼X)>0 because X and Y are logically independent of each other, it follows that the first addend in (A1.5) is positive. Given this, and given that Pr(Z | ∼X) equals (A1.5), it follows that:

(A1.6) $${\rm{Pr}}\left( {{\rm{Z | \sim X}}} \right){\rm{ \gt Pr}}\left( {{\rm{Z | \sim Y}}} \right)$$

(A1.3), (A1.6), and TLI-1 together imply that EXPO(Z, Y) > EXPO(Z, X). It’s a theorem of the probability calculus that (for any propositions P1, P2, and P3):

(A1.7) $${\rm{\;Pr(P1 | P2}} \vee {\rm{P3) = Pr(P2 | P2}} \vee {\rm{P3)Pr}}\left( {{\rm{P1 | P2}}} \right){\rm{ + Pr(P3\& \sim P2 | P2}} \vee {\rm{P3)Pr}}\left( {{\rm{P1 | P3\& \sim P2}}} \right) $$

Pr(Z | ∼(X&Y)) equals Pr(Z | ∼X∨∼Y)), and so by (A1.7) also equals:

(A1.8) $${\rm{Pr(\sim X | \sim X}} \vee {\rm{\sim Y)Pr}}\left( {{\rm{Z | \sim X}}} \right){\rm{ + Pr(\sim Y\& X | \sim X}} \vee {\rm{\sim Y)Pr}}\left( {{\rm{Z | \sim Y\& X}}} \right) $$

Because Pr(Z | ∼Y&X) = Pr(Z | ∼Y), it follows that (A1.8) equals:

(A1.9) $${\rm{Pr(\sim X | \sim X}} \vee {\rm{\sim Y)Pr}}\left( {{\rm{Z | \sim X}}} \right){\rm{ + Pr(\sim Y\& X | \sim X}} \vee {\rm{\sim Y)Pr}}\left( {{\rm{Z | \sim Y}}} \right) $$

Given that Pr(Z | ∼Y) < Pr(Z | ∼X), it follows that (A1.9) is greater than:

(A1.10) $${\rm{Pr(\sim X | \sim X}} \vee {\rm{\sim Y)Pr}}\left( {{\rm{Z | \sim Y}}} \right){\rm{ + Pr(\sim Y\& X | \sim X}} \vee {\rm{\sim Y)Pr}}\left( {{\rm{Z | \sim Y}}} \right)$$

This sum equals Pr(Z | ∼Y), and so:

(A1.11) $${\rm{Pr}}\left( {{\rm{Z | \sim }}\left( {{\rm{X\& Y}}} \right)} \right){\rm{ \gt Pr}}\left( {{\rm{Z | \sim Y}}} \right)$$

Given this, and given that, as per condition (vii), Pr(Z | X&Y) = Pr(Z | Y), it follows by TLI-2 that EXPO(Z, Y) ≥ EXPO(Z, X&Y). Hence any measure meeting TLI-1 and TLI-2 implies that EXPO(Z, Y) > EXPO(Z, X) and EXPO(Z, Y) ≥ EXPO(Z, X&Y).

Proof of Theorem 2

The aim is to show:

Theorem 2 In cases of Type 2, TLI measures entail that EXPO(Z, Y) ≥ EXPO(Z, X) and EXPO(Z, Y) ≥ EXPO(Z, X&Y).

Consider a case of Type 2 in which (i) X causes Y, (ii) Y causes Z, (iii) X entails and increases the probability of Y, (iv) Y increases the probability of Z, and (v) X is screened-off from Z by Y in that Pr(Z | Y&X) = Pr(Z | Y). Given that X entails Y, it follows that X is logically equivalent to Y&X. Hence:

(A1.12) $${\rm{Pr}}\left( {{\rm{Z | X}}} \right){\rm{ = Pr}}\left( {{\rm{Z | Y\& X}}} \right)$$

This and the fact that Pr(Z | Y&X) = Pr(Z | Y) imply:

(A1.13) $${\rm{Pr}}\left( {{\rm{Z | X}}} \right){\rm{ = Pr}}\left( {{\rm{Z | Y}}} \right)$$

The law of total probability implies:

(A1.14) $${\rm{Pr}}\left( {{\rm{Z | \sim X}}} \right){\rm{ = Pr}}\left( {{\rm{Y | \sim X}}} \right){\rm{Pr}}\left( {{\rm{Z | Y\& \sim X}}} \right){\rm{ + Pr}}\left( {{\rm{\sim Y | \sim X}}} \right){\rm{Pr}}\left( {{\rm{Z | \sim Y\& \sim X}}} \right) $$

Pr(Z | Y&X) = Pr(Z | Y) only if Pr(Z | Y&X) = Pr(Z | Y&∼X). So, since Pr(Z | Y&X) = Pr(Z | Y), the right-hand side of (A1.14) equals:

(A1.15) $${\rm{Pr}}\left( {{\rm{Y | \sim X}}} \right){\rm{Pr}}\left( {{\rm{Z | Y}}} \right){\rm{ + Pr}}\left( {{\rm{\sim Y | \sim X}}} \right){\rm{Pr}}\left( {{\rm{Z | \sim Y\& \sim X}}} \right) $$

X entails Y, and so ∼Y is logically equivalent to ∼Y&∼X. Hence (A1.15) is equal to:

(A1.16) $${\rm{Pr}}\left( {{\rm{Y | \sim X}}} \right){\rm{Pr}}\left( {{\rm{Z | Y}}} \right){\rm{ + Pr}}\left( {{\rm{\sim Y | \sim X}}} \right){\rm{Pr}}\left( {{\rm{Z | \sim Y}}} \right)$$

This simplifies to:

(A1.16) $${\rm{Pr}}\left( {{\rm{Y | \sim X}}} \right)\left[ {{\rm{Pr}}\left( {{\rm{Z | Y}}} \right){\rm{ - Pr}}\left( {{\rm{Z | \sim Y}}} \right)} \right]{\rm{ + Pr}}\left( {{\rm{Z | \sim Y}}} \right)$$

Since Pr(Z | Y) > Pr(Z), Pr(Z | Y) > Pr(Z | ∼Y). Hence, as Pr(Y | ∼X)≥0, the first addend in (A1.16) is nonnegative. Hence:

(A1.17) $${\rm{Pr}}\left( {{\rm{Z | \sim X}}} \right) \ge {\rm{Pr}}\left( {{\rm{Z | \sim Y}}} \right)$$

(A1.13), (A1.17), and TLI-2 together imply that EXPO(Z, Y) ≥ EXPO(Z, X). Hence any measure meeting TLI-2 implies that EXPO(Z, Y) ≥ EXPO(Z, X).

Appendix 2

The aim is to show:

Theorem 3 In cases of Type 3, TLI measures entail that EXPO(X, Y) > EXPO(X, Y&Z) and EXPO(X, Z) > EXPO(X, Y&Z).

Consider a case of Type 3 where (i) Y causes X, (ii) Z causes X, and (iii) there’s a negative causal interaction in that (a) Pr(X | Y) is greater than Pr(X), (b) Pr(X | Z) is greater than Pr(X), (c) Pr(X | ∼Y&∼Z) is less than Pr(X | Y&Z), and (d) the latter is less than each of Pr(X | Y&∼Z) and Pr(X | ∼Y&Z). Given that Pr(X | Y&Z) is less than Pr(X | Y&∼Z), it follows that:

(A2.1) $${\rm{Pr}}\left( {{\rm{X | Y}}} \right){\rm{ \gt Pr}}\left( {{\rm{X | Y\& Z}}} \right)$$

Because Pr(X | Y&Z) is less than Pr(X | ∼Y&Z), it follows that:

(A2.2) $${\rm{Pr}}\left( {{\rm{X | Z}}} \right){\rm{ \gt Pr}}\left( {{\rm{X | Y\& Z}}} \right)$$

∼(Y&Z) is logically equivalent to ∼Y∨∼Z, and so:

(A2.3) $${\rm{Pr}}\left( {{\rm{X | \sim }}\left( {{\rm{Y\& Z}}} \right)} \right){\rm{ = Pr(X | \sim Y}} \vee {\rm{\sim Z)}}$$

Given this, and given (A1.7) in Appendix 1, it follows that Pr(X | ∼(Y&Z)) equals each of the following:

(A2.4) $${\rm{Pr(\sim Y | \sim Y}} \vee {\rm{\sim Z)Pr(X | \sim Y) + Pr(\sim Z\& Y | \sim Y}} \vee {\rm{\sim Z)Pr(X | \sim Z\& Y)}} $$
(A2.5) $${\rm{Pr(\sim Z | \sim Y}} \vee {\rm{\sim Z)Pr(X | \sim Z) + Pr(\sim Y\& Z | \sim Y}} \vee {\rm{\sim Z)Pr(X | \sim Y\& Z)}} $$

(A2.4) simplifies to:

(A2.6) $${\rm{Pr(X | \sim Y) + Pr(Y | \sim Y}} \vee {\rm{\sim Z)[Pr(X | \sim Z\& Y) - Pr(X | \sim Y)]}} $$

(A2.5) simplifies to:

(A2.7) $${\rm{Pr(X | \sim Z) + Pr(Z | \sim Y}} \vee {\rm{\sim Z)[Pr(X | \sim Y\& Z) - Pr(X | \sim Z)]}} $$

Now focus on the differences in the second addends in (A2.6) and (A2.7). First, given that Pr(X | Y&∼Z) > Pr(X | Y&Z), it follows that:

(A2.8) $${\rm{Pr(X | \sim Z\& Y) \gt Pr(X | Y)}}$$

Given this, given that Pr(X | Y) > Pr(X), and given that Pr(X | Y) > Pr(X) precisely when Pr(X | Y) > Pr(X | ∼Y), it follows that (A2.6) is greater than Pr(X | ∼Y). Hence:

(A2.9) $${\rm{Pr}}\left( {{\rm{X | \sim Y}}} \right){\rm{ \lt Pr}}\left( {{\rm{X | \sim }}\left( {{\rm{Y\& Z}}} \right)} \right)$$

Second, given that Pr(X | ∼Y&Z) > Pr(X | Y&Z), it follows that:

(A2.10) $${\rm{Pr(X | \sim Y\& Z) \gt Pr(X | Z)}}$$

Given this, given that Pr(X | Z) > Pr(X), and given that Pr(X | Z) > Pr(X) precisely when Pr(X | Z) > Pr(X | ∼Z), it follows that (A2.7) is greater than Pr(X | ∼Z). Hence:

(A2.11) $${\rm{Pr}}\left( {{\rm{X | \sim Z}}} \right){\rm{ \lt Pr}}\left( {{\rm{X | \sim }}\left( {{\rm{Y\& Z}}} \right)} \right)$$

Given (A2.1), (A2.9), and any TLI measure, it follows that EXPO(X, Y) > EXPO(X, Y&Z). Given (A2.2), (A2.10), and any TLI measure, it follows that EXPO(X, Z) > EXPO(X, Y&Z).

Appendix 3

Take a causal chain from X to Y to Z, and consider the following probability distribution:

X Y Z Pr
T T T 1/4
T T F 14669/211750
T F T 3/14
T F F 3/20
F T T 4/39
F T F 117352/4129125
F F T 6/55
F F F 21/275

This distribution is such that:

(A3.1) $$ \rm Pr(Y | X) \approx 0.467 \gt 0.450 \approx Pr(Y)$$
(A3.2) $$\rm Pr(Z | Y) \approx 0.783 \gt 0.676 \approx Pr(Z)$$
(A3.3) $$ \rm Pr(Z | Y\& X) = Pr(Z | Y) \approx 0.783$$
(A3.4) $$\Pr \left( {Z{\rm{ }}|{\rm{ }}\sim Y\& X} \right){\rm{ }} = {\rm{ }}\Pr \left( {Z{\rm{ }}|{\rm{ }}\sim Y} \right) \approx 0.588$$
(A3.5) $${\rm{Pr}}\left( {{\rm{Y | Z}}} \right) \approx {\rm{0}}{\rm{.522 \gt 0}}{\rm{.467}} \approx {\rm{Pr}}\left( {{\rm{Y | X}}} \right)$$
(A3.6) $${\rm{Pr}}\left( {{\rm{Y | \sim Z}}} \right) \approx {\rm{0}}{\rm{.301 \lt 0}}{\rm{.414}} \approx {\rm{Pr}}\left( {{\rm{Y | \sim X}}} \right)$$

(A3.1) implies that X increases the probability of Y. (A3.2) implies that Y increases the probability of Z. The fact that none of the probabilities in the distribution equals 0 implies that X and Y are logically independent of each other, and that the same is true of Y and Z. (A3.3) and (A3.4) imply that X is screened-off from Z both by Y and by ∼Y. (A3.5) implies (5.1). (A3.6) implies (5.2). This is a case, then, of Type 1 where both (5.1) and (5.2) hold.

Appendix 4

Suppose that H explains and increases the probability of E, that I is explanatorily irrelevant to E, and that I is probabilistically independent of E, H, and E&H. Given this last supposition, it follows that:

(A4.1) $${\rm{Pr}}\left( {{\rm{E | H}}} \right){\rm{ = Pr}}\left( {{\rm{E | H\& I}}} \right) $$

Given this, in turn, it follows that EXPOSS(E, H) = EXPOSS(E, H&I), EXPOCT(E, H) = EXPOCT(E, H&I), EXPOG(E, H) = EXPOG(E, H&I), and EXPOD(E, H) = EXPOD(E, H&I). This follows because EXPOSS, EXPOCT, EXPOG, and EXPOD meet the following strengthened version of OLI-1:

$${\bf OLI - 1*.}\ {\rm{If}}\,{\rm{Pr}}\left( {{\rm{E | H1}}} \right){\rm{ \gt / = / \lt Pr}}\left( {{\rm{E | H2}}} \right){\rm{, then EXPO}}\left( {{\rm{E, H1}}} \right){\rm{ \gt / = / \gt EXPO}}\left( {{\rm{E, H2}}} \right){\rm{.}}$$

Next, continuing to suppose that H explains and increases the probability of E, that I is explanatorily irrelevant to E, and that I is probabilistically independent of E, H, and E&H, consider the following probability distribution:

E H I Pr
T T T 25/38
T T F 8/39
T F T 1/199
T F F 304/194025
F T T 0
F T F 0
F F T 961323/9671798
F F F 2563528/82719325

This distribution is such that:

(A4.2) $${\rm{EXP}}{{\rm{O}}_{\rm{C}}}\left( {{\rm{E}},{\rm{ H}}} \right) \approx 0.952 \gt 0.381 \approx {\rm{EXP}}{{\rm{O}}_{\rm{C}}}\left( {{\rm{E}},{\rm{ H\& I}}} \right) $$
(A4.3) $${\rm{EXP}}{{\rm{O}}_{{\rm{LR}}}}\left( {{\rm{E}},{\rm{ H}}} \right) \approx 4.377 \gt {\rm{ }}0.692 \approx {\rm{EXP}}{{\rm{O}}_{{\rm{LR}}}}\left( {{\rm{E}},{\rm{ H\& I}}} \right) $$

Hence Adding Irrelevance Sometimes Hurts is met by EXPOC and EXPOLR but not by EXPOSS, EXPOCT, EXPOG, or EXPOD.

Footnotes

1 EXPOSS is also defended in Schupbach (Reference Schupbach, McCain and Poston2017).

2 There is room in logical space for partially probabilistic measures of explanatory power. These are measures on which EXPO(E, H) is partially determined by probabilities. Imagine, for instance, a variant of EXPOSS on which H’s degree of explanatory power with respect to E is a weighted average of EXPOSS(E, H) and H’s score in terms of explanatory virtues such as mechanism and unification (on which see Cabrera [Reference Cabrera2017]). If H’s score in terms of some set of explanatory virtues is not fully determined by probabilities, then this variant of EXPOSS is a partially probabilistic measure of explanatory power. There is also room in logical space for (a) nonprobabilistic measures of explanatory power and (b) accounts of explanatory power that don’t involve a measure. See Northcott (Reference Northcott2013) and Ylikoski and Kuorikoski (Reference Ylikoski and Kuorikoski2010) for examples.

3 Brössel (Reference Brössel2015) notes that EXPOG is ordinally equivalent to the following measure from Popper (Reference Popper1954):

$${\rm{EXP}}{{\rm{O}}_{\rm{P}}}\left( {{\rm{E}},{\rm{\;H}}} \right) = {{\Pr {\rm{(E\;|H}}) - {\rm{Pr}}\left( {\rm{E}} \right)} \over {\Pr {\rm{(E\;|H}}) + {\rm{Pr}}\left( {\rm{E}} \right)}}$$

4 There’s a growing literature on purely probabilistic measures of explanatory power. In addition to the works discussed in this article, see Cohen (Reference Cohen2015, Reference Cohen2016, Reference Cohen2018), Eva and Stern (Reference Eva and Stern2019), and Ströing (Reference Ströing2018).

5 Three comments are in order. First, OLI-1 is in effect Requirement 3 in Brössel (Reference Brössel2015). Second, it might not be obvious that EXPOSS meets OLI-1 and OLI-2. If so, see Schupbach (Reference Schupbach, McCain and Poston2017, 48) for a proof. Third, TLI-1, TLI-2, OLI-1, and OLI-2 have confirmation-theoretic analogues. TLI-1’s analogue, for example, is:

If (a) Pr(E | H1) > Pr(E | H2) and (b) Pr(E | ∼H1) < Pr(E | ∼H2), then CONF(H1, E) > CONF(H2, E).

This is the so-called Weak Law of Likelihood. It is similar to the “Weak Likelihood Principle” defended by Joyce (Reference Joyce2019, sec. 3). See Roche and Shogenji (Reference Roche and Shogenji2014) for further discussion and references.

6 See, for example, Eells and Fitelson (Reference Eells and Fitelson2002), Huber (Reference Huber2008), and Zalabardo (Reference Zalabardo2009).

7 That said, it’s worth noting that Mayr’s use of the term “ultimate” sends a different message. The word doesn’t suggest that proximate and distal explanations are coequal partners. Rather, it suggests that proximate explanations play second fiddle. In this respect, Mayr’s “ultimate” stands side by side with Theodosius Dobzhansky’s (Reference Dobzhansky1973) slogan that “nothing in biology makes sense except in the light of evolution.”

8 If X entails Y, then X is logically equivalent to X&Y and so X and X&Y are interchangeable in the context of probability. It follows that if X entails Y, then EXPO(Z, X) = EXPO(Z, X&Y) and thus EXPO(Z, Y) ≥ EXPO(Z, X) if and only if EXPO(Z, Y) ≥ EXPO(Z, X&Y). So the second inequality in Theorem 2 is redundant given the first. We included both inequalities in case readers didn’t realize that EXPO(Z, X) = EXPO(Z, X&Y) if X entails Y.

9 There is a similar problem for purely probabilistic causal theories of mental content. See Roche and Sober (Reference Roche and Sober2021).

10 This is not to deny that a theory that provides a unifying explanation of diverse phenomena is often better supported by those phenomena (or has greater predictive accuracy) than a theory that provides a disunifying explanation of those same phenomena. However, our subject here isn’t the evidential support of theories, but concerns how powerfully theories would explain if they were true. See Sober (Reference Sober and Stadler2003) for discussion of this distinction.

11 For the sake of a simple example, we’re discussing dichotomous causal variables here, but negative causal interactions can also arise for ordered n-state variables and for continuous variables.

12 Humphreys (Reference Humphreys1989, 101) says that if one asks for an explanation of why Y is true (more precisely, why Y in S at t), an appropriate answer will take the form “Y is true because of X, and despite Z.” He says that Z isn’t part of the explanation of Y proper; rather, Z’s role is to make it clearer how the events described in X brought about Y. He also distinguishes between explanation and understanding, saying that only contributing causes explain the outcome, but a specification of the counteracting causes is necessary for a full understanding of how the outcome came about. We agree with Humphreys that “Y is true because X and in spite of Z” usually means that X promoted and Z tended to prevent. However, we think that “because” in this context is not a synonym for explanation. Humphreys is denying that X and Z together explain Y. We think they do. What is more, we suspect that the distinction between X’s explaining why Y is true and X’s allowing us to understand why Y is true is a distinction without a difference.

13 This objection to TLI measures supplements a nice objection that Glymour (Reference Glymour2015, 596) develops. Suppose X causes Z, X causes Y, and Y causes Z. As noted in the preceding text, EXPOSS(Z, Y) > EXPOSS(Z, X) precisely when Pr(Z | Y) > Pr(Z | X). The problem is that “Pr(Z | Y)” fails to hold fixed the value of X, so it fails to distinguish the causal-explanatory power of Y from its noncausal predictive power.

14 It might seem that TLI measures could be restricted in scope to cases where the hypotheses at issue are incompatible with each other, and it might seem that if they were thus understood, then cases of Type Footnote 3 wouldn’t be a problem for them because H1&H2 isn’t incompatible with H1 or with H2. It’s true that TLI measures could be understood in that fashion. But if they were, cases of Type 3 would still be a problem for them because they would then be unable to capture facts about explanatory power in cases of Type Footnote 3. They wouldn’t say false things about such cases, but they would fail to report some truths. Further, there are lots of other kinds of cases in which the hypotheses at issue aren’t incompatible with each other. The previous section on temporal shallowness provides examples. TLI measures would fall silent there as well.

15 The situation with respect to Hempel’s I-S model is less straightforward. Hempel (Reference Hempel1965, sec. 3.4.2) explicitly notes that I-S explanations at a given time are relative to the knowledge situation at that time, where this consists of “the class of all statements asserted or accepted by empirical science” (395) at that time. But this doesn’t make I-S explanations relative to any particular subject. Hempel (Reference Hempel and Fetzer2001) later acknowledges that he would prefer a model on which statistical explanations are not relative to knowledge situations, and that he never showed that there can be no such model. See Salmon (Reference Salmon2006) for helpful discussion of Hempel’s I-S model and the issue of epistemic relativity. Relatedly, Colombo et al. (Reference Colombo, Bucher and Sprenger2017, 2) claim that “[a]ccording to several philosophers, the power of an explanation is manifest in the amount of statistical information that an explanans H provides about an explanandum E, given some class or population S. In particular, it has to be the case that Prob(E | H&S) > Prob(E | S).” They cite Jeffrey (Reference Jeffrey and Salmon1971) and Salmon (Reference Salmon and Salmon1971) as two examples of such philosophers. However, Schupbach and Sprenger (Reference Schupbach and Sprenger2011, 107, n. 3) correctly note that this isn’t true; Jeffrey and Salmon don’t require an increase in probability. The same is true of Salmon’s (1984) causal model of explanation.

16 Colombo (Reference Colombo2017, 510) holds that Schupbach and Sprenger’s idea of explanatory power as surprise reduction coheres nicely with Hempel’s “Deductive-Nomological model, which can be understood as claiming that an explanatory hypothesis will necessarily increase the degree to which the explanandum is expected.” Hempel often writes in terms of “expectedness.” But, as Hempel’s point that explanations are like proofs makes clear, he thinks that explanations should not be understood in terms of anything psychological. Compare: If an argument is logically valid, then the conclusion was to be expected given the premises; but this doesn’t mean that logical validity is psychological in any important sense.

17 For further discussion of Peirce on explanation and how Peirce’s views relate to Hempel’s, see Niiniluoto (Reference Niiniluoto and Fetzer2000).

18 By “explanation” we mean “potential explanation” where H is a potential of E precisely when H would explain E if H were true (and assuming E is true). This way of understanding potential explanations is standard but might need to be tweaked; see Lipton (Reference Lipton2004, 58–59) for helpful discussion. If tweaking is needed, Nonexplanations Worse would need tweaking as well.

19 This problem does not arise in the context of “systematic power” where this involves not just explanatory power, but also predictive and retrodictive power (Brössel Reference Brössel2015).

20 Schupbach and Sprenger (Reference Schupbach and Sprenger2011, 109) also require that H and E have unconditional probabilities less than 1 and greater than 0. They note, though, that this requirement isn’t needed to derive their results, and is imposed so as to make their “calculations and motivations much more elegant.”

21 Schupbach and Sprenger (Reference Schupbach and Sprenger2011) don’t always adhere to their restriction to potential explanations. For example, they claim that EXPOSS(E, H) is maximal (at 1) precisely when H entails E, and that EXPOSS(E, H) is minimal (at –1) precisely when H entails ∼E. However, their Potential-Explanation Restriction implies that there can be no cases where H entails ∼E and EXPOSS(E, H) = –1. Suppose that H entails ∼E. It follows that it is not the case that H would explain E if H were true. For, if H were true, then E would be false and so would not be explained by anything.

22 Potential-Explanation Restriction isn’t the only possible way to restrict PPMs. Here are two alternatives:

Potential-Explanation Restriction*. EXPO(E, H) is defined precisely when at least part of H is a potential explanation of at least part of E in that at least part of H would explain at least part of E if H were true.

Potential-Explanation Restriction**. EXPO(E, H) is defined precisely when there is a partition Γ such that (i) H is a member of Γ and (ii) at least one member of Γ is a potential explanation of E.

Each of these alternatives is more permissive than Potential-Explanation Restriction. However, they would lead to further problems for TLI measures.

23 The fact that Schupbach and Sprenger’s EXPOSS measure violates our Adding Irrelevance Sometimes Hurts condition will come as no surprise to those authors. We say this because Schupbach and Sprenger (Reference Schupbach and Sprenger2011) introduce the following as a condition of adequacy that any probabilistic measure of explanatory power must satisfy:

Irrelevant Conjunction. If (i) Pr(E&H2) = Pr(E)Pr(H2), (ii) Pr(H1&H2) = Pr(H1)Pr(H2), and (iii) Pr(E&H1&H2) = Pr(E&H1)Pr(H2), then EXPO(E, H1) = EXPO(E, H1&H2).

Schupbach and Sprenger don’t discuss Kyburg’s and Salmon’s influential examples. Rather, they motivate their Irrelevant Conjunction by appeal to the idea that explanatory power is a matter of surprise reduction, and they motivate that idea by appeal to Hempel, Peirce, and Glymour. We argued in section 4 that this latter appeal is problematic.

References

Brössel, Peter. 2015. “On the Role of Explanatory and Systematic Power in Scientific Reasoning.” Synthese 192 (12):3877–913.CrossRefGoogle Scholar
Cabrera, Frank. 2017. “Can There Be a Bayesian Explanationism? On the Prospects of a Productive Partnership.” Synthese 194 (4):1245–72.CrossRefGoogle Scholar
Cohen, Michael. 2015. “On Schupbach and Sprenger’s Measures of Explanatory Power.” Philosophy of Science 82 (1):97109.CrossRefGoogle Scholar
Cohen, Michael. 2016. “On Three Measures of Explanatory Power with Axiomatic Representations.” British Journal for the Philosophy of Science 67 (4):1077–89.CrossRefGoogle Scholar
Cohen, Michael. 2018. “Explanatory Justice: The Case of Disjunctive Explanation.” Philosophy of Science 85 (3):442–54.CrossRefGoogle Scholar
Colombo, Matteo. 2017. “Experimental Philosophy of Explanation Rising: The Case for a Plurality of Concepts of Explanation.” Cognitive Science 41 (2):503–8.CrossRefGoogle ScholarPubMed
Colombo, Matteo, Bucher, Leandra, and Sprenger, Jan. 2017. “Determinants of Judgments of Explanatory Power: Credibility, Generality, and Statistical Relevance.” Frontiers in Psychology 8:1430 (1-11).CrossRefGoogle ScholarPubMed
Crupi, Vincenzo, and Tentori, Katya. 2012. “A Second Look at the Logic of Explanatory Power (with Two Novel Representation Theorems).” Philosophy of Science 79 (3):365–85.CrossRefGoogle Scholar
Dobzhansky, Theodosius. 1973. “Nothing in Biology Makes Sense Except in the Light of Evolution.” American Biology Teacher 35 (3):125–29.CrossRefGoogle Scholar
Eells, Ellery, and Fitelson, Branden. 2002. “Symmetries and Asymmetries in Evidential Support.” Philosophical Studies 107 (2):129–42.CrossRefGoogle Scholar
Eva, Benjamin, and Stern, Reuben. 2019. “Causal Explanatory Power.” British Journal for the Philosophy of Science 70 (4):1029–50.CrossRefGoogle Scholar
Glymour, Clark. 1980. “Explanations, Tests, Unity and Necessity.” Nous 14 (1):3150.CrossRefGoogle Scholar
Glymour, Clark. 2015. “Probability and the Explanatory Virtues.” British Journal for the Philosophy of Science 66 (3):591604.CrossRefGoogle Scholar
Good, Irving. 1960. “Weight of Evidence, Corroboration, Explanatory Power, Information and the Utility of Experiments.” Journal of the Royal Statistical Society (Series B) 22 (2):319–31.Google Scholar
Hempel, Carl. 1965. “Aspects of Scientific Explanation.” In Aspects of Scientific Explanation and Other Essays in the Philosophy of Science, 331496. New York: Free Press.Google Scholar
Hempel, Carl. 2001. “Postrscript 1976: More Recent Ideas on the Problem of Statistical Explanations.” In The Philosophy of Carl G. Hempel: Studies in Science, Explanation, and Rationality, edited by Fetzer, James, 165–86. Oxford: Oxford University Press.Google Scholar
Huber, Franz. 2008. “Milne’s Argument for the Log-Ratio Measure.” Philosophy of Science 75 (4):413–20.CrossRefGoogle Scholar
Humphreys, Paul. 1989. The Chances of Explanation. Princeton: Princeton University Press.Google Scholar
Jeffrey, Richard. 1971. “Statistical Explanation vs. Statistical Inference.” In Statistical Explanation and Statistical Relevance, edited by Salmon, Wesley, 1928. Pittsburgh: University of Pittsburgh Press.CrossRefGoogle Scholar
Joyce, James. 2019. “Bayes’ Theorem.” In The Stanford Encyclopedia of Philosophy, edited by Edward Zalta. Stanford: Stanford University Press. https://plato.stanford.edu/archives/spr2019/entries/bayes-theorem/.Google Scholar
Kyburg, Henry. 1965. “Salmon’s Paper.” Philosophy of Science 32 (2):147–51.CrossRefGoogle Scholar
Lipton, Peter. 2004. Inference to the Best Explanation.2nd ed. London: Routledge.Google Scholar
Mayr, Ernst. 1961. “Cause and Effect in Biology.” Science 134 (3489):1501–6.CrossRefGoogle ScholarPubMed
Niiniluoto, Ilkka. 2000. “Hempel’s Theory of Statistical Explanation.” In Science, Explanation, and Rationality: Aspects of the Philosophy of Carl G. Hempel, edited by Fetzer, James, 138–63. Oxford: Oxford University Press.Google Scholar
Northcott, Robert. 2013. “Degree of Explanation.” Synthese 190 (15):3087–105.CrossRefGoogle Scholar
Peirce, Charles. 1931–35. Collected Papers of Charles Sanders Peirce. Vols. 1–6, edited by Hartshorne, Charles and Weiss, Paul. Cambridge, MA: Harvard University Press.Google Scholar
Peirce, Charles. 1958. Collected Papers of Charles Sanders Peirce. Vols. 7–8, edited by Burks, Arthur. Cambridge, MA: Harvard University Press.Google Scholar
Popper, Karl. 1954. “Degree of Confirmation.” British Journal for the Philosophy of Science 5 (18):143–49.CrossRefGoogle Scholar
Roche, William, and Shogenji, Tomoji. 2014. “Dwindling Confirmation.” Philosophy of Science 81 (1):114–37.CrossRefGoogle Scholar
Roche, William, and Sober, Elliott. 2021. “Disjunction and Distality: The Hard Problem for Purely Probabilistic Causal Theories of Mental Content.” Synthese 198 (8):7197–230.CrossRefGoogle Scholar
Russell, Bertrand. 1919. Introduction to Mathematical Philosophy. London: George Allen & Unwin.Google Scholar
Salmon, Wesley. 1971. “Statistical Explanation.” In Statistical Explanation and Statistical Relevance, edited by Salmon, Wesley, 2987. Pittsburgh: University of Pittsburgh Press.CrossRefGoogle Scholar
Salmon, Wesley. 1984. Scientific Explanation and the Causal Structure of the World. Princeton: Princeton University Press.Google Scholar
Salmon, Wesley. 2006. Four Decades of Scientific Explanation. Pittsburgh: University of Pittsburgh Press.CrossRefGoogle Scholar
Schupbach, Jonah. 2011. “Comparing Probabilistic Measures of Explanatory Power.” Philosophy of Science 78 (5):813–29.CrossRefGoogle Scholar
Schupbach, Jonah. 2017. “Inference to the Best Explanation, Cleaned Up and Made Respectable.” In Best Explanations: New Essays on Inference to the Best Explanation, edited by McCain, Kevin and Poston, Ted, 3961. Oxford: Oxford University Press.Google Scholar
Schupbach, Jonah, and Sprenger, Jan. 2011. “The Logic of Explanatory Power.” Philosophy of Science 78 (1):105–27.CrossRefGoogle Scholar
Shogenji, Tomoji. 2003. “A Condition for Transitivity in Probabilistic Support.” British Journal for the Philosophy of Science 54 (4):613–16.CrossRefGoogle Scholar
Sober, Elliott. 1999. “The Multiple Realizability Argument against Reductionism.” Philosophy of Science 66 (4):542–64.CrossRefGoogle Scholar
Sober, Elliott. 2003. “Two Uses of Unification.” In The Vienna Circle and Logical Empiricism, edited by Stadler, Friedrich, 205–16. Dordrecht: Kluwer.CrossRefGoogle Scholar
Ströing, Pascal. 2018. “Data, Evidence, and Explanatory Power.” Philosophy of Science 85 (3):422–41.CrossRefGoogle Scholar
Ylikoski, Petri, and Kuorikoski, Jaakko. 2010. “Dissecting Explanatory Power.” Philosophical Studies 148 (2):201–19.CrossRefGoogle Scholar
Zalabardo, José. 2009. “An Argument for the Likelihood-Ratio Measure of Confirmation.” Analysis 69 (4):630–35.CrossRefGoogle Scholar
Figure 0

Table 1. The probability of Jones’s having lung cancer conditional on four conjunctions