1. Introduction
There are clear cases in which two hypotheses each explain an observation, but one of them does so with greater explanatory power (or strength). Here’s one example:
Suppose that a randomized control trial is performed to see how the incidence of heart disease is affected by eliminating fried foods from one’s diet, and that the result is that the reduction is 10%. Now consider two causal models that might explain why this is the outcome. The first says that the reduction should have been between 9% and 11% while the second says that the reduction should have been between 2% and 20%. Suppose that both models are theoretically well-grounded, and that each is explanatorily relevant to the observed reduction. We think that there is a clear sense in which the first model has greater explanatory power than the second with respect to the observed reduction. Each model explains the observed reduction to some extent, but the explanation provided by the first is stronger than the one provided by the second.
Here’s another:
Suppose Jones smoked cigarettes and inhaled asbestos, and that both of these causally contributed to her getting lung cancer, where a person’s risk of lung cancer is greater if she inhales both carcinogens than if she inhales just one. Jones’s getting lung cancer is explained by her smoking cigarettes, by her inhaling asbestos, and by her doing both, but the smoking + asbestos explanation is more powerful than the smoking explanation is and it’s also more powerful than the asbestos explanation is.
We’ll consider other examples in what follows, but these two suffice to set the stage for our main question: Is there a plausible mathematical measure of the power with which a hypothesis H explains an explanadum E?
That the answer is affirmative has prima facie plausibility. Each of the previous examples can be fleshed out so that the different explanans statements have different probabilities and confer different probabilities on their explananda. Perhaps these probabilistic differences ground differences in explanatory power, and perhaps more generally explanatory power can be measured in terms of probabilities.
The idea of a mathematical measure of explanatory power carries baggage that more modest ideas about explanatory power do not shoulder. Here is a less demanding question:
For any H1, H2, and O, if H1 and H2 each explains O, is there a plausible criterion that determines whether H1’s explanation of O is more powerful than H2’s?
This is the question of whether there is a criterion for comparative explanatory power, where both hypotheses address the same observation. No numbers are involved. A measure of explanatory power does much more. For example, it would describe whether the power of H1’s explanation of O1 exceeds the power of H2’s explanation of O2, and by how much. That may sound like a paradigmatic example of apples and oranges, but that’s what a mathematical measure delivers. If you think that the power of an explanation has multiple orthogonal dimensions, with the result that H1 may be a more powerful explanation of O than H2 is on some dimensions, while the reverse is true on others (as do Ylikoski and Kuorikoski Reference Ylikoski and Kuorikoski2010), you may think that the quest for such a measure is quixotic.
That said, several able philosophers have dared to plunge ahead. Are they tilting at windmills? One such attempt is Schupbach and Sprenger’s (Reference Schupbach and Sprenger2011). They set out several conditions of adequacy on measures of explanatory power, which they use to derive the following measure:
EXPOSS is a purely probabilistic measure of explanatory power (a PPM, for short) in that the value of EXPOSS(E, H) is fully determined by probabilities.Footnote 1, Footnote 2
Schupbach and Sprenger’s measure is not the only game in town. There are many—indeed, very many—purely probabilistic measures of explanatory power in logical space in addition to EXPOSS. Here are four:
EXPOCT is defended in Crupi and Tentori (Reference Crupi and Tentori2012). EXPOG is defended in Good (Reference Good1960).Footnote 3 EXPOC and EXPOD are mentioned but not defended in Schupbach (Reference Schupbach2011). In what follows, we’ll use “The Five” to refer to these five measures.Footnote 4
No two of The Five are ordinally equivalent, but they all satisfy the following two conditions, each of which involves two likelihood inequalities:
TLI-1. If (a) Pr(E | H1) > Pr(E | H2) and (b) Pr(E | ∼H1) < Pr(E | ∼H2), then EXPO(E, H1) > EXPO(E, H2).
TLI-2. If (a) Pr(E | H1) ≥ Pr(E | H2) and (b) Pr(E | ∼H1) ≤ Pr(E | ∼H2), then EXPO(E, H1) ≥ EXPO(E, H2).
EXPOSS, EXPOCT, EXPOG, and EXPOD meet these conditions because they meet the following stronger conditions, each of which involves one likelihood inequality:
OLI-1. If Pr(E | H1) > Pr(E | H2), then EXPO(E, H1) > EXPO(E, H2).
OLI-2. If Pr(E | H1) ≥ Pr(E | H2), then EXPO(E, H1) ≥ EXPO(E, H2).
Neither OLI-1 nor OLI-2 is satisfied by EXPOC because this measure says that H’s explanatory power with respect to E depends not just on Pr(E | H) but also on Pr(E | ∼H).Footnote 5
Our focus in what follows will be on “TLI” measures, defined as purely probabilistic measures of explanatory power that meet both TLI-1 and TLI-2. These include each of The Five along with many additional measures. Consider, for example, the following:
This measure and measures ordinally equivalent to it have been discussed in the context of confirmation (as a measure of the degree to which E incrementally confirms H),Footnote 6 but not, as far as we are aware, in the context of explanatory power. But, as EXPOLR meets both TLI-1 and TLI-2, it is included in our target class of purely probabilistic measures of explanatory power. The question is whether any TLI measure is adequate.
We want to emphasize that the issue here is not about inference. For example, in the cancer case described at the outset, we aren’t claiming that the smoking + asbestos explanation is more probable than the smoking explanation. It isn’t. The claim, rather, is that the smoking + asbestos explanation is stronger than the smoking explanation in that it sheds more light on why Jones has lung cancer. Both explanations shed light on why Jones has lung cancer, but the first does this to a greater extent.
The remainder of this article is organized as follows. In section 2, we present the Problem of Temporal Shallowness. In section 3, we introduce the Problem of Negative Causal Interactions. In section 4, we consider a reply that might be made to the objections we presented in the previous two sections. In section 5, we describe the Problem of Nonexplanations. In section 6, we consider a possible reply to what we’ve said about that problem. In section 7, we discuss the Problem of Explanatory Irrelevance. In section 8, we conclude.
2. The problem of temporal shallowness
Consider a causal chain from a distal cause Cd, to a proximate cause Cp, and then to an effect E. For example, you throw a baseball, which causes the baseball to hit a window, which causes the window to break. Someone then asks: Why did the window break? Consider two answers:
-
(2.1) A baseball hit the window and was going 65 mph.
-
(2.2) You threw a baseball, which caused the baseball to hit the window at 65 mph.
Notice that proposition (2.2) entails proposition (2.1) but not conversely. Explanation (2.1) is temporally shallow, in that it describes only the more proximate cause. Explanation (2.2) has more temporal depth, in that it traces the effect back to a proximate cause, and traces that proximate cause back to your dastardly deed. This is a case, we suggest, in which greater temporal depth makes for greater explanatory power: Explanation (2.2) is stronger than explanation (2.1).
The baseball example is mundane, but the point of the example applies to scientific examples as well. Evolutionary biologists and philosophers of biology routinely distinguish “proximate” from “ultimate” explanations. Ernst Mayr (Reference Mayr1961) used this terminology to distinguish two ways of explaining the phenotypes of organisms. For example, why do sunflowers turn toward the sun? A proximate explanation of that behavior would describe the mechanisms inside individual sunflowers that cause phototropism. An ultimate explanation would describe the evolutionary processes, including natural selection, that caused that trait to increase in frequency and remain at high frequency in the lineage leading to present day sunflowers. As Mayr (Reference Mayr1961, 1503) says, “[t]here is always a proximate set of causes and an ultimate set of causes; both have to be explained and interpreted for a complete understanding of the given phenomenon.” Mayr opposes temporal shallowness, but he also rejects the idea that only evolutionary considerations should be mentioned in explaining the present phenotypes of organisms.Footnote 7
To put the sunflower example into the baseball format, consider these two explanations of why sunflowers now turn toward the sun:
-
(2.3) Sunflowers now have mechanism M inside their cells (where M has features F1, F2, …, Fn).
-
(2.4) The ancestors of present day sunflowers were subject to selection for phototropism with the result that genes (G1, G2, …, Gm) that cause sunflowers to develop mechanism M (where M has features F1, F2, …, Fn) increased in frequency. These genes were passed down the generations, and present-day sunflowers inherited them, with the result that they have mechanism M.
What goes for the broken window also goes for the sunflowers. Propositions (2.3) and (2.4) are both explanatory, but we suggest that (2.4) has more explanatory power than (2.3).
Some readers may balk at this last, absolute, pronouncement. Perhaps they believe that what counts as the more powerful explanation depends on your interests. Sometimes all you want is the temporally shallow cause; at other times you want more. Neither is better than the other unconditionally. We sympathize with that pluralistic claim, but we do feel the tug of the idea that temporally more inclusive explanations are often more explanatorily powerful than explanations that are temporally less inclusive (Sober Reference Sober1999).
How do TLI measures address the question of temporal shallowness? Are some or all of them pluralistic? Consider two types of cases in which there is a causal chain from X to Y to Z (or, strictly speaking, from the events described by X to the events described by Y to the events described by Z):
Type 1 (i) X causes Y, (ii) Y causes Z, (iii) X and Y are logically independent of each other, (iv) Y and Z are logically independent of each other, (v) X increases the probability of Y, (vi) Y increases the probability of Z, and (vii) X is screened-off from Z both by Y and by ∼Y in that Pr(Z | Y&X) = Pr(Z | Y) and Pr(Z | ∼Y&X) = Pr(Z |∼Y).
Type 2 (i) X causes Y, (ii) Y causes Z, (iii) X entails and increases the probability of Y, (iv) Y increases the probability of Z, and (v) X is screened-off from Z by Y in that Pr(Z | Y&X) = Pr(Z | Y).
We show in Appendix 1 that TLI measures are heavily biased in favor of temporally shallow explanations in the following sense:
Theorem 1 In cases of Type 1, TLI measures entail that EXPO(Z, Y) > EXPO(Z, X) and EXPO(Z, Y) ≥ EXPO(Z, X&Y).
Theorem 2 In cases of Type 2, TLI measures entail that EXPO(Z, Y) ≥ EXPO(Z, X) and EXPO(Z, Y) ≥ EXPO(Z, X&Y).Footnote 8
Surely, however, an adequate measure of explanatory power should allow EXPO(Z, Y) to be less than EXPO(Z, X) or less than EXPO(Z, X&Y) in causal chains of either kind.Footnote 9 In such causal chains, TLI measures say that explanatory power is never increased by adding information about distal causes. Ernst Mayr is not the only scientist who would scoff at that suggestion.
It might be objected that our argument here depends on a wholesale commitment to the thesis that a more complete explanation of E is always more powerful than a less complete explanation of E. Philosophers often argue against this thesis by pointing out that more complete explanations are sometimes harder to understand and work with, and that less complete explanations sometimes apply to more explananda (see, e.g., Ylikoski and Kuorikoski [Reference Ylikoski and Kuorikoski2010, 216–17]). We agree with both these points, but they aren’t relevant to the thesis we want to endorse. We’re interested in whether more complete explanations ever have greater explanatory power than less complete explanations of the same explanandum when both are easy to grasp. In addition, L is not shown to be a better explanation of E1 than M is by the fact that L explains E1, E2, …, En whereas M explains only E1.Footnote 10 In any event, the examples we have discussed in this section do not fall in the cracks that Ylikoski and Kuorikoski’s observations reveal. Neither of the explanations of window breaking that we described strains our cognitive capacities to the breaking point, and the same point holds of the two explanations we described of sunflower phototropism.
3. The problem of negative causal interactions
At the start of this article, we considered an example in which asbestos exposure and smoking cigarettes both causally promote lung cancer. We assumed that lung cancer goes up in probability if an individual smokes, if an individual inhales asbestos, and if an individual does both, with lung cancer being more probable if both causes are present than it would be if there were just one.
Here we tweak the example, as shown in Table 1. Suppose that c, a, and i are positive, and that i > c and i > a. Now there is a causal interaction between smoking and asbestos; the model is not additive (which is the condition that obtains precisely when the interaction term i=0). In particular, suppose the interaction is negative. Although smoking cigarettes increases the probability of lung cancer if Jones doesn’t inhale asbestos, and inhaling asbestos increases the probability of lung cancer if Jones doesn’t smoke, smoking cigarettes decreases the probability of lung cancer if Jones inhales asbestos, and inhaling asbestos decreases the probability of lung cancer if Jones smokes. Somehow if Jones inhales cigarette smoke and asbestos particles, each inhibits the other’s carcinogenic propensities.Footnote 11
Pr(Jones has lung cancer | –) | ||
---|---|---|
Jones inhales asbestos | Jones does not inhale asbestos | |
Jones smokes cigarettes | x + c + a – i | x + a |
Jones does not smoke cigarettes | x + c | x |
Now suppose Jones has lung cancer (this is the explanandum E) and each of the following three hypotheses is true:
H1: Jones smoked cigarettes.
H2: Jones inhaled asbestos.
H1&H2: Jones smoked cigarettes and Jones inhaled asbestos.
Jones’s cancer is explained by each of these hypotheses, but H1 and H2 provide censored pictures of the causal facts whereas H1&H2’s explanation is more causally complete. H1 and H2 each resemble the one-sided story presented by the prosecutor in a trial who neglects to mention details that would cast doubt on the defendant’s guilt. Doing so may be rhetorically powerful, but it is evidentially inferior to a fuller accounting that mentions both the bits of evidence that favor guilt and the bits that favor innocence. So it is with explanation—powerful explanations should describe both the causal promoters and the causal inhibitors that impacted the explanandum event. Removing all the latter reduces explanatory power.Footnote 12
How do TLI measures handle cases like that of Jones? Consider, more generally, cases of the following type:
Type 3 (i) Y causes X, (ii) Z causes X, and (iii) there’s a negative causal interaction in that (a) Pr(X | Y) is greater than Pr(X), (b) Pr(X | Z) is greater than Pr(X), (c) Pr(X | ∼Y&∼Z) is less than Pr(X | Y&Z), and (d) the latter is less than each of Pr(X | Y&∼Z) and Pr(X | ∼Y&Z).
We show in Appendix 2 that TLI measures are heavily biased in favor of censored explanations in that:
Theorem 3 In cases of Type 3, TLI measures entail that EXPO(X, Y) > EXPO(X, Y&Z) and EXPO(X, Z) > EXPO(X, Y&Z).
The upshot is that each of The Five, and all other TLI measures, entail an erroneous interpretation of negative causal interactions of the kind we’ve described. They all say that a hypothesis that describes both positive and negative causal factors has less explanatory power than a hypothesis that describes the positive and ignores the negative.Footnote 13, Footnote 14
We note that this point about negative causal interactions involves no commitment to the general thesis that more complete explanations are always more powerful. What we said at the end of the previous section about window-breaking and phototropism also applies to the present example of cigarettes and asbestos.
4. A possible reply to the temporal shallowness and the negative causal interactions problems
The arguments offered in the previous two sections concerning temporal shallowness and negative causal interactions could be defused if there were an important sense of explanatory power that demands the verdicts that TLI measures deliver on those topics. A possible defusion of this type can be found in the motivation that Schupbach and Sprenger (Reference Schupbach and Sprenger2011) offer for their proposed measure of explanatory power. They begin with the idea that there’s a sense of explanatory power on which explanatory power is a matter of surprise reduction. They further hold that surprise reduction is psychological and so should be understood in terms of actual or rational degrees of belief. They defend this idea by arguing that their interpretation of explanatory power “finds precedence in many classical discussions of explanation”:
Perhaps its clearest historical expression occurs when Peirce (Reference Peirce, Hartshorne and Weiss1931–35, 5.189) identifies the explanatoriness of a hypothesis with its ability to render an otherwise “surprising fact” as “a matter of course.” This sense of explanatory power may also be seen as underlying many of the most popular accounts of explanation. Most obviously, Deductive-Nomological and Inductive-Statistical accounts (Hempel Reference Hempel1965) and necessity accounts (Glymour Reference Glymour1980) explicitly analyze explanation in such a way that a theory that is judged to be explanatory of some explanandum will necessarily increase the degree to which we expect that explanandum. (Schupbach and Sprenger Reference Schupbach and Sprenger2011, 108)
We find this dubious, especially when it comes to Hempel’s D-N and I-S models. First, it’s clear from passages such as the following that Hempel sees his D-N and I-S models in terms of a nonpsychological, or objective, conception of explanation:
Scientific research seeks to give an account—both descriptive and explanatory—of empirical phenomena which is objective in the sense that its implications and its evidential support do not depend essentially on the individuals who happen to apply or to test them. This ideal suggests the problem of constructing a nonpragmatic conception of scientific explanation—a conception that requires reference to puzzled individuals no more than does the concept of mathematical proof. And it is this nonpragmatic conception of explanation with which the two covering-law models are concerned. (Hempel Reference Hempel and Fetzer2001, 82)
Second, in Hempel’s D-N model of explanation, what matters (roughly) is that the explanans propositions entail the explanandum proposition. Your degrees of belief (normative or actual) play no role, much less the idea that learning the explanans must change your degrees of belief in the explanandum. Even if it’s true that you increase your degree of belief in E when you use Hempel’s D-N model and conclude that H explains E, it doesn’t follow that Hempel’s model asserts that explanatory power is a matter of increase in degree of belief. Compare: Even if it’s true that you become excited when you use Hempel’s D-N model and conclude that H explains E, it doesn’t follow that Hempel’s model asserts that explanatory power is a matter of becoming excited.Footnote 15, Footnote 16
What about Peirce (Reference Peirce, Hartshorne and Weiss1931–35, Reference Peirce and Burks1958)? It’s true that the passage from Peirce that Schupbach and Sprenger cite includes the expressions “surprising fact” and “a matter of course.” However, the main subject of that passage is abduction, not the question of what an explanation is. Peirce thinks that when H is abductively inferred (or conjectured) from E, H explains E and reduces E’s surprisingness, but it doesn’t follow that he takes this to be essential to explanation. In fact, there are many passages where Peirce characterizes explanation with no appeal to surprise reduction or to an agent’s degrees of belief. One such passage is cited by Schupbach and Sprenger (Reference Schupbach and Sprenger2011, 108, n. 5). They quote Peirce as saying that “to explain a fact is to show that it is a necessary or, at least, a probable result from another fact, known or supposed.” This sounds like Hempel but without all the formal niceties.Footnote 17
In addition to appealing to the precedents of Hempel and Peirce, Schupbach and Sprenger cite Glymour (Reference Glymour1980) as a precursor of the idea of explanatory power as surprise reduction. Glymour’s basic idea is that when an explanans is deduced from necessary truths involving property identities, the resulting explanans shows that the explanandum is necessary rather than contingent. This is a contingency-reducing explanation, which Glymour doesn’t think of psychologistically.
Schupbach and Sprenger (Reference Schupbach and Sprenger2011, 108) also appeal to statistical practice, saying that the surprise-reduction concept of explanatory power “dominates statistical reasoning in which scientists are ‘explaining away’ surprise in the data by means of assuming a specific statistical model (e.g., in the omnipresent linear regression procedures).” Our response is that regression procedures are part of the frequentist toolkit, and frequentists have no truck with degrees of belief.
In summary, the historical precedents that Schupbach and Sprenger (Reference Schupbach and Sprenger2011) note do not lend credence to the idea that there is an important sense of “explanatory power” according to which explanatory power should be understood in terms of surprise reduction where the latter concept is understood in terms of the changes in credence that an agent makes or ought to make. This leaves it open that their starting idea can be motivated in some other way. We now turn to our third argument against TLI measures of explanatory power.
5. The problem of nonexplanations
Take some case of Type 1 (as described in section 2)—that is, an example in which X causes and increases the probability of Y, Y causes and increases the probability of Z, X and Y are logically independent of each other, Y and Z are logically independent of each other, and X is screened-off from Z both by Y and by ∼Y. Is X’s explanatory power with respect to Y greater than Z’s explanatory power with respect to Y? The answer is clearly yes. Because X is a cause of Y whereas Z is an effect of Y, X explains Y at least to some extent, whereas Z explains Y not at all.
This simple example in which an explanation is contrasted with a nonexplanation suggests a condition of adequacy on measures of explanatory power:
Nonexplanations Worse. If (i) X is an explanation of E in that X explains E at least to some extent and (ii) Y is a nonexplanation with respect to E in that Y explains E not at all, then EXPO(E, X) > EXPO(E, Y).Footnote 18
Nonexplanations have no explanatory power. Explanations have some explanatory power. Hence nonexplanations have less explanatory power than explanations.
TLI measures fail to satisfy Nonexplanations Worse. It’s straightforward to construct cases of Type 1, for example, in which:
We describe an example of this sort in Appendix 3. Because X is an explanation of Y whereas Z is not an explanation of Y, Nonexplanations Worse implies that EXPO(Y, X) > EXPO(Y, Z). However, given (5.1) and (5.2), TLI measures deliver the result that EXPO(Y, X) < EXPO(Y, Z).Footnote 19
Consider, for comparison, the following condition of adequacy on probabilistic measures of confirmation in the sense of increase in probability:
Nonconfirmers Worse. If (i) E confirms X in that E increases the probability of X at least to some extent and (ii) E is a nonconfirmer with respect to Y in that E decreases or has no impact on the probability of Y, then CONF(X, E) > CONF(Y, E).
This condition, qua condition of adequacy, is uncontroversial in Bayesian confirmation theory. The situation with respect to explanatory power seems similar: just as confirmers confirm to a greater degree than nonconfirmers, explanations explain more powerfully than nonexplanations.
6. A possible reply to the problem of nonexplanations
We have been understanding PPMs as unrestricted in the sense that the EXPO(E, H) is well defined even if H is not an explanation of E. The reason why is straightforward: Some of our beliefs about explanatory power concern nonexplanations, and we want to know how well PPMs do in terms of all our considered judgments about explanatory power. In contrast, Schupbach and Sprenger (Reference Schupbach and Sprenger2011) restrict EXPOSS, along with rival PPMs, as follows:
Potential-Explanation Restriction. EXPO(E, H) is defined precisely when H is a potential explanation of E in that H would explain E if H were true.Footnote 20
Given this way of understanding PPMs, TLI measures do not deliver the result that EXPO(Y, X) < EXPO(Y, Z) in our example of a causal chain from X to Y to Z because Z isn’t a potential explanation of Y, and so EXPO(Y, Z) isn’t defined.Footnote 21 The restriction allows TLI measures to avoid generating false claims about explanatory power. But these theories still fail to satisfy Nonexplanations Worse because they fail to deliver some manifestly true propositions about explanatory power.Footnote 22 The result is that the measures are incomplete, not that they entail falsehoods.
We argued in the previous section that it is a condition of adequacy on a quantitative measure of explanatory power that it judge nonexplanations as having less explanatory power than genuine explanations. This condition of adequacy of course obviously clashes with the restriction that Schupbach and Sprenger embrace. We hope the following comments will help adjudicate.
It is one thing for a mathematical criterion to fail to generate a result, as when the ratio x/y fails to be defined when y=0. It is another thing for a mathematical criterion to entail a result, which friends of the criterion wish to disown. Schupbach and Sprenger will understandably dislike what their mathematical criterion says about the causal chain from X to Y to Z discussed in section 5, but to stipulate that their criterion does not apply to such cases is not a legitimate solution. Where do such stipulations stop? Will friends of TLI measures of explanatory power urge that their theories do not apply to cases involving temporal shallowness or to cases involving negative causal factors? We are not saying that they are inclined to do so. Our question concerns the ground rules for using restrictive stipulations to avoid apparent counterexamples.
An anonymous referee for this journal suggests that friends of TLI measures of explanatory power could set aside the restriction strategy and instead simply stipulate that if H is not an explanation of E, then EXPO(E, H) has the lowest possible value. Consider, once more, the example of Schupbach and Sprenger’s proposed measure. EXPOSS(E, H) might be understood in terms of a function from inputs to outputs such that:
(i) If the input consists of (a) explanandum E and (b) hypothesis H where H is not an explanation of E, then the output is –1.
(ii) If the input consists of (a) explanandum E, (b) hypothesis H where H is an explanation of E, and (c) probability function Pr(–), then the output is a number equal to ${{\Pr ({\rm{H}}\;{\rm{|E}}) - \Pr ({\rm{H}}\;{\rm{|}}\sim{\rm{E}})} \over {\Pr ({\rm{H}}\;{\rm{|E}}) + \Pr ({\rm{H}}\;{\rm{|}}\sim{\rm{E}})}}$ .
The same referee further notes that this suggestion is similar to the fact that the likelihood ratio measure of confirmation—CONF(H, E) = ${{\Pr ({\rm{E}}\;{\rm{|H}})} \over {\Pr ({\rm{E}}\;{\rm{|}}\sim{\rm{H}})}}$ —is standardly understood so that CONF(H, E) is maximal at ∞ when Pr(E | H)>0 and Pr(E | ∼H)=0, even though, strictly speaking, division by 0 is undefined. There is an important difference here, however. Take some value for Pr(E | H) greater than Pr(E | ∼H). The limit of ${{\Pr ({\rm{E}}\;{\rm{|H}})} \over {\Pr ({\rm{E}}\;{\rm{|}}\sim{\rm{H}})}}$ approaches positive infinity if Pr(E | H) is held fixed and Pr(E | ∼H) is made smaller and smaller. This is the rationale behind the standard way of understanding the likelihood ratio measure. However, nothing about surprise reduction or ${{\Pr ({\rm{H}}\;{\rm{|E}}) - \Pr ({\rm{H}}\;{\rm{|}}\sim{\rm{E}})} \over {\Pr ({\rm{H}}\;{\rm{|E}}) + \Pr ({\rm{H}}\;{\rm{|}}\sim{\rm{E}})}}$ justifies the stipulation put forward in (i). Rather, what motivates (i) is the idea that nonexplanations have no explanatory power. We of course agree with that idea, but we think it’s reasonable to expect that a mathematical measure should capture that idea without recourse to a special stipulation.
We think our criticism of the referee’s suggested stipulation can be further motivated by the following “continuity argument.” Consider a sequence of hypotheses H1, H2, etc. where each entry has less explanatory power than the one before, relative to a fixed explanandum proposition E. According to Schupbach and Sprenger’s measure, these hypotheses can be ranked for their explanatory power just by looking at the value of Pr(E|Hi)—the bigger this is, the more explanatory it is, with the result that Pr(E|BestPossible)=1, and Pr(E|worst possible) = 0. The point of relevance here is that numerous nonexplanations are nowhere “close” to Pr(E|WorstPossible) = 0. Consider, for instance, some hypothesis H about the present that deductively retrodicts E. Although H fails to explain E (assuming that the past isn’t explained by the present), the fact remains that Pr(E | H)=1, which is also the value of Pr(E|BestPossible). Here the reduction in E’s surprisingness is maximal, and yet the referee’s suggested stipulation puts H in the same camp as hypotheses that entail ∼E.
In summary, Schupbach and Sprenger (Reference Schupbach and Sprenger2011) stipulate that their measure applies only to potential explanations. The referee mentioned suggests that if this stipulation fails to satisfy, proponents of TLI measures can stipulate instead that a nonexplanation has the minimum possible degree of explanatory power. These are different stipulations, but there is a common flaw, or so we have suggested. The use of stipulations in either of these two ways reminds us of Russell’s (Reference Russell1919, 71) famous quip that “[t]he method of ‘postulating’ what we want has many advantages; they are the same as the advantages of theft over honest toil.”
7. The problem of explanatory irrelevance
In the previous two sections we discussed the dichotomy between an explanation of E and a nonexplanation of E. We now want to consider something that is a matter of degree. If H is a genuine explanation of E, and you conjoin explanatorily irrelevant proposition I to H, does the resulting conjunction, H&I, provide a less powerful explanation of E than H provides?
A related question has a distinguished history in philosophical discussions of explanation. Consider two influential examples:
This sample of table salt dissolves in water, for it had a dissolving spell cast upon it, and all samples of table salt that have had dissolving spells cast on them dissolve in water. (Kyburg Reference Kyburg1965, 147)
John Jones avoided becoming pregnant during the past year, for he has taken his wife’s birth control pills regularly, and every man who regularly takes birth control pills avoids pregnancy. (Salmon Reference Salmon and Salmon1971, 34)
Kyburg and Salmon present these examples to illustrate how a genuine explanation can be turned into a nonexplanation by adding information that is explanatorily irrelevant. They seem to be committed to the general thesis that if H contains even one detail that is explanatorily irrelevant to E, then H is not an explanation of E. We think this may go too far.
Shifting from the question of whether a hypothesis explains to the question of how powerfully it does so, our modest thesis about irrelevant additions is this:
Adding Irrelevance Sometimes Hurts. EXPO(E, H) is sometimes greater than EXPO(E, H&I) when (a) I is explanatorily irrelevant to E and (b) I is probabilistically independent of E, H, and E&H.
Salmon and Kyburg appear to be purists; for them adding even a teaspoon of irrelevancy to a sea of explanatorily relevant information is enough to turn the sea into a nonexplanation. We aren’t purists about such additions when the question is degree of explanatory power. That’s why we say sometimes. Here we have in mind examples in which a brief, telling H is conjoined with a giant, complicated, and doubly irrelevant proposition I.
How do TLI measures fair in terms of the Adding Irrelevance Sometimes Hurts condition? The situation is mixed, in that some TLI measures meet Adding Irrelevance Sometimes Hurts whereas others do not. We show in Appendix 4 that Adding Irrelevance Sometimes Hurts is met by EXPOC and EXPOLR but not by EXPOSS, EXPOCT, EXPOG, or EXPOD.Footnote 23
TLI measures would satisfy the Adding Irrelevance Sometimes Hurts condition if they were understood so that H&I fails to explain E when I is explanatorily irrelevant to E, and if nonexplanations are stipulated to have the minimum possible degree of explanatory power. However, this way of understanding those measures leads back to our earlier discussion of the role of stipulation in the formulation of PPMs of explanatory power.
8. Concluding comments
We have focused on TLI measures of explanatory power, which include all extant purely probabilistic measures of explanatory power. All TLI measures face three problems—the Problem of Temporal Shallowness, the Problem of Negative Causal Interactions, and the Problem of Nonexplanations. Some TLI measures face a fourth problem—the Problem of Explanatory Irrelevance. These problems together raise the suspicion that an adequate measure of explanatory power can’t be purely probabilistic. However, the door has not slammed shut absolutely. It is left open that an adequate PPM of explanatory power can be found; the place to search, we suggest, is in measures that violate the TLI constraints.
Acknowledgments
We thank the three anonymous referees for the journal, and also the journal’s editorial team, for helpful comments on prior versions of the paper.
Appendix 1
Proof of Theorem 1
The aim is to show:
Theorem 1 In cases of Type 1, TLI measures entail that EXPO(Z, Y) > EXPO(Z, X) and EXPO(Z, Y) ≥ EXPO(Z, X&Y).
Consider a case of Type 1 in which (i) X causes Y, (ii) Y causes Z, (iii) X and Y are logically independent of each other, (iv) Y and Z are logically independent of each other, (v) X increases the probability of Y, (vi) Y increases the probability of Z, and (vii) X is screened-off from Z both by Y and by ∼Y in that Pr(Z | Y&X) = Pr(Z | Y) and Pr(Z | ∼Y&X) = Pr(Z | ∼Y). It follows from condition (vii) that (see Shogenji Reference Shogenji2003):
The fact that Pr(Z | ∼Y&X) = Pr(Z | ∼Y) implies that Pr(∼Y&X)>0, and so Pr(Y | X)<1. This, in turn, implies:
(A1.1) and (A1.2) together imply:
Next, note that condition (vii) implies:
The sum after the second equality sign in (A1.4) equals:
Because Pr(Z | Y) > Pr(Z), it follows that Pr(Z | Y) > Pr(Z | ∼Y). Given this, and given that Pr(Y | ∼X)>0 because X and Y are logically independent of each other, it follows that the first addend in (A1.5) is positive. Given this, and given that Pr(Z | ∼X) equals (A1.5), it follows that:
(A1.3), (A1.6), and TLI-1 together imply that EXPO(Z, Y) > EXPO(Z, X). It’s a theorem of the probability calculus that (for any propositions P1, P2, and P3):
Pr(Z | ∼(X&Y)) equals Pr(Z | ∼X∨∼Y)), and so by (A1.7) also equals:
Because Pr(Z | ∼Y&X) = Pr(Z | ∼Y), it follows that (A1.8) equals:
Given that Pr(Z | ∼Y) < Pr(Z | ∼X), it follows that (A1.9) is greater than:
This sum equals Pr(Z | ∼Y), and so:
Given this, and given that, as per condition (vii), Pr(Z | X&Y) = Pr(Z | Y), it follows by TLI-2 that EXPO(Z, Y) ≥ EXPO(Z, X&Y). Hence any measure meeting TLI-1 and TLI-2 implies that EXPO(Z, Y) > EXPO(Z, X) and EXPO(Z, Y) ≥ EXPO(Z, X&Y).
Proof of Theorem 2
The aim is to show:
Theorem 2 In cases of Type 2, TLI measures entail that EXPO(Z, Y) ≥ EXPO(Z, X) and EXPO(Z, Y) ≥ EXPO(Z, X&Y).
Consider a case of Type 2 in which (i) X causes Y, (ii) Y causes Z, (iii) X entails and increases the probability of Y, (iv) Y increases the probability of Z, and (v) X is screened-off from Z by Y in that Pr(Z | Y&X) = Pr(Z | Y). Given that X entails Y, it follows that X is logically equivalent to Y&X. Hence:
This and the fact that Pr(Z | Y&X) = Pr(Z | Y) imply:
The law of total probability implies:
Pr(Z | Y&X) = Pr(Z | Y) only if Pr(Z | Y&X) = Pr(Z | Y&∼X). So, since Pr(Z | Y&X) = Pr(Z | Y), the right-hand side of (A1.14) equals:
X entails Y, and so ∼Y is logically equivalent to ∼Y&∼X. Hence (A1.15) is equal to:
This simplifies to:
Since Pr(Z | Y) > Pr(Z), Pr(Z | Y) > Pr(Z | ∼Y). Hence, as Pr(Y | ∼X)≥0, the first addend in (A1.16) is nonnegative. Hence:
(A1.13), (A1.17), and TLI-2 together imply that EXPO(Z, Y) ≥ EXPO(Z, X). Hence any measure meeting TLI-2 implies that EXPO(Z, Y) ≥ EXPO(Z, X).
Appendix 2
The aim is to show:
Theorem 3 In cases of Type 3, TLI measures entail that EXPO(X, Y) > EXPO(X, Y&Z) and EXPO(X, Z) > EXPO(X, Y&Z).
Consider a case of Type 3 where (i) Y causes X, (ii) Z causes X, and (iii) there’s a negative causal interaction in that (a) Pr(X | Y) is greater than Pr(X), (b) Pr(X | Z) is greater than Pr(X), (c) Pr(X | ∼Y&∼Z) is less than Pr(X | Y&Z), and (d) the latter is less than each of Pr(X | Y&∼Z) and Pr(X | ∼Y&Z). Given that Pr(X | Y&Z) is less than Pr(X | Y&∼Z), it follows that:
Because Pr(X | Y&Z) is less than Pr(X | ∼Y&Z), it follows that:
∼(Y&Z) is logically equivalent to ∼Y∨∼Z, and so:
Given this, and given (A1.7) in Appendix 1, it follows that Pr(X | ∼(Y&Z)) equals each of the following:
(A2.4) simplifies to:
(A2.5) simplifies to:
Now focus on the differences in the second addends in (A2.6) and (A2.7). First, given that Pr(X | Y&∼Z) > Pr(X | Y&Z), it follows that:
Given this, given that Pr(X | Y) > Pr(X), and given that Pr(X | Y) > Pr(X) precisely when Pr(X | Y) > Pr(X | ∼Y), it follows that (A2.6) is greater than Pr(X | ∼Y). Hence:
Second, given that Pr(X | ∼Y&Z) > Pr(X | Y&Z), it follows that:
Given this, given that Pr(X | Z) > Pr(X), and given that Pr(X | Z) > Pr(X) precisely when Pr(X | Z) > Pr(X | ∼Z), it follows that (A2.7) is greater than Pr(X | ∼Z). Hence:
Given (A2.1), (A2.9), and any TLI measure, it follows that EXPO(X, Y) > EXPO(X, Y&Z). Given (A2.2), (A2.10), and any TLI measure, it follows that EXPO(X, Z) > EXPO(X, Y&Z).
Appendix 3
Take a causal chain from X to Y to Z, and consider the following probability distribution:
X | Y | Z | Pr |
---|---|---|---|
T | T | T | 1/4 |
T | T | F | 14669/211750 |
T | F | T | 3/14 |
T | F | F | 3/20 |
F | T | T | 4/39 |
F | T | F | 117352/4129125 |
F | F | T | 6/55 |
F | F | F | 21/275 |
This distribution is such that:
(A3.1) implies that X increases the probability of Y. (A3.2) implies that Y increases the probability of Z. The fact that none of the probabilities in the distribution equals 0 implies that X and Y are logically independent of each other, and that the same is true of Y and Z. (A3.3) and (A3.4) imply that X is screened-off from Z both by Y and by ∼Y. (A3.5) implies (5.1). (A3.6) implies (5.2). This is a case, then, of Type 1 where both (5.1) and (5.2) hold.
Appendix 4
Suppose that H explains and increases the probability of E, that I is explanatorily irrelevant to E, and that I is probabilistically independent of E, H, and E&H. Given this last supposition, it follows that:
Given this, in turn, it follows that EXPOSS(E, H) = EXPOSS(E, H&I), EXPOCT(E, H) = EXPOCT(E, H&I), EXPOG(E, H) = EXPOG(E, H&I), and EXPOD(E, H) = EXPOD(E, H&I). This follows because EXPOSS, EXPOCT, EXPOG, and EXPOD meet the following strengthened version of OLI-1:
Next, continuing to suppose that H explains and increases the probability of E, that I is explanatorily irrelevant to E, and that I is probabilistically independent of E, H, and E&H, consider the following probability distribution:
E | H | I | Pr |
---|---|---|---|
T | T | T | 25/38 |
T | T | F | 8/39 |
T | F | T | 1/199 |
T | F | F | 304/194025 |
F | T | T | 0 |
F | T | F | 0 |
F | F | T | 961323/9671798 |
F | F | F | 2563528/82719325 |
This distribution is such that:
Hence Adding Irrelevance Sometimes Hurts is met by EXPOC and EXPOLR but not by EXPOSS, EXPOCT, EXPOG, or EXPOD.