Hostname: page-component-55f67697df-sqlfs Total loading time: 0 Render date: 2025-05-08T11:53:32.370Z Has data issue: false hasContentIssue false

Accurate Updating

Published online by Cambridge University Press:  08 May 2025

Ginger Schultheis*
Affiliation:
University of Chicago Division of the Humanities, Chicago, IL, USA
Rights & Permissions [Opens in a new window]

Abstract

A number of authors have observed that epistemic externalists seem to face a dilemma: Either deny that Conditionalization is the rational update rule, thereby rejecting traditional Bayesian epistemology, or deny that the rational update rule maximizes expected accuracy, thereby rejecting accuracy-first epistemology. Call this the Bayesian Dilemma. I’m not convinced by this argument. Once we make the premises explicit, we see that it relies on assumptions the externalist rejects. In this paper, I argue that the Bayesian Dilemma is nevertheless a genuine dilemma. My argument does not make any assumptions that the externalist rejects.

Type
Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2025. Published by Cambridge University Press on behalf of the Philosophy of Science Association

1 Introduction

Accuracy-first epistemology aims to justify all epistemic norms by showing that they can be derived from the rational pursuit of accuracy. Take, for example, probabilism—the norm that credence functions should be probability functions. Accuracy-firsters say non-probabilistic credences are irrational because they’re accuracy-dominated: For every non-probabilistic credence function, there’s some probabilistic credence function that’s more accurate no matter what.Footnote 1 Or take norms of updating, my topic in this paper. Accuracy-firsters aim to derive the rational updating rule by way of accuracy; specifically, they claim that the rational updating rule is the rule that maximizes expected accuracy.Footnote 2

Externalism, put roughly, says that we do not always know what our evidence is. Though far from universally accepted, externalism is a persuasive and widely held thesis, supported by a compelling vision about the kinds of creatures we are—creatures whose information-gathering mechanisms are fallible, and whose beliefs about most subject matters are not perfectly sensitive to the facts.

Schoenfield (Reference Schoenfield2017) has shown that following the update rule Metaconditionalization maximizes expected accuracy.Footnote 3 However, as she and many other authors note, if externalism is true, Metaconditionalization is not Bayesian Conditionalization. Therefore, the externalist seems to face a dilemma: Either deny that Conditionalization is the rational update rule, thereby rejecting traditional Bayesian epistemology, or else deny that the rational update rule is the rule that maximizes expected accuracy, thereby rejecting the accuracy-first program. Call this the Bayesian Dilemma.Footnote 4

I’m not convinced by this argument. We’ll see that once we make the premises fully explicit, the argument relies on assumptions that the externalist should reject. Still, I think that the Bayesian Dilemma is a genuine dilemma. I give a new argument—I call it the continuity argument—that does not make any assumptions that the externalist rejects. Roughly, what I show is that if you’re sufficiently confident that you would follow Metaconditionalization if you adopted Metaconditionalization, then you’ll expect adopting a rule I’ll call Accurate Metaconditionalization to be more accurate than adopting Bayesian Conditionalization.

I’ll start in section 2 by introducing an accuracy-based framework for evaluating updating rules in terms of what I will call actual inaccuracy. In section 3, I’ll introduce externalism. In section 4, I turn to the Bayesian Dilemma. I present an argument purporting to show that the externalist must choose between Bayesian Conditionalization and accuracy-first epistemology, and I explain why the argument does not succeed. In section 5, I present the continuity argument showing that the Bayesian Dilemma is nevertheless a genuine dilemma. Section 6 concludes.

2 The accuracy framework: Actual inaccuracy

Accuracy-first epistemology says that our beliefs and credal states aim at accuracy, or closeness to the truth; that is, our beliefs and credal states aim to avoid inaccuracy, or distance from the truth. We said that, according to accuracy-firsters, the rational update rule is the rule that maximizes expected accuracy. There are different ways of making that thesis precise. In this section, I’ll present my own preferred way. We’ll start by getting the basics of the accuracy-first framework on the table.

2.1 Basics of the accuracy framework

For technical purposes, it is better to work with measures of inaccuracy rather than measures of accuracy. An inaccuracy measure I is a function that takes a world $$$$$$$$$ $w$ from a set of worlds ${\rm{\Omega }}$ , and a probability function $C$ defined over ${\cal P}\left( {\rm{\Omega }} \right)$ , and returns a number between 0 and 1. This number represents how inaccurate $C$ is in $w$ . $C$ is minimally inaccurate if it assigns 1 to all truths and 0 to all falsehoods; $C$ is maximally inaccurate if it assigns 1 to all falsehoods and 0 to all truths.

The expected inaccuracy of a probability function $C$ —relative to another probability function $P$ —is a weighted average of $C$ ’s inaccuracy in all worlds, weighted by how likely it is, according to $P$ , that those worlds obtain. Formally:

(1) $${{\mathbb E}_P}\left[ {{\bf{I}}\left( C \right)} \right] = \mathop \sum \limits_{w \in {\rm{\Omega }}} P\left( w \right) \cdot {\bf{I}}\left( {C,w} \right).$$

I will make three assumptions about inaccuracy measures. Though these assumptions are not incontrovertible, they are standard in the accuracy-first literature, and I will not say much to justify them.Footnote 5

The first assumption is:

Strict Propriety

For any two distinct probability functions $P$ and $C$ , ${\mathbb{E}_C}\left[ {{\bf{I}}\left( C \right)} \right] \lt {\mathbb{E}_C}\left[ {{\bf{I}}\left( P \right)} \right]$ .

Strict Propriety says that probabilistic credence functions expect themselves to minimize inaccuracy. Strict Propriety is often motivated by appeal to the norm of immodesty—roughly, that rational agents should be doing best, by their own lights, in their pursuit of accuracy.

The second assumption is Additivity, which says, roughly, that the total inaccuracy score of a credence function at a world is the sum of the inaccuracy scores of each of its individual credences. More precisely:

Additivity

For any $H \in {\cal P}\left( {\rm{\Omega }} \right)$ , there is a local inaccuracy measure ${i^H}$ that takes a world $w \in {\rm{\Omega }}$ and a credence $C\left( H \right)$ in the proposition H to a real number such that:

$${\bf{I}}\left( {C,w} \right) = \mathop \sum \limits_{H \in {\cal P}\left( {\rm{\Omega }} \right)} i_w^H\left( {C\left( H \right)} \right)$$

The third assumption is a continuity assumption for local inaccuracy measures. Specifically:

Continuity

$i_w^H\left( x \right)$ is a continuous function of $x$ .

Now that we know how to measure the inaccuracy of a credence function, we turn to updating rules. I will assume that a learning experience can be characterized by a unique proposition—the subject’s evidence. We define a learning situation as a complete specification of all learning experiences that an agent thinks she might undergo during a specific period of time—a specification of all of the propositions that the agent thinks she might learn during that time. Formally, a learning situation is an evidence function ${\textsf{E}}$ that maps each world $w$ to a proposition ${\textsf{E}}\left( w \right)$ , the subject’s evidence in w. I will write $\left[ {{\textsf{E}} = {\textsf{E}}\left( w \right)} \right]$ for the proposition that the subject’s evidence is ${\textsf{E}}\left( w \right)$ :

(2) $$\left[ {{\textsf{E}} = {\textsf{E}}\left( w \right)} \right] = \left\{ {w{\rm{'}} \in {\rm{\Omega }}:{\textsf{E}}\left( {w{\rm{'}}} \right) = {\textsf{E}}\left( w \right)} \right\}.$$

We define an evidential updating rule as a function $g$ that takes a prior probability function $C$ and an evidence proposition ${\textsf{E}}\left( w \right)$ , and returns a credence function.Footnote 6 In the next two sections of the paper, we will be talking about two updating rules. The first is Bayesian Conditionalization.

Bayesian Conditionalization

$${g_{{\rm{cond}}}}\left( {C,{\textsf{E}}\left( w \right)} \right) = C( \cdot |{\textsf{E}}\left( w \right)).$$

Bayesian Conditionalization says that you should respond to your evidence ${\textsf{E}}\left( w \right)$ by conditioning on your evidence; for any proposition $H$ , your new credence in $H$ , upon receiving your new evidence, should be equal to your old credence in $H$ conditional on your new evidence. The second rule is Metaconditionalization.

Metaconditionalization

$${g_{{\rm{meta}}}}\left( {C,{\textsf{E}}\left( w \right)} \right) = C( \cdot |{\textsf{E}} = {\textsf{E}}\left( w \right)).$$

Metaconditionalization says that you should respond to your evidence ${\textsf{E}}\left( w \right)$ by conditioning on the proposition that your evidence is $E\left( w \right)$ .

2.2 Adopting rules and following rules

I will distinguish adopting an updating rule from following an updating rule. If you follow a rule, then your posterior credence function is the credence function that the rule recommends. If you adopt an updating rule, then you intend or plan to follow the rule. Of course, in general, we can intend or plan to do things without succeeding in doing those doing things. Intending or planning to follow an updating rule is no exception. We can intend or plan to follow an updating rule—in my terminology, we can adopt an updating rule—without following it.Footnote 7

To see how this might happen, consider Williamson’s well-known case of the unmarked clock.Footnote 8 Off in the distance you catch a brief glimpse of an unmarked clock. You can tell that the hand is pointing to the upper-right quadrant of the clock, but you can’t discern its exact location—your vision is good, but not perfect. What do you learn from this brief glimpse? What evidence do you gain? That—according to Williamson—depends on what the clock really reads. If the clock really reads that it is 4:05, the evidence you gain is that the time is between (say) 4:04 and 4:06. If the clock really reads 4:06, the evidence you gain is that the time is between (say) 4:05 and 4:07. Suppose that you adopt Bayesian Conditionalization as your update rule, and that the clock in fact reads 4:05. Your evidence is that the time is between 4:04 and 4:06, but you mistakenly think that your evidence is that the time is between 4:05 and 4:07. As a result you misapply Bayesian Conditionalization; you condition on the wrong proposition.Footnote 9 Despite having adopted Bayesian Conditionalization as your update rule, you did not follow the rule.

The accuracy-first epistemologist says that the rational updating rule is the rule that minimizes expected inaccuracy. I said that there are different ways to make this precise. According to one common way of making it precise, the thesis is a claim about following updating rules (although the distinction between adopting and following is often not made explicit). At a first pass, we might understand this thesis as saying that we are rationally required to follow an updating rule that minimizes expected inaccuracy. But there is an immediate problem with this first-pass thesis, which others have recognized. Consider the omniscient updating rule, which tells you to assign credence 1 to all and only true propositions. The omniscient updating rule is less inaccurate than any other rule at every world, and so every probabilistic credence function expects it to uniquely minimize inaccuracy. But we do not want to say that we are rationally required to follow the omniscient updating rule. To avoid this implication, theorists refine the thesis by appeal to the notion of an available updating rule. The refined thesis says that we’re rationally required to follow an updating rule that is such that (i) following that rule is an available option, and (ii) following that rule minimizes expected inaccuracy among the available options.Footnote 10 Following the omniscient updating rule is not an available option and so we are not required to follow it.

To evaluate this proposal, we need to investigate the notion of availability at issue. A natural thought is that an act is available to you only if you are able to perform the act, and that you are able to perform an act if and only if, if you tried to perform the act, you would.Footnote 11 But on this understanding, even following Bayesian Conditionalization is not always an available option, according to the externalist. Return to the example of the unmarked clock. The clock in fact reads 4:05. Your evidence is therefore that the time is between 4:04 and 4:06. How do you update your credences? There are two cases. In the first case, you correctly identify your evidence, and as a result, you condition on your evidence. In this case, it is true that if you tried to follow Bayesian Conditionalization, you would. In the second case, you mistakenly take your evidence to be that the time is between 4:05 and 4:07, and as a result, you condition on the wrong proposition. In this case, it is not true that if you tried to follow Bayesian Conditionalization then you would, and so it is not true that you are able to follow Bayesian Conditionalization.

Of course, one might object to this account of ability. Rather than wade any further into this debate, I will simply observe that however we define availability, if we state the accuracy-first thesis in terms of following, we’ll be taking for granted that if you adopt an available updating rule, you will follow it; we’ll be ignoring possibilities in which you do not succeed in following your updating rule because you mistake your evidence. But the example of the unmarked clock suggests that cases like this are commonplace. We should take them into account. In light of this, I suggest that we understand the accuracy-first thesis as a thesis about which updating rule we are rationally required to adopt. To that end, we need to say how to evaluate the inaccuracy of adopting an updating rule.

2.3 Actual inaccuracy

I propose to measure the inaccuracy of adopting an updating rule in terms of what I will call actual inaccuracy.Footnote 12 Roughly, the actual inaccuracy of adopting an updating rule $g$ in a world $w$ is the inaccuracy, in $w$ , of the credence function you would have if you adopted $g$ in $w$ .Footnote 13 To give a more precise definition, I need to introduce credal selection functions.

A credal selection function is a function $f$ that takes an evidential updating rule $g$ and a world $w$ , and returns a credence function—the credence function that the subject would have if she were to adopt the rule g in world $w$ .Footnote 14 Of course, any number of factors might play a role in determining what credence function a given subject would have if she were to adopt a certain updating rule. To keep things manageable, I am going to make some simplfiying assumptions about how we are disposed to change our credal states if we adopt Bayesian Conditionalization or Metaconditionalization.

Return to the example of the unmarked clock. Suppose you adopt Bayesian Conditionalization. In fact, the clock reads 4:05 and so your evidence is that the time is between 4:04 and 4:06. How do you update your credences? There are, as before, two cases. In one case, you correctly identify your evidence: to use the terminology that I will from now on adopt, you guess correctly that your evidence is that the time is between 4:04 and 4:06. In this case, the conditional

(I) If you adopted Bayesian Conditionalization, you would follow Bayesian Conditionalization.

is true of you. In the second case, you guess incorrectly that your evidence is that the time is between 4:05 and 4:07. In this case, the conditional (I) is false—if you adopted Bayesian Conditionalization you would condition on the wrong proposition. I will assume that these are the only two cases. Either you guess correctly and condition on the right proposition, or else you guess incorrectly and condition on the wrong proposition.Footnote 15

To make this more precise, fix a set of worlds ${\rm{\Omega }}$ and an evidence function ${\textsf{E}}$ defined on ${\rm{\Omega }}$ . We will let ${{\textsf{G}}^{\textsf{E}}}$ be a guess function defined on ${\rm{\Omega }}$ . This is a function that takes each world $w$ to a proposition ${{\textsf{G}}^{\textsf{E}}}\left( w \right)$ : the subject’s guess about what her evidence is in $w$ .Footnote 16 Then, where ${f_{C,{{\textsf{G}}^{\textsf{E}}}}}$ is the credal selection function for any subject with guess function ${{\textsf{G}}^{\textsf{E}}}$ and prior $C$ :Footnote 17

(3) $${f_{C,{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{cond}}}},w} \right) = {g_{{\rm{cond}}}}\left( {C,{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right),$$
(4) $${f_{C,{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{meta}}}},w} \right) = {g_{{\rm{meta}}}}\left( {C,{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right).$$

Equation (3) says that the credence function you would have if you adopted Bayesian Conditionalization in a world $w$ , given that you have prior $C$ and guess function ${{\textsf{G}}^{\textsf{E}}}$ , is the result of conditioning your prior $C$ on ${{\textsf{G}}^{\textsf{E}}}\left( w \right)$ , your guess about what your evidence is in $w$ . Likewise, (4) says that the credence function you would have if you adopted Metaconditionalization in a world $w$ , given that you have prior $C$ and guess function ${{\textsf{G}}^{\textsf{E}}}$ , is the result of conditioning your prior $C$ on the proposition that your evidence is ${G^E}\left( w \right)$ , your guess about what your evidence is in $w$ .

We will now use credal selection functions to define the actual accuracy of adopting an evidential updating rule. Let $g$ be any evidential updating rule. Let ${{\textsf{G}}^{\textsf{E}}}$ be any guess function. Let $C$ be any prior. We define ${V_{{\rm{C}},{{\textsf{G}}^{\textsf{E}}}}}\left( {g,w} \right)$ , the actual inaccuracy, in $w$ , of adopting rule $g$ given prior $C$ and guess function ${{\textsf{G}}^{\textsf{E}}}$ , as follows:

Actual Inaccuracy

$${V_{{\rm{C}},{{\textsf{G}}^{\textsf{E}}}}}\left( {g,w} \right) = {\bf{I}}\left[ {{f_{{\rm{C}},{{\textsf{G}}^{\textsf{E}}}}}\left( {g,w} \right),w} \right].$$

The actual inaccuracy, in $w$ , of adopting the updating rule $g$ given that you have guess function ${{\textsf{G}}^{\textsf{E}}}$ and prior $C$ is the inaccuracy, in $w$ , of the credence function you would have if you adopted rule $g$ in $w$ , given that $C$ is your prior and ${{\textsf{G}}^{\textsf{E}}}$ is your guess function.Footnote 18

Assuming (3), the actual inaccuracy of adopting Bayesian Conditionalization in a world $w$ for a subject with prior $C$ and guess function ${{\textsf{G}}^{\textsf{E}}}$ is

(5) $${\bf{I}}\left[{{f_{C,{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{cond}}}},w} \right),w\left] { = {\bf{I}}} \right[{g_{{\rm{cond}}}}\left( {C,{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right),w} \right].$$

Assuming (4), the actual inaccuracy of adopting Metaconditionalization in a world $w$ for a subject with prior $C$ and guess function ${{\textsf{G}}^{\textsf{E}}}$ is

(6) $${\bf{I}}\left[{{f_{C,{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{meta}}}},w} \right),w\left]\ { = {\bf{I}}} \right[{g_{{\rm{meta}}}}\left({C,{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right),w} \right].$$

The expected actual inaccuracy of adopting Bayesian Conditionalization and of adopting Metaconditionalization are defined in (7) and (8), respectively:

(7) $$\mathop \sum \limits_{w \in {\rm{\Omega }}} C\left( w \right) \cdot {\bf{I}}\left[ {{f_{C,{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{cond}}}},w} \right),w} \right] = \mathop \sum \limits_{w \in {\rm{\Omega }}} \,C\left( w \right) \cdot {\bf{I}}\left[ {{g_{{\rm{cond}}}}\left( {C,{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right),w} \right],$$
(8) $$\mathop \sum \limits_{w \in {\rm{\Omega}}} C\left( w \right) \cdot {\bf{I}}\left[{{f_{C,{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{meta}}}},w} \right),w} \right] = \mathop \sum \limits_{w \in {\rm{\Omega }}} C\left( w \right) \cdot {\bf{I}}\left[ {{g_{{\rm{meta}}}}\left( {C,{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right),w} \right].$$

Returning to the accuracy-first thesis that the rational updating rule is the rule that does best in terms of accuracy. I have argued that this claim is best understood as a claim about which updating rule we should adopt. We can now make this claim more precise using the notion of actual inaccuracy. I propose to formulate the accuracy-first thesis, which I call Accuracy-First Updating, as follows:

Accuracy-First Updating

You are rationally required to adopt an evidential updating rule that minimizes expected actual inaccuracy.

Let’s turn now to epistemic externalism.

3 Externalism

To characterize externalism, we need to first characterize internalism. Internalism says, roughly, that for certain special propositions, when those propositions are true we have a special kind of access to their truth. Let’s say that you have access to a proposition if and only if, whenever it is true, your evidence entails that it is true. Then internalism says that, for certain special propositions, whenever those propositions are true, your evidence entails that they are true. There are different brands of internalism, depending on what kinds of propositions are taken to be special. According to some, the special propositions are propositions about our own minds, such as the proposition that I am in pain. These internalists say that whenever I am in pain, my evidence entails that I am in pain—I can always tell that I am in pain by carefully attending to this evidence, my own experiences. In this paper we will be mainly interested in one form of internalism—evidence internalism. On this view, propositions about what our evidence is are special propositions in the sense that whenever they’re true, our evidence entails that they are true.

Evidence Internalism

If your evidence is the proposition ${\textsf{E}}\left( w \right)$ , then your evidence entails that your evidence is ${\textsf{E}}\left( w \right)$ .

Let evidence externalism be the denial of evidence internalism. More precisely:

Evidence Externalism

Sometimes, your evidence is some proposition ${\textsf{E}}\left( w \right)$ , but your evidence does not entail that your evidence is ${\textsf{E}}\left( w \right)$ .

Why accept evidence externalism? One standard argument appeals to our fallibility. The externalist says that all of our information-gathering mechanisms are fallible. Now, it is no surprise that our mechanisms specialized for detecting the state of our external environment—such as whether it is raining, or whether there is a computer on my desk—can lead us astray. What is controversial about externalism is its insistence that what is true of these propositions about my external environment is true of nearly all propositions, including the proposition that I am in pain or that I feel cold. The externalist says that, sometimes, I am feeling cold, but my mechanisms specialized for detecting feelings of coldness misfire, telling me that I am not feeling cold.

The externalist asks us to consider a case in which my information-gathering mechanisms have misfired. As a matter of fact, I’m feeling cold, but my mechanisms specialized for detecting feelings of coldness misfire, telling me that I’m not feeling cold. Since it is false that I’m not feeling cold, it is not part of my evidence that I’m not feeling cold. But I have no reason to believe that anything is amiss—it is not part of my evidence that it is not part of my evidence that I’m not feeling cold.Footnote 19

4 The Bayesian Dilemma and the externalist reply

In the introduction I said that some have argued that externalists face a dilemma, the Bayesian Dilemma: Either deny that we are rationally required to adopt Bayesian Conditionalization as our update rule or else deny that the rational update rule is the rule that maximizes expected accuracy, thereby rejecting the accuracy-first program. In this section, I present a core piece of that argument, Schoenfield’s result that you can expect following Metaconditionalization to be more accurate than following any other updating rule. But as we’ll see, this result cannot do the work that others have thought it can. It doesn’t follow from Schoenfield’s result that you expect adopting Metaconditionalization to be more accurate than adopting Bayesian Conditionalization, and I have argued that that it is adopting, not following, that the accuracy-first updating thesis should concern.

Let’s begin by stating Schoenfield’s result.

Theorem 1 Let $E$ be any learning situation. Consider any updating rule $g$ and any prior $C$ such that $g\left( {C,E\left( w \right)} \right) \ne {g_{meta}}\left( {C,E\left( w \right)} \right)$ for some $w$ such that $C\left( w \right) \gt 0$ . Then

$$\mathop \sum \limits_{w \in {\rm{\Omega }}} C\left( w \right) \cdot {\bf{I}}[{g_{{\rm{meta}}}}\left( {C,{\textsf{E}}\left( w \right)} \right] \lt \mathop \sum \limits_{w \in {\rm{\Omega }}} C\left( w \right) \cdot {\bf{I}}[g\left( {C,{\textsf{E}}\left( w \right)} \right].$$

Here is what Theorem 1 says. Consider any evidential updating rule $g$ that disagrees with Metaconditionalization in learning situation ${\textsf{E}}$ . Consider any subject who leaves open worlds where $g$ and Metaconditionalization disagree. Then, Theorem 1 says, that the subject will expect the recommendation of Metaconditionalization to be strictly less inaccurate than the recommendation of $g$ in that learning situation.

But, as Schoenfield and others observe, if evidence externalism is true, Metaconditionalization is not Bayesian Conditionalization. Remember, Baysian Conditionalization says that you should respond to your evidence ${\textsf{E}}\left( w \right)$ by conditioning on ${\textsf{E}}\left( w \right)$ . Metaconditionalization says that you should respond to ${\textsf{E}}\left( w \right)$ by conditioning on the proposition that your evidence is ${\textsf{E}}\left( w \right)$ , the proposition $\left[ {{\textsf{E}} = {\textsf{E}}\left( w \right)} \right]$ . If evidence externalism is true, then ${\textsf{E}}\left( w \right)$ is not always the same proposition as $\left[ {{\textsf{E}} = {\textsf{E}}\left( w \right)} \right]$ . In particular, sometimes ${\textsf{E}}\left( w \right)$ will not entail the proposition $\left[ {{\textsf{E}} = {\textsf{E}}\left( w \right)} \right]$ , and when this happens, Metaconditionalization and Bayesian Conditionalization will disagree.

Let ${\textsf{E}}$ be any learning situation in which $\left[ {{\textsf{E}} = {\textsf{E}}\left( w \right)} \right] \ne {\textsf{E}}\left( w \right)$ for some world $w$ . Consider any subject who leaves open some such worlds. Then Theorem 1 entails that the subject will expect the recommendation of Metaconditionalization to be less inaccurate than the recommendation of Bayesian Conditionalization in learning situation ${\textsf{E}}$ . Formally:

(9) $$\mathop \sum \limits_{w \in {\rm{\Omega }}} C\left( w \right) \cdot {\bf{I}}[{g_{{\rm{meta}}}}\left( {C,{\textsf{E}}\left( w \right)} \right] \lt \mathop \sum \limits_{w \in {\rm{\Omega }}} C\left( w \right) \cdot {\bf{I}}[{g_{{\rm{cond}}}}\left( {C,{\textsf{E}}\left( w \right)} \right].$$

But it doesn’t follow from Theorem 1 that the subject expects adopting—intending or planning to follow—Metaconditionalization to be less inaccurate than adopting Bayesian Conditionalization. That would follow from Theorem 1 only if we knew that the subject would follow Metaconditionalization if she adopted Metaconditionalization, and that she would follow Bayesian Conditionalization if she adopted Bayesian Conditionalization.

To see this, let ${{\textsf{G}}^{\textsf{E}}}$ be the subject’s guess function in learning situation ${\textsf{E}}$ . Let Guess Right be the proposition that the subject’s guess about her evidence in learning situation ${\textsf{E}}$ is right. Formally:

(10) $$Guess{\rm{\;\;}}Right = \left\{ {w \in {\rm{\Omega }}:{{\textsf{G}}^{\textsf{E}}}\left( w \right) = {\textsf{E}}\left( w \right)} \right\}.$$

Let Guess Wrong be the proposition that the subject’s guess about her evidence in ${\textsf{E}}$ is not right. Formally:

(11) $$Guess{\rm{\;\;}}Wrong = \left\{ {w \in {\rm{\Omega }}:{{\textsf{G}}^{\textsf{E}}}\left( w \right) \ne {\textsf{E}}\left( w \right)} \right\}.$$

Say that a subject with guess function ${{\textsf{G}}^{\textsf{E}}}$ is infallible in learning situation ${\textsf{E}}$ if, for any $w \in {\rm{\Omega }}$ , Guess Right is true in $w$ . If we assume that our subject is infallible in learning situation ${\textsf{E}}$ , then for all $w \in {\rm{\Omega }}$ ,

(12) $${f_{C,{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{cond}}}},w} \right) = {g_{{\rm{cond}}}}\left( {C,{\textsf{E}}\left( w \right)} \right),$$
(13) $${f_{C,{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{meta}}}},w} \right) = {g_{{\rm{meta}}}}\left( {C,{\textsf{E}}\left( w \right)} \right).$$

If (12) and (13) are true, then Theorem 1 entails that the subject expects adopting Metaconditionalization to be less inaccurate than adopting Bayesian Conditionalization. Formally:

(14) $$\mathop \sum \limits_{w \in {\rm{\Omega }}} C\left( w \right) \cdot {\bf{I}}\left[ {{f_{C,{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{meta}}}},w} \right),w} \right] \lt \mathop \sum \limits_{w \in {\rm{\Omega }}} C\left( w \right) \cdot {\bf{I}}\left[ {{f_{C,{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{cond}}}},w} \right),w} \right].$$

But of course the externalist will insist that creatures like us are not infallible. Remember, the externalist says my beliefs about what evidence I have are not perfectly sensitive to the facts about what evidence I have. Return to the case of the unmarked clock. In fact my evidence is that the time is between 4:04 and 4:06. But my mechanisms specialized for detecting what evidence I have misfire, and so I mistakenly think that my evidence is some other proposition—that the time is between 4:05 and 4:07. Importantly, the externalist maintains that no amount of careful attention to my evidence will insure me against error. For the externalist, even ideally rational, maximally attentive agents are not always certain of the true answer to the question of what their evidence is. That is just to say that even ideally rational, maximally attentive agents are not always such that, if they adopted Metaconditionalization, they would follow Metaconditionalization.

In short, (13) is often false for agents like us—agents with fallible information-gathering mechanisms. But without (13), we can’t derive (14) from (9). We can’t conclude that, for fallible agents like us, adopting Metaconditionalization has lower expected actual inaccuracy than adopting Bayesian Conditionalization.

Let me summarize. If evidence externalism is true, then Theorem 1 tells us that, under certain conditions, we will expect following Metaconditionalization to be less inaccurate than following any other evidential updating rule. It doesn’t follow, however, that we expect adopting Metaconditionalization to be less inaccurate than adopting any other rule.Footnote 20 In particular, it doesn’t follow that we expect adopting Metaconditionalization to be less inaccurate than adopting Bayesian Conditionalization. That would follow only if we knew that we’re infallible, but we cannot, on pain of begging the question against the externalist, simply assume that this is so. So we have not shown that if evidence externalism is true, then we must choose between the rule that maximizes expected accuracy and Bayesian Conditionalization.Footnote 21

5 The Bayesian Dilemma reconsidered

In this section, I show that we can establish the Bayesian Dilemma without the assumption of infallibility. I give a new argument—I call it the continuity argument—showing that if you are sufficiently confident that you will correctly identify your evidence, then you will expect a rule that I call Accurate Metaconditionalization to have less expected inaccuracy than adopting Bayesian Conditionalization. In section 5.1 I’ll begin by saying what Accurate Metaconditionalization is, and then I’ll present the continuity argument. In section 5.2 I will consider whether other rules are immune to the continuity argument.

5.1 The continuity argument

Metaconditionalization said that you should respond to your evidence ${\textsf{E}}\left( w \right)$ by conditioning on the proposition that your evidence is ${\textsf{E}}\left( w \right)$ . Accurate Metaconditionalization says that you should respond to your evidence ${\textsf{E}}\left( w \right)$ by conditioning on the proposition that your evidence is ${\textsf{E}}\left( w \right)$ and that you have guessed right. (Remember, $Guess{\rm{\;\;}}Right = \left\{ {w \in {\rm{\Omega }}:{{\textsf{G}}^{\textsf{E}}}\left( w \right) = {\textsf{E}}\left( w \right)} \right\}$ .) More precisely:

Accurate Metaconditionalization

Where $C$ is any prior such that $C({\textsf {E}} = {\textsf{E}}\left( w \right)|Guess{\rm{\;\;}}Right) \gt 0$ for all $w \in {\rm{\Omega }}$ ,

$${g_{{\rm{acc}} - {\rm{meta}}}}\left( {C,{\textsf E}\left( w \right)} \right) = C( \cdot |Guess{\rm{\;\;}}Right \wedge {\textsf{E}} = {\textsf{E}}\left( w \right)).$$

For simplicity, I will assume:

(15) $${f_{C,{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{acc}} - {\rm{meta}}}},w} \right) = {g_{{\rm{acc}} - {\rm{meta}}}}\left( {C,{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right) = C( \cdot |Guess{\rm{\;\;}}Right \wedge {\textsf{E}} = {{\textsf{G}}^{\textsf{E}}}\left( w \right)).$$

Equation (15) says that the credence function you would have if you adopted Accurate Metaconditionalization is the result of conditioning your prior on the proposition that your evidence is ${{\textsf{G}}^{\textsf{E}}}\left( w \right)$ , your guess about what your evidence is in $w$ , and that you have guessed right.

I am going to show that for a wide class of fallible subjects, if the subject is sufficiently confident that she will correctly identify her evidence, then adopting Accurate Metaconditionalization will have lower expected actual inaccuracy than adopting Bayesian Conditionalization for her. Here is roughly how the argument will go. I will begin by showing that we can state the expected actual inaccuracy of adopting an updating rule as a function of your credence $x$ in the proposition Guess Right. In particular, we can state the expected actual inaccuracy of adopting Accurate Metaconditionalization as a function of $x$ , and we can state the expected actual inaccuracy of adopting Bayesian Conditionalization as a function of $x$ . Importantly, both functions are continuous functions of $x$ . We will show that when $x = 1$ , adopting Bayesian Conditionalization has greater expected actual inaccuracy than adopting Accurate Metaconditionalization. Since both functions are continuous, it follows there is some $\delta \gt 0$ such that if $x \gt 1 - \delta $ , then adopting Bayesian Conditionalization has greater expected actual inaccuracy than adopting Accurate Metaconditionalization.

Let’s now turn to the details. To begin, I am going to introduce and define a new kind of function, which I’ll call a probability extension function. We can think of a probability extension function as a specification of the conditional credences of some hypothetical subject, conditional on each member of the partition $\left\{ {Guess{\rm{\;\;}}Right,Guess{\rm{\;\;}}Wrong} \right\}$ that the subject leaves open. We then feed the probability extension function a possible credence $x$ in Guess Right (a real number between 0 and 1) and the function returns a (complete) probability function—the probability function determined by the conditional credence specifications, together with $x$ .

To make this more precise, fix a set of worlds ${\rm{\Omega }}$ . Let ${\textsf{E}}$ be any evidence function, and let ${{\textsf{G}}^{\textsf{E}}}$ be any guess function. Let ${\rm{\Delta }}$ be the set of probability functions over ${\cal P}\left( {\rm{\Omega }} \right)$ . We define ${{\rm{\Delta }}_{{\rm{Right}}}}$ as

(16) $${{\rm{\Delta }}_{{\rm{Right}}}} = \left\{ {{P_{\rm{R}}}:{P_{\rm{R}}} \in {\rm{\Delta\ and}}\ {P_{\rm{R}}}\left( {Guess{\rm{\;\;}}Right} \right) = 1} \right\},$$

and we define ${{\rm{\Delta }}_{{\rm{Wrong}}}}$ in a similar way:

(17) $${{\rm{\Delta }}_{{\rm{Wrong}}}} = \left\{ {{P_{\rm{W}}}:{P_W} \in {\rm{\Delta\ and}}\ {P_{\rm{W}}}\left( {Guess{\rm{\;\;}}Wrng} \right) = 1} \right\}.$$

For each pair $\langle {P_{\rm{R}}},{P_{\rm{W}}}\rangle $ consisting of a ${P_{\rm{R}}} \in {{\rm{\Delta }}_{{\rm{Right}}}}$ and a ${P_{\rm{W}}} \in {{\rm{\Delta }}_{{\rm{Wrong}}}}$ , we define a probability extension function ${\lambda _{\langle {P_{\rm{R}}},{P_{\rm{W}}}\rangle }}$ as a function that takes a real number $x$ between 0 and 1 and returns a probability function ${\lambda _{\langle {P_{\rm{R}}},{P_{\rm{W}}}\rangle }}\left( x \right)$ over ${\cal P}\left( {\rm{\Omega }} \right)$ defined as follows:

(18) $${\lambda _{\langle {P_{\rm{R}}},{P_{\rm{W}}}\rangle }}\left( x \right)\left( \cdot \right) = {P_{\rm{R}}}\left( \cdot \right)x + {P_{\rm{W}}}\left( \cdot \right)\left( {1 - x} \right).$$

Each probability extension function is indexed to a pair $\langle {P_{\rm{R}}},{P_{\rm{W}}}\rangle $ . In what follows I will leave off the subscripts for the sake of readability.

We can use probability extension functions to specify the expected actual inaccuracy of adopting an updating rule, for some subject, as a function of her credence in Guess Right. To see this, fix a learning situation ${\textsf{E}}$ , a guess function ${{\textsf{G}}^{\textsf{E}}}$ , and an evidential updating rule $g$ . Each probability extension function $\lambda $ determines a function that takes a credence $x$ in Guess Right and returns the expectation, relative to $\lambda \left( x \right)$ , of the actual inaccuracy of adopting rule $g$ given guess function ${{\textsf{G}}^{\textsf{E}}}$ . For example, consider

(19) $$\mathop \sum \limits_{w \in {\rm{\Omega }}} \lambda \left( x \right)\left( w \right) \cdot {\bf{I}}\left[ {{f_{\lambda \left( x \right),{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{meta}}},w} ),w} \right)}\right] = \mathop \sum \limits_{w \in {\rm{\Omega }}} \lambda \left( x \right)\left( w \right) \cdot {\bf{I}}[{g_{meta}}\left( {\lambda \left( x \right),{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right),w].$$

This is a function that takes a credence $x$ in Guess Right and returns the expectation, relative to $\lambda \left( x \right)$ , of the actual inaccuracy of adopting Metaconditionalization given guess function ${{\textsf{G}}^{\textsf{E}}}$ . Similarly, we have

(20) $$\mathop \sum \limits_{w \in {\rm{\Omega }}} \lambda \left( x \right)\left( w \right) \cdot {\bf{I}}\left[ {{f_{\lambda \left( x \right),{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{cond}}},w }),w}\right)}\right]= \mathop \sum \limits_{w \in {\rm{\Omega }}} \lambda \left( x \right)\left( w \right) \cdot {\bf{I}}[{g_{{\rm{cond}}}}\left( {\lambda \left( x \right),{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right),w].$$

This is a function that takes a credence $x$ in Guess Right and returns the expectation, relative to $\lambda \left( x \right)$ , of the actual inaccuracy of adopting Bayesian Conditionalization given guess function ${{\textsf{G}}^{\textsf{E}}}$ .

For any probability extension function $\lambda $ , we define ${{\cal C}_\lambda }$ as follows:

(21) $${{\cal C}_\lambda } = \left\{ {C \in {\rm{\Delta }}:C = \lambda \left( {C\left( {{{Guess\ Right}}} \right)} \right)} \right\}.$$

We’re thinking of $\lambda $ as a specification of the conditional credences of some hypothetical subject, conditional on each member of $\left\{ {Guess{\rm{\;}}Right,Guess{\rm{\;\;}}Wrong} \right\}$ that the subject leaves open. We can then think of ${{\cal C}_\lambda }$ as the set of all probability functions that agree with $\lambda $ with respect to those assignments of conditional credences. Importantly, every probability function $C \in {\rm{\Delta }}$ belongs to ${{\cal C}_\lambda }$ for some probability extension function $\lambda $ .Footnote 22

We will show that for any probability extension function $\lambda $ satisfying certain constraints, and any probability function $C$ in ${{\cal C}_\lambda }$ , if $C\left( {Guess{\rm{\;}}Right} \right)$ is sufficiently high, then the expected actual inaccuracy, relative to $C$ , of adopting Accurate Metaconditionalization will be lower than the expected actual inaccuracy of adopting Bayesian Conditionalization. More precisely:

Theorem 2 Let $E$ be any learning situation, ${G^E}$ any guess function, and $\lambda $ any probability extension function such that:

  1. (i) $\lambda \left( 1 \right)\left( {{\textsf{E}}\left( w \right)} \right) \gt 0$ for all $w \in {\rm{\Omega }}$ ;

  2. (ii) $\lambda \left( 1 \right)\left( {{\textsf{E}} = {\textsf{E}}\left( w \right)} \right) \gt 0$ for all $w \in {\rm{\Omega }}$ ;

  3. (iii) ${g_{{\rm{meta}}}}\left( {\lambda \left( 1 \right),{\textsf{E}}\left( w \right)} \right) \ne {g_{{\rm{cond}}}}\left( {\lambda \left( 1 \right),{\textsf{E}}\left( w \right)} \right)$ for some $w \in Guess{\rm{\;\;}}Right$ .

Then there’s a ${\delta _\lambda } \gt 0$ such that, for all $C \in {{\cal C}_\lambda }$ , if $C\left( {Guess{\rm{\;\;}}Right} \right) \gt 1 - {\delta _\lambda }$ , then

$$\mathop \sum \limits_{w \in {\rm{\Omega }}} C\left( w \right) \cdot {\bf{I}}\left[ {{f_{C,{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{acc}} - {\rm{meta}}}},w} \right),w} \right] \lt \mathop \sum \limits_{w \in {\rm{\Omega }}} C\left( w \right) \cdot {\bf{I}}\left[ {{f_{C,{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{cond}}}},w} \right),w} \right]$$

The proof of Theorem 2 relies on a lemma.

Lemma 1 Let $E$ be any learning situation, ${G^E}$ any guess function, and $\lambda $ any probability extension function satisfying conditions (i) and (ii) in our statement of Theorem 2. Then

  1. (i) $\mathop \sum \limits_{w \in {\rm{\Omega }}} \lambda \left( x \right)\left( w \right) \cdot {\bf{I}}\left[ {{f_{\lambda \left( 1 \right),{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{meta}}}},w} \right),w} \right)]$

  2. (ii) $\mathop \sum \limits_{w \in {\rm{\Omega }}} \lambda \left( x \right)\left( w \right) \cdot {\bf{I}}\left[ {{f_{\lambda \left( x \right),{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{cond}},w} \right),w} \right)]$

are both continuous at 1.

I leave the proof of Lemma 1 to the appendix.

Proof of Theorem 2. Consider any learning situation ${\textsf{E}}$ , any guess function ${{\textsf{G}}^{\textsf{E}}}$ , and any probability extension function $\lambda $ satisfying (i), (ii), and (iii). It follows from Theorem 1 that

(22) $$\mathop \sum \limits_{w \in {\rm{\Omega }}} \lambda \left( 1 \right)\left( w \right) \cdot {\bf{I}}\left[ {{g_{{\rm{meta}}}}\left( {\lambda \left( 1 \right),{\textsf{E}}\left( w \right)} \right),w} \right] \lt \mathop \sum \limits_{w \in {\rm{\Omega }}} \lambda \left( 1 \right)\left( w \right) \cdot {\bf{I}}\left[ {{g_{{\rm{cond}}}}\left( {\lambda \left( 1 \right),{\textsf{E}}\left( w \right)} \right),w} \right].$$

This says that any subject whose prior is $\lambda \left( 1 \right)$ expects following Metaconditionalization in learning situation ${\textsf{E}}$ to have lower expected inaccuracy than following Bayesian Conditionalization in learning situation ${\textsf{E}}$ . Note that $\lambda \left( 1 \right)\left( {Guess{\rm{\;\;}}Right} \right) = 1$ . This means that, for all $w \in {\rm{\Omega }}$ such that $\lambda \left( 1 \right)\left( w \right) \gt 0$ ,

(23) $${g_{{\rm{meta}}}}\left( {\lambda \left( 1 \right),{\textsf{E}}\left( w \right)} \right) = {f_{\lambda \left( 1 \right),{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{meta}}}},w} \right),$$
(24) $${g_{{\rm{cond}}}}\left( {\lambda \left( 1 \right),{\textsf{E}}\left( w \right)} \right) = {f_{\lambda \left( 1 \right),{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{cond}}}},w} \right).$$

Given (23) and (24), (22) entails

(25) $$\mathop \sum \limits_{w \in {\rm{\Omega }}} \lambda \left( 1 \right)\left( w \right) \cdot {\bf{I}}\left[ {{f_{\lambda \left( 1 \right),{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{meta}}}},w} \right),w} \right] \lt \mathop \sum \limits_{w \in {\rm{\Omega }}} \lambda \left( 1 \right)\left( w \right) \cdot {\bf{I}}\left[ {{f_{\lambda \left( 1 \right),{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{cond}}}},w} \right),w} \right].$$

This says that any subject whose prior is $\lambda \left( 1 \right)$ and whose guess function is ${{\textsf{G}}^{\textsf{E}}}$ expects adopting Metaconditionalization in learning situation ${\textsf{E}}$ to have lower expected inaccuracy than adopting Bayesian Conditionalization in learning situation ${\textsf{E}}$ .

Equation (25) and Lemma 1 together entail that

(26) $$\mathop \sum \limits_{w \in {\rm{\Omega }}} \lambda \left( x \right)\left( w \right) \cdot {\bf{I}}\left[ {{f_{\lambda \left( 1 \right),{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{meta}}}},w} \right),w} \right] \lt \mathop \sum \limits_{w \in {\rm{\Omega }}} \lambda \left( x \right)\left( w \right) \cdot {\bf{I}}\left[ {{f_{\lambda \left( x \right),{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{cond}}}},w} \right),w} \right].$$

We know that for all $C \in {{\cal C}_\lambda }$ , $C = \lambda \left( {C\left( {Guess{\rm{\;\;}}Right} \right)} \right)$ . Therefore, it follows from (26) that

There’s a ${\delta _\lambda } \gt 0 {\rm s.t.,for\ all}\ C \in {{\cal C}_\lambda },{\rm{if\ C}}\left( {Guess{{\;\;}}Right} \right) \gt 1 - {\delta _\lambda },{\rm{then}}$

(27) $$\mathop \sum \limits_{w \in {\rm{\Omega }}} C\left( w \right) \cdot {\bf{I}}\left[ {{f_{\lambda \left( 1 \right),{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{meta}}}},w} \right),w} \right] \lt \mathop \sum \limits_{w \in {\rm{\Omega }}} C\left( w \right) \cdot {\bf{I}}\left[ {{f_{C,{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{cond}}}},w} \right),w} \right].$$

This says that for any subject whose prior probability functions is in ${{\cal C}_\lambda }$ , if the subject is sufficiently confident in Guess Right, then she will expect adopting Metaconditionalization with respect to $\lambda \left( 1 \right)$ to have strictly lower actual inaccuracy than adopting Bayesian Conditionalization with respect to her own prior. Remember, we’re assuming that

(28) $${f_{C,{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{acc}} - {\rm{meta}}}},w} \right) = {g_{{\rm{acc}} - {\rm{meta}}}}\left( {C,{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right) = C( \cdot |Guess{\rm{\;\;}}Right \wedge {\textsf{E}} = {{\textsf{G}}^{\textsf{E}}}\left( w \right)).$$

We are also assuming that

(29) $${f_{C,{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{meta}}}},w} \right) = {g_{{\rm{meta}}}}\left( {C,{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right) = C( \cdot |{\textsf{E}} = {{\textsf{G}}^{\textsf{E}}}\left( w \right)).$$

It follows that

(30) $${f_{C,{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{acc}} - {\rm{meta}}}},w} \right) = {f_{C( \cdot {\rm{\;}}|{\rm{\;}}Guess{\rm{\;\;}}Right),{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{meta}}}},w} \right).$$

We know that, for all $C \in {{\cal C}_\lambda }$ , if $C\left( {Guess{\rm{\;\;}}Right} \right) \gt 0$ then

(31) $$C( \cdot |Guess{\rm{\;\;}}Right) = \lambda \left( 1 \right).$$

Equations (28) and (29) together entail that for all $C \in {{\cal C}_\lambda }$ , if $C\left( {Guess{\rm{\;\;}}Right} \right) \gt 0$ , then

(32) $${f_{C,{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{acc}} - {\rm{meta}}}},w} \right) = {f_{\lambda \left( 1 \right),{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{meta}}}},w} \right).$$

Given (30), (27) entails

There’s a ${\delta _\lambda } \gt 0\ {\rm s.t.,for\ all}\ C \in {{\cal C}_\lambda },{\rm{if\ C}}\left( {Guess{\rm{\;\;}}Right} \right) \gt 1 - {\delta _\lambda },{\rm{then}}$

(33) $$\mathop \sum \limits_{w \in {\rm{\Omega }}} C\left( w \right) \cdot {\bf{I}}\left[ {{f_{C,{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{acc}} - {\rm{meta}}}},w} \right),w} \right] \lt \mathop \sum \limits_{w \in {\rm{\Omega }}} C\left( w \right) \cdot {\bf{I}}\left[ {{f_{C,{{\textsf{G}}^{\textsf{E}}}}}\left( {{g_{{\rm{cond}}}},w} \right),w} \right].$$

This says that for any subject whose prior probability functions is in ${{\cal C}_\lambda }$ and whose guess function is ${{\textsf{G}}^{\textsf{E}}}$ , if the subject is sufficiently confident in Guess Right, then she will expect adopting Accurate Metaconditionalization in learning situation ${\textsf{E}}$ to have strictly lower actual inaccuracy than adopting Bayesian Conditionalization in learning situation ${\textsf{E}}$ . This completes the proof of Theorem 2.

Let’s take stock. In section 4, I presented Schoenfield’s showing that following Metaconditionalization has greater expected actual accuracy than following Bayesian Conditionalization. But, I argued, we cannot conclude from this fact that adopting Metaconditionalization has greater expected actual accuracy than adopting Bayesian Conditionalization. That would follow only if we said that we’re infallible in every learning situation, and we cannot, on pain of begging the question against the externalist, assume that this is so. In this section I have shown that we can do without the assumption of infallibility. Theorem 2 shows that for a wide class of fallible subjects and learning situations, if the subject is sufficiently confident that she will correctly identify her evidence in that learning situation, then adopting Accurate Metaconditionalization will have greater expected actual accuracy for her than adopting Bayesian Conditionalization.Footnote 23

This is not good news for the project of reconciling accuracy-first externalism with Bayesian epistemology. The externalist who wishes to justify Bayesian Conditionalization on the basis of accuracy should hope to find a natural class of fallible agents for whom Bayesian Conditionalization is the most accurate updating procedure in expectation. We should be pessimistic about the prospects for this project on the basis of the results of this paper. Theorem 2 shows that adopting Accurate Metaconditionalization will have greater expected actual accuracy than adopting Conditionalization for some agents in any such class—so long as it includes agents who are sufficiently confident that they will correctly identify their evidence, and I can see no principled reason to exclude all such agents.

5.2 Guess conditionalization

Let me take a moment to address a concern about the significance of this result, and its relationship to other results in the literature. Those who have read Gallow (Reference Gallow2021) or Isaacs and Russell (Reference Isaacs and Sanford Russell2023) might wonder: Haven’t these authors already shown us how fallible agents should update their credences? Gallow (Reference Gallow2021) argues that we can use a version of a result due to Greaves and Wallace (Reference Greaves and Wallace2006) to show that a rule that I will call Guess Conditionalization is the best rule for fallible agents.Footnote 24

Guess Conditionalization

$${g_{{\textsf{G}} - {\rm{cond}}}}\left( {C,{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right) = C( \cdot |{{\textsf{G}}^{\textsf{E}}} = {{\textsf{G}}^{\textsf{E}}}\left( w \right)).$$

However, I believe that the argument that Gallow is alluding to requires certain assumptions about the nature of our fallibility that the externalist should reject. To see this, remember that our guess function ${{\textsf{G}}^{\textsf{E}}}$ is a function that takes each world $w$ to the subject’s guess, in $w$ , about what her evidence is. If we are interested in subjects who are trying to follow Guess Conditionalization, we need another guess function ${{\textsf{G}}^{{{\textsf{G}}^{\textsf{E}}}}}$ that takes each world $w$ to the subject’s guess, in $w$ , about what her guess is in $w$ . Let us assume that

(34) $${f_{C,{{\textsf{G}}^{{{\textsf{G}}^{\textsf{E}}}}}}}\left( {{g_{{\textsf{G}} - {\rm{cond}}}},w} \right) = {g_{{\textsf{G}} - {\rm{cond}}}}\left( {C,{{\textsf{G}}^{{{\textsf{G}}^{\textsf{E}}}}}\left( w \right)} \right).$$

This says that the credence function you would have if you adopted Guess Conditionalization in learning situation ${\textsf{E}}$ is the result of conditioning your prior on your guess about what your guess is.Footnote 25 With this assumption in place, the Greaves and Wallace–style argument that Guess Conditionalization is the best rule for fallible agents requires us to assume that subjects are guess-infallible:

(35) $${\rm{For\ all\ worlds\;\;}}w,{\rm{\;\;}}{{\textsf{G}}^{{{\textsf{G}}^{\textsf{E}}}}}\left( w \right) = {{\textsf{G}}^{\textsf{E}}}\left( w \right).$$

If we assume that our subject is guess-infallible then, for all $w \in {\rm{\Omega }}$ ,

(36) $${f_{C,{{\textsf{G}}^{{{\textsf{G}}^{\textsf{E}}}}}}}\left( {{g_{{\textsf{G}} - {\rm{cond}}}},w} \right) = {g_{{\textsf{G}} - {\rm{cond}}}}\left( {C,{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right).$$

This says that if the subject adopts Guess Conditionalization, then she would follow Guess Conditionalization.

But the externalist should insist that creatures like us are not guess-infallible. According to the externalist, my beliefs about what I have guessed are not perfectly sensitive to the facts about what I have guessed, and, importantly, no amount of careful attention to my guesses will insure me against error. Even ideally rational, maximally attentive agents are not always certain of the true answer to the question of what their guess is. That is just to say that even ideally rational, maximally attentive agents are not always such that, if they adopted Guess Conditionalization, they would follow Guess Conditionalization. In short, (34) is often false for agents like us—agents with fallible information-gathering mechanisms. But without (34), we can’t use the Greaves and Wallace–style argument that Gallow is alluding to in order to show that adopting Guess Conditionalization has lower expected actual inaccuracy than adopting any other rule.

6 Conclusion

It’s been said that accuracy-first epistemology poses a special threat to externalism. Schoenfield (Reference Schoenfield2017) shows that the rule that maximizes expected accuracy is Metaconditionalization. But if externalism is true, Metaconditionalization is not Bayesian Conditionalization. Thus, externalists seem to face a dilemma, which I have called the Bayesian Dilemma: Either deny that Bayesian Conditionalization is required or else deny that the rational update rule is the rule that maximizes expected accuracy. I am not convinced by this argument. Schoenfield’s result shows that following Metaconditionalization has greater expected accuracy than following Bayesian Conditionalization. It doesn’t follow that adopting Metaconditionalization has greater expected accuracy than adopting Bayesian Conditionalization. That would follow only if we also said that if you adopted Metaconditionalization, you would follow Metaconditionalization. But the externalist has every reason to deny that this is always so. I have argued that the Bayesian Dilemma is nevertheless a genuine dilemma. I presented a new argument that does not make any assumptions that the externalist must reject. This argument shows that, for a wide class of fallible subjects, if the subject is sufficiently confident that she will correctly identify her evidence, then adopting Accurate Metaconditionalization will have greater expected accuracy for her than adopting Bayesian Conditionalization.

Acknowledgments

Thanks to David Boylan, Kevin Dorst, Matt Mandelkern, Alex Meehan, Dmitri Gallow, and Bernhard Salow for helpful conversations. I am especially grateful to two people: Milo Phillips-Brown for extremely helpful feedback on earlier drafts, and to Snow Zhang for many conversations about this material, including conversations that led me to starting work on accuracy and updating in the first place.

Appendix A

In this appendix, we prove Lemma 1.

Proof of Lemma 1. We start by showing that (i) is continuous. Observe that (i) is a sum of terms of the form

(37) $$\lambda \left( x \right)\left( w \right) \cdot {\bf{I}}\left[ {{g_{{\rm{meta}}}}\left( {\lambda \left( 1 \right),{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right),w} \right].$$

Notice that $\lambda \left( x \right)\left( w \right) = {P_{\rm{R}}}\left( w \right) \cdot x + {P_{\rm{W}}}\left( w \right)\left( {1 - x} \right)$ is a polynomial and so is continuous everywhere. Moreover, ${\bf{I}}\left[ {{g_{{\rm{meta}}}}\left( {\lambda \left( 1 \right),{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right),w} \right]$ is a constant. Therefore, (i) is a linear combination of continuous functions and therefore is itself continuous.

Next we will show that (ii) is continuous at 1. To begin, observe that (ii) is a sum of terms of the form

(38) $$\lambda \left( x \right)\left( w \right) \cdot {\bf{I}}\left[ {{g_{{\rm{cond}}}}\left( {\lambda \left( x \right),{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right),w} \right].$$

Thus, to show that (ii) is continuous at 1, it suffices to show that (38) is a continuous function at 1 for all $w \in {\rm{\Omega }}$ . We have seen that $\lambda \left( x \right)\left( w \right)$ is a polynomial and so is continuous everywhere. Thus, to show that (38) is continuous at 1 it suffices to show that

(39) $${\bf{I}}\left[ {{g_{{\rm{cond}}}}\left( {\lambda \left( x \right),{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right),w} \right]$$

is continuous at 1. By our assumption that I satisfies Additivity, we have that ${\bf{I}}\left[ {{g_{{\rm{cond}}}}\left( {\lambda \left( x \right),{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right),w} \right]$ is equal to

(40) $$\mathop \sum \limits_{H \in {\cal P}\ {\rm{\Omega }}} i_w^H\left[ {{g_{{\rm{cond}}}}\left( {\lambda \left( x \right),{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right)} \right].$$

Fix an arbitrary $H \in {\cal P}\left( {\rm{\Omega }} \right)$ . To show that (39) is continuous at 1 it suffices to show that

(41) $$f\left( x \right) = i_w^H\left[ {{g_{{\rm{cond}}}}\left( {\lambda \left( x \right),{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right)} \right]$$

is continuous at 1. Define $h\left( x \right)$ as follows:

(42) $$h\left( x \right) = {g_{{\rm{cond}}}}\left( {\lambda \left( x \right),{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right)\left( H \right) = \lambda \left( x \right)(H|{{\textsf{G}}^{\textsf{E}}}\left( w \right)).$$

Then $f\left( x \right) = i_w^H \circ h\left( x \right)$ . By our assumption of Continuity for the local inaccuracy measure $i_w^H$ , we know that $i_w^H$ is a continuous function of $h\left( x \right)$ . Thus, to show that $f\left( x \right)$ is continuous at 1, it suffices to show that $h$ is continuous at 1. By the definition of $\lambda \left( x \right)(H|{{\textsf{G}}^{\textsf{E}}}\left( w \right))$ , we have

(43) $$h\left( x \right) = \lambda \left( x \right)(H|{{\textsf{G}}^{\textsf{E}}}\left( w \right))= {{\lambda \left( x \right)(H \wedge {{\textsf{G}}^{\textsf{E}}}\left( w \right)} \over {\lambda \left( x \right)\left( {{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right)}} = {{{P_{\rm{R}}}\left( {H \wedge {{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right)x + {P_{\rm{W}}}\left( {H \wedge {{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right)\left( {1 - x} \right)} \over {{P_{\rm{R}}}\left( {{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right)x + {P_{\rm{W}}}\left( {{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right)\left( {1 - x} \right)}}.$$

It follows from our assumption that $\lambda \left( 1 \right)\left( {{\textsf{E}}\left( w \right)} \right) \gt 0$ for all $w \in {\rm{\Omega }}$ that $\lambda \left( 1 \right)\left( {{{\textsf{G}}^{\textsf{E}}}\left( w \right)} \right) \gt 0$ for all $w \in {\rm{\Omega }}$ . Since the numerator and denominator are both continuous at 1 and the denominator is greater than zero when $x = 1$ , it follows that $h\left( x \right)$ is continuous at 1. This completes the proof of Lemma 1.

Footnotes

2 See Greaves and Wallace (Reference Greaves and Wallace2006) and Easwaran (Reference Easwaran2013). Not all arguments for updating norms appeal to the norm that one should maximize expected accuracy. Briggs and Pettigrew (Reference Briggs and Pettigrew2020) give an accuracy-dominance argument for Conditionalization. See also Nielsen (Reference Nielsen2021).

3 The name “Metaconditionalization” is due to Das (Reference Das2019). I believe that this rule was first introduced and defended by Matthias Hild; see Hild (Reference Hild1998a,b). Hild calls the rule “Auto-Epistemic Conditionalization.”

4 For recent work on the relationship between accuracy-first epistemology and externalism, see Bronfman (Reference Bronfman2014), Schoenfield (Reference Schoenfield2017), Das (Reference Das2019), Gallow (Reference Gallow2021), and Zendejas Medina (Reference Zendajas Medina2023). Note that not all of these authors argue for the Bayesian Dilemma as I have presented it. For example, Zendejas Medina argues that the Bayesian Dilemma is not a genuine dilemma; in particular, he claims that the dilemma only arises if we accept a certain plan coherence principle, and he argues that we should reject this principle. Das (Reference Das2019) focuses on the relationship between externalism and accuracy-first arguments for Ur-Prior Conditionalization (instead of the rule of Conditionalization).

5 See, among others, Joyce (Reference Joyce1998) and Pettigrew (Reference Pettigrew2016) for defenses of Additivity and Continuity. See Joyce (Reference Joyce1998) and Campbell-Moore and Levinstein (Reference Campbell-Moore and Levinstein2021) for defenses of Strict Propriety.

6 Not all Bayesians accept the assumption that a learning experience can be characterized by a unique proposition. Jeffrey (Reference Jeffrey1965) believed that, sometimes, we undergo a learning experience, but we do not learn with certainty that a unique proposition is true; instead, the experience tells us that a set of propositions ${A_1},{A_2}, \ldots, {A_n}$ should be assigned probabilities ${\alpha _1},{\alpha _2}, \ldots, {\alpha _n}$ . I believe that my arguments can be recast in Jeffrey’s framework, but I do not have the space to explore this question in this paper.

7 My distinction between adopting a plan and following a plan is similar to Schoenfield’s (Reference Schoenfield2015) distinction between the best plan to follow and the best plan to make. See Gallow (Reference Gallow2021), who appeals to a related distinction between flawless dispositions and (potentially) misfiring dispositions. See also Isaacs and Russell (Reference Isaacs and Sanford Russell2023).

8 Williamson (Reference Williamson2000).

9 This analysis of the case of the unmarked clock is due to Gallow (Reference Gallow2021).

10 This is roughly how Greaves and Wallace (Reference Greaves and Wallace2006), Schoenfield (Reference Schoenfield2017), and Das (Reference Das2019) understand it.

11 For defenses of the view that the scope of our options is limited to the scope of our abilities, see Jeffrey (Reference Jeffrey1965, Reference Jeffrey1992), Lewis (Reference Lewis1981), Hedden (Reference Hedden2012), and Koon (Reference Koon2020). For example, Jeffrey (Reference Jeffrey1965) regards options as propositions and writes, “An act is then a proposition which is within the agent’s power to make true if he pleases.”

12 This term comes from Andrew Bacon’s notion of actual value; see Bacon (Reference Bacon2022).

13 Note that when I talk about “worlds” I am talking about big worlds—maximally specific worlds that settle answers to all questions, including questions about what your evidence is and what credence function you adopt.

14 Credal selection functions can be defined in terms of Stalnakerian selection functions. A Stalnakerian selection function $h$ —used in Stalnaker’s (Reference Stalnaker and Rescher1968) semantics for conditionals—is a function that takes a proposition $A$ and a world $w$ and returns another world $h\left( {A,w} \right)$ —intuitively, the world that would have obtained if $A$ had been true in $w$ . Then where Adopt-g is the proposition that the subject adopts updating rule g, we can define $f\left( {g,w} \right)$ as the credence function you have in $h\left( {{\rm{Adopt}} - {\rm{g}},w} \right)$ .

15 In the main text I am assuming that when you adopt Bayesian Conditionalization, you follow a three-step process: (i) you receive some evidence, (ii) you guess what your evidence is, and (iii) you condition your prior on your guess. But it is far from obvious that adopting an updating rule always involves the intermediate step (ii). As an anonymous referee points out, it may be that you simply respond to your evidence without forming any (explicit or implicit) beliefs about what your evidence is. In the end, I want to agree with this. I do not think that the externalist has to think of adopting Conditionalization as involving my intermediate step (ii). What I do think is that the kinds of motivations that lead us to accept externalism should also lead us to believe that, at least sometimes, you will adopt Conditionalization as your update rule yet fail to follow Conditionalization because you condition your prior on the wrong proposition. If we accept that your evidence can come apart from the proposition that you condition on in this way, we can think of ${{\rm{G}}^{\rm{E}}}$ as representing the proposition that you condition on (in the case of Conditionalization). For simplicity, I will continue to talk about guesses about your evidence in the main text, but it is important to remember that the formalism does not have to be interpreted in this way.

16 Isaacs and Russell (Reference Isaacs and Sanford Russell2023) also use the term “guess function.” Note, however, that they use the term differently from how I am using it here. In particular, their guess functions are used to model guesses about which world you are in. (In their framework, worlds are coarse—they settle some questions, but not all.) There are many interesting connections between my framework and the framework used in Isaacs and Russell, but I do not have the space to address them here.

17 Here I assume that ${{\rm{G}}^{\rm{E}}}\left( w \right) = {\rm{E}}\left( {w{\rm{'}}} \right)$ for some $w{\rm{'}} \in {\rm{\Omega }}$ .

18 Note that the actual inaccuracy of adopting $g$ in $w$ is not always the inaccuracy of your credence function in $w$ . Suppose you do not adopt $g$ in $w$ . Then the actual inaccuracy of adopting $g$ in $w$ is the inaccuracy, in $w$ , of the credence function you would have if you had adopted $g$ in $w$ .

19 Versions of this argument can be found in McDowell (Reference McDowell1982, Reference McDowell2011), Williamson (Reference Williamson2000), Weatherson (Reference Weatherson2011), and Salow (Reference Salow2019).

20 Steel (Reference Steel2018) makes this same point in a different context. He observes that the Greaves and Wallace accuracy argument for Bayesian Conditionalization at best shows that Bayesian Conditionalization is the optimal rule to follow; it does not show that Bayesian Conditionalization is the optimal rule to try to follow.

21 Here I state the Bayesian Dilemma in terms of adopting an updating rule because I prefer to state the accuracy-first thesis as a thesis about rule adoption, not a thesis about rule following. As I mentioned in section 2, many theorists (implicitly) take the accuracy-first thesis to be a thesis about following. For these theorists, the Bayesian Dilemma is a choice between (i) the claim that we’re required to follow Bayesian Conditionalization and (ii) the claim that we’re required to follow a rule that minimizes expected inaccuracy. The argument for this version of the Bayesian Dilemma runs as follows. Following Metaconditionalization is an available option, and following Metaconditionalization minimizes expected inaccuracy among the available options. Therefore, if accuracy-first epistemology is true, we’re required to follow Metaconditionalization. But if externalism is true, Metaconditionalization is not Bayesian Conditionalization. So the externalist must choose between accuracy-first epistemology and Bayesian Conditionalization. I don’t think the externalist should be persuaded by this version of the argument, either. In particular, they should deny that following Metaconditionalization is always an available option. Earlier I said that a standard constraint on option availability is that an act is available only if you are able to perform the act. But, for the reasons I discuss in the main text, the externalist should deny that we are always able to follow Metaconditionalization.

22 If $C\left( {Guess{\rm{\;\;}}Right} \right) \gt 0$ and $C\left( {Guess{\rm{\;\;}}Wrong} \right) \gt 0$ , then let $\lambda = {\lambda _{\langle {P_{\rm{R}}},{P_{\rm{W}}}\rangle }}$ where ${P_{\rm{R}}}\left( \cdot \right) = C( \cdot |Guess{\rm{\;\;}}Right)$ and ${P_{\rm{W}}}\left( \cdot \right) = C( \cdot |Guess{\rm{\;\;}}Wrong)$ . If $C\left( {Guess{\rm{\;\;}}Wrong} \right) = 1$ , then let $\lambda = {\lambda _{\langle {P_{\rm{R}}},{P_{\rm{W}}}\rangle }}$ where ${P_{\rm{R}}}$ is any probability function in ${{\rm{\Delta }}_{{\rm{right}}}}$ , and ${P_{\rm{W}}}\left( \cdot \right) = C\left( \cdot \right)$ . If $C\left( {Guess{\rm{\;\;}}Right} \right) = 1$ , let $\lambda = {\lambda _{\langle {P_{\rm{R}}},{P_{\rm{W}}}\rangle }}$ where ${P_{\rm{W}}}$ is any probability function in ${{\rm{\Delta }}_{{\rm{wrong}}}}$ and ${P_{\rm{R}}}\left( \cdot \right) = C\left( \cdot \right)$ .

23 It is worth emphasizing that you don’t have to be that confident that you will correctly identify your evidence. There are models of the unmarked clock in which anything over 50% will do. It is also worth taking a moment to see how this result interacts with considerations of availability that are often discussed in the context of Schoenfield’s result. We said that many theorists (implictly) take the accuracy-first thesis to be a thesis about which rule to follow. On this understanding, the thesis says, roughly, that we’re rationally required to follow an updating rule that is such that (i) following that rule is an available option, and (ii) following that rule minimizes expected inaccuracy among the available options. In footnote 4 I said that the externalist should deny that Metaconditionalization is (always) an available option. My result does not assume that following Accurate Metaconditionalization (or Metaconditionalization for that matter) is an available option; I assume only that adopting Accurate Metaconditionalization is an available option. I see no principled reasons for denying that this is so. The externalist says that I cannot make it the case that I am always certain of the true answer to the question of what my evidence is. They do not deny that I can try or plan to be certain of the true answer to the question of what my evidence is. It is also worth noting that the results in this section do not depend on any assumptions about the structure of evidence. In particular, I have not assumed that evidence obeys introspection principles and I have not assumed that evidence is factive.

24 Note that Gallow himself is actually interested in a slightly different rule, which he calls Update Conditionalization. The differences between Update Conditionalization and Guess Conditionalization do not matter for my purposes.

25 In the main text I assume (i) that you use your first-order guesses (your guesses about what your evidence is) when you try to follow Conditionalization, Metaconditionalization, or Accurate Conditionalization, and (ii) that you use your second-order guesses (your guesses about what your guesses are) when you try to follow Guess Conditionalization. One might question this assumption: Why couldn’t I use my first-order guesses for Guess Conditionalization, too? I agree that my assumption that we always use our second-order guesses for Guess Conditionalization is not necessarily true, and I have made this assumption primarily to simplify the presentation of the argument in the main text. To avoid the argument that Guess Conditionalization is best, we do not have to assume that you always use your second-order guesses when you are trying to follow Guess Conditionalization. Rather, all we have to assume is that it is not always true that if you tried to follow Guess Conditionalization, then you would. For example, here is one way to say that this conditional is not always true without assuming that the agent always uses her second-order guesses when she tries to follow Guess Conditionalization. We could say that in some worlds where the agent’s first-order guesses and second-order guesses come apart, she will try to follow Guess Conditionalization by using her first-order guesses, and in other worlds where her first-order guesses and her second-order guesses come apart, she will try to follow Guess Conditionalization by using her second-order guesses. Thanks to an anonymous referee.

References

Bacon, Andrew. 2022. “Actual Value in Decision Theory.” Analysis 82 (4):617–29. https://doi.org/10.1093/analys/anac014 CrossRefGoogle Scholar
Briggs, R. A. and Pettigrew, Richard. 2020. “An Accuracy-Dominance Argument for Conditionalization.” Nous 54 (1):162–81. https://www.doi.org/10.1111/nous.12258 CrossRefGoogle Scholar
Bronfman, Aaron. 2014. “Conditionalization and Not Knowing That One Knows.” Erkenntnis 79 (4):871–92. https://doi.org/10.1007/s10670-013-9570-0 CrossRefGoogle Scholar
Campbell-Moore, Catrin and Levinstein, Benjamin A.. 2021. “Strict Propriety Is Weak.” Analysis 81 (1):813. https://www.doi.org/10.1093/analys/anaa001 CrossRefGoogle Scholar
Das, Nilanjan. 2019. “Accuracy and Ur-Prior Conditionalization.” Review of Symbolic Logic 12 (1):6296. https://doi.org/10.1017/S1755020318000035 CrossRefGoogle Scholar
Easwaran, Kenny. 2013. “Expected Accuracy Supports Conditionalization—and Conglomerability and Reflection.” Philosophy of Science 80 (1):119142. https://doi.org/10.1086/668879 CrossRefGoogle Scholar
Gallow, Dmitri. 2021. “Updating for Externalists.” Noûs 55 (3):487516. https://doi.org/10.1111/nous.12307 CrossRefGoogle Scholar
Greaves, Hilary and Wallace, David. 2006.“Justifying Conditionalisation: Conditionalisation Maximizes Expected Epistemic Utility.” Mind 115 (459):607–32. https://doi.org/10.1093/mind/f CrossRefGoogle Scholar
Hedden, Brian. 2012. “Options and the Subjective Ought.” Philosophical Studies 158 (2):343–60. https://doi.org/10.1007/s11098-012-9880-0 CrossRefGoogle Scholar
Hild, Matthias. 1998a. “Auto-Epistemology and Updating.” Philosophical Studies 92 (3):321–61. https://doi.org/10.1023/A:1004229808144 CrossRefGoogle Scholar
Hild, Matthias. 1998b. “The Coherence Argument Against Conditionalization.” Synthese 115 (2):229–58. https://doi.org/10.1023/A:1005082908147 CrossRefGoogle Scholar
Isaacs, Yoaav and Sanford Russell, Jeffrey. 2023. “Updating Without Evidence.” Nous 57 (3):576–99. https://doi.org/10.1111/nous.12426 CrossRefGoogle Scholar
Jeffrey, Richard. 1965. The Logic of Decision. Chicago IL: University of Chicago Press.Google Scholar
Jeffrey, Richard. 1992. “Preference Among Preferences.” In Probability and the Art of Judgment, 154–69. Cambridge: Cambridge University Press.CrossRefGoogle Scholar
Joyce, James. 1998. “A Non-Pragmatic Vindication of Probabilism.” Philosophy of Science 65 (4):575603.CrossRefGoogle Scholar
Koon, Justis. 2020. “Options Must Be External.” Philosophical Studies 177 (5):1175–89. https://doi.org/10.1007/s11098-019-01240-0 CrossRefGoogle Scholar
Lewis, David. 1981. “Causal Decision Theory.” Australian Journal of Philosophy 59 (1):530. https://doi.org/10.1080/00048408112340011 CrossRefGoogle Scholar
McDowell, John. 1982. “Criteria, Defeasibility, and Knowledge.” Proceedings of the British Academy 68:455–79.Google Scholar
McDowell, John. 2011. Perception as a Capacity for Knowledge. Milwaukee, WI: Marquette University Press.Google Scholar
Nielsen, Michael. 2021. Accuracy-Dominance and conditionalization. Philosophical Studies 178 (10):32173236.CrossRefGoogle Scholar
Pettigrew, Richard. 2016. Accuracy and the Laws of Credence. Oxford: Oxford University Press.CrossRefGoogle Scholar
Salow, Bernhard. 2019. “Elusive Externalism.” Mind 128 (510):397427. https://doi.org/10.1093/mind/fzx015 CrossRefGoogle Scholar
Schoenfield, Miriam. 2015. “Bridging Rationality and Accuracy.” Journal of Philosophy 112 (12):633–57. https://doi.org/10.5840/jphil20151121242 CrossRefGoogle Scholar
Schoenfield, Miriam. 2017. “Conditionalization Does Not Maximize Expected Accuracy.” Mind 126 (504):1155–87. https://doi.org/10.1093/mind/fzw027 CrossRefGoogle Scholar
Stalnaker, Robert. 1968. “A Theory of Conditionals.” In Studies in Logical Theory, edited by Rescher, Nicholas, 98112. Oxford: Blackwell.Google Scholar
Steel, Robert. 2018. “Anticipating Failure and Avoiding It.” Philosophers’ Imprint 18 (14):128.Google Scholar
Weatherson, Brian. 2011. “Stalnaker on Sleeping Beauty.” Philosophical Studies 155:445–56. https://doi.org/10.1007/s11098-010-9613-1 CrossRefGoogle Scholar
Williamson, Timothy. 2000. Knowledge and its Limits. Oxford: Oxford University Press.Google Scholar
Zendajas Medina, Pablo. 2023. “Just as Planned: Bayesianism, Externalism, and Plan Coherence.” Philosophical Imprint 23:28. https://doi.org/10.3998/phimp.1300 Google Scholar