Decisions and Higher-Order Knowledge

Similar documents
Gandalf s Solution to the Newcomb Problem. Ralph Wedgwood

NOTES ON WILLIAMSON: CHAPTER 11 ASSERTION Constitutive Rules

NICHOLAS J.J. SMITH. Let s begin with the storage hypothesis, which is introduced as follows: 1

Williamson, Knowledge and its Limits Seminar Fall 2006 Sherri Roush Chapter 8 Skepticism

1 Introduction. Cambridge University Press Epistemic Game Theory: Reasoning and Choice Andrés Perea Excerpt More information

Lost in Transmission: Testimonial Justification and Practical Reason

Philosophical Perspectives, 16, Language and Mind, 2002 THE AIM OF BELIEF 1. Ralph Wedgwood Merton College, Oxford

Comments on Lasersohn

PHL340 Handout 8: Evaluating Dogmatism

Keywords precise, imprecise, sharp, mushy, credence, subjective, probability, reflection, Bayesian, epistemology

Bayesian Probability

Luminosity, Reliability, and the Sorites

Evidential Support and Instrumental Rationality

COMPARING CONTEXTUALISM AND INVARIANTISM ON THE CORRECTNESS OF CONTEXTUALIST INTUITIONS. Jessica BROWN University of Bristol

Small Stakes Give You the Blues: The Skeptical Costs of Pragmatic Encroachment

THE ROLE OF COHERENCE OF EVIDENCE IN THE NON- DYNAMIC MODEL OF CONFIRMATION TOMOJI SHOGENJI

BLACKWELL PUBLISHING THE SCOTS PHILOSOPHICAL CLUB UNIVERSITY OF ST ANDREWS

Are There Reasons to Be Rational?

Probability and Prodigality

ON PROMOTING THE DEAD CERTAIN: A REPLY TO BEHRENDS, DIPAOLO AND SHARADIN

Skepticism and Internalism

RATIONALITY AND SELF-CONFIDENCE Frank Arntzenius, Rutgers University

CRUCIAL TOPICS IN THE DEBATE ABOUT THE EXISTENCE OF EXTERNAL REASONS

Does Deduction really rest on a more secure epistemological footing than Induction?

Impermissive Bayesianism

what makes reasons sufficient?

Philosophical reflection about what we call knowledge has a natural starting point in the

Boghossian & Harman on the analytic theory of the a priori

Moral Argumentation from a Rhetorical Point of View

2.1 Review. 2.2 Inference and justifications

Is it rational to have faith? Looking for new evidence, Good s Theorem, and Risk Aversion. Lara Buchak UC Berkeley

Varieties of Apriority

Stout s teleological theory of action

The Paradox of the Question

Uncommon Priors Require Origin Disputes

Can A Priori Justified Belief Be Extended Through Deduction? It is often assumed that if one deduces some proposition p from some premises

A Priori Bootstrapping

Introduction: Belief vs Degrees of Belief

Knowledge, Trade-Offs, and Tracking Truth

REASONING ABOUT REASONING* TYLER BURGE

THE CONCEPT OF OWNERSHIP by Lars Bergström

WHY THERE REALLY ARE NO IRREDUCIBLY NORMATIVE PROPERTIES

Understanding Truth Scott Soames Précis Philosophy and Phenomenological Research Volume LXV, No. 2, 2002

Ramsey s belief > action > truth theory.

CAN TWO ENVELOPES SHAKE THE FOUNDATIONS OF DECISION- THEORY?

Lecture 3. I argued in the previous lecture for a relationist solution to Frege's puzzle, one which

Rawls s veil of ignorance excludes all knowledge of likelihoods regarding the social

DESIRES AND BELIEFS OF ONE S OWN. Geoffrey Sayre-McCord and Michael Smith

MULTI-PEER DISAGREEMENT AND THE PREFACE PARADOX. Kenneth Boyce and Allan Hazlett

Received: 30 August 2007 / Accepted: 16 November 2007 / Published online: 28 December 2007 # Springer Science + Business Media B.V.

Against Coherence: Truth, Probability, and Justification. Erik J. Olsson. Oxford: Oxford University Press, Pp. xiii, 232.

Choosing Rationally and Choosing Correctly *

1 What is conceptual analysis and what is the problem?

Reason and Explanation: A Defense of Explanatory Coherentism. BY TED POSTON (Basingstoke,

Akrasia and Uncertainty

How and How Not to Take on Brueckner s Sceptic. Christoph Kelp Institute of Philosophy, KU Leuven

The epistemology of the precautionary principle: two puzzles resolved

The St. Petersburg paradox & the two envelope paradox

Class #14: October 13 Gödel s Platonism

The paradox we re discussing today is not a single argument, but a family of arguments. Here s an example of this sort of argument:!

Philosophy Epistemology Topic 5 The Justification of Induction 1. Hume s Skeptical Challenge to Induction

The Rightness Error: An Evaluation of Normative Ethics in the Absence of Moral Realism

Detachment, Probability, and Maximum Likelihood

The paradox we re discussing today is not a single argument, but a family of arguments. Here are some examples of this sort of argument:

Rawls, rationality, and responsibility: Why we should not treat our endowments as morally arbitrary

Philosophy 5340 Epistemology Topic 4: Skepticism. Part 1: The Scope of Skepticism and Two Main Types of Skeptical Argument

Epistemic utility theory

Analyticity and reference determiners

Bayesian Probability

Imprecise Bayesianism and Global Belief Inertia

2014 THE BIBLIOGRAPHIA ISSN: Online First: 21 October 2014

KNOWLEDGE ON AFFECTIVE TRUST. Arnon Keren

WORLD UTILITARIANISM AND ACTUALISM VS. POSSIBILISM

Etchemendy, Tarski, and Logical Consequence 1 Jared Bates, University of Missouri Southwest Philosophy Review 15 (1999):

Well-Being, Disability, and the Mere-Difference Thesis. Jennifer Hawkins Duke University

Deontological Perspectivism: A Reply to Lockie Hamid Vahid, Institute for Research in Fundamental Sciences, Tehran

5 A Modal Version of the

The Connection between Prudential Goodness and Moral Permissibility, Journal of Social Philosophy 24 (1993):

Epistemic Consequentialism, Truth Fairies and Worse Fairies

In Epistemic Relativism, Mark Kalderon defends a view that has become

TWO APPROACHES TO INSTRUMENTAL RATIONALITY

Dogmatism and Moorean Reasoning. Markos Valaris University of New South Wales. 1. Introduction

Why Have Consistent and Closed Beliefs, or, for that Matter, Probabilistically Coherent Credences? *

An Inferentialist Conception of the A Priori. Ralph Wedgwood

Does the Skeptic Win? A Defense of Moore. I. Moorean Methodology. In A Proof of the External World, Moore argues as follows:

KNOWING AGAINST THE ODDS

Foreknowledge, evil, and compatibility arguments

Scientific Progress, Verisimilitude, and Evidence

PROSPECTS FOR A JAMESIAN EXPRESSIVISM 1 JEFF KASSER

Ayer and Quine on the a priori

RALPH WEDGWOOD. Pascal Engel and I are in agreement about a number of crucial points:

A Review of Neil Feit s Belief about the Self

the negative reason existential fallacy

A Puzzle about Knowing Conditionals i. (final draft) Daniel Rothschild University College London. and. Levi Spectre The Open University of Israel

Let us begin by first locating our fields in relation to other fields that study ethics. Consider the following taxonomy: Kinds of ethical inquiries

REPUGNANT ACCURACY. Brian Talbot. Accuracy-first epistemology is an approach to formal epistemology which takes

Epistemic Contextualism as a Theory of Primary Speaker Meaning

A Defense of the Significance of the A Priori A Posteriori Distinction. Albert Casullo. University of Nebraska-Lincoln

British Journal for the Philosophy of Science, 62 (2011), doi: /bjps/axr026

World without Design: The Ontological Consequences of Natural- ism , by Michael C. Rea.

Transcription:

Decisions and Higher-Order Knowledge Moritz Schulz, University of Tübingen Penultimate draft. Forthcoming in Nôus. Abstract A knowledge-based decision theory faces what has been called the prodigality problem (Greco 2013): given that many propositions are assigned probability 1, agents will be inclined to risk everything when betting on propositions which are known. In order to undo probability 1 assignments in high risk situations, the paper develops a theory which systematically connects higher level goods with higher-order knowledge. 1 Knowledge and Decision It is not unlikely that the prime reason for seeking knowledge is to make good decisions. When Sally checks the timetable to see when the bus is coming, she will typically do so because she can make a better decision about when to leave the house with the relevant information at hand. Needless to say, this does not imply that it is the only reason. Knowledge surely has some intrinsic value, at least in certain cases, and knowledge can also serve other goals one might have, for example, becoming famous by winning the Nobel prize. If there is some truth to this conjecture, one would expect that knowledge plays a key role in decision making. There is evidence that this is indeed the case. For instance, Williamson (2005a: 231) suggests that a proposition p is an appropriate premise for one s practical reasoning iff one knows p. In a similar vein, Hawthorne and Stanley (2008) argue that one should treat the proposition that p as a reason for acting only if one knows that p (similar claims have been defended by Fantl and McGrath 2002, Hyman 1999 and Unger 1975: ch. 5). If decisions should be based on what one knows, knowledge should play a fundamental role in decision theory. Assuming that the right decision to make is one which maximizes (in one way or other) expected utility, the question would be of what kind the probabilities are which figure in the calculation of 1

1 KNOWLEDGE AND DECISION 2 expected utility. If rational decisions are based on knowledge, a minimal requirement should be that the probabilities are taken conditional on the knowledge the agent has. This has the consequence that what the agent knows always has probability 1 (because P(p K) = 1 if K = p). Williamson (2000) has developed a picture of knowledge and probability which accommodates a decision theory of this kind very naturally. 1 According to Williamson, there is an evidential probability function P measuring, roughly speaking, the objective degree of justification a body of evidence E lends to a proposition p. 2 The evidential probability for a proposition p relative to an agent S and a time t is then given by P(p K), where K is the knowledge S has at t. 3 On this picture, knowledge always has evidential probability 1. The decision theory sketched so far conflicts with standard interpretations of Bayesian decision theory. According to many Bayesians, probability 1 should only be given away very sparingly, to logical truths and perhaps to propositions reporting the contents of our subjective experiences (cp. Jeffrey 1965 and Lewis 1981). One prominent reason for this sparingness is decision-theoretic: if a proposition p has probability 1, betting on p has positive expected utility no matter how daunting the odds are (as long as the utility of winning the bet is itself positive), for the chances of losing the bet will be 0. 4 In the form of a popular slogan: one would have to bet one s life for a penny (see sec. 2 for a more detailed analysis). Given that a knowledge based decision theory without any skeptical inclinations allows assignments of probability 1 across a wide range of cases, including substantial claims about the external world, the question how such a theory deals with high-loss low-gain bets of the aforementioned kind is particularly pressing. In light of this problem, Greco (2013) and Kaplan (2009) have argued that a knowledge based decision theory is ultimately untenable. There are various possible responses available to a defender of a knowl- 1 See also Levi (1980), who takes knowledge to be the standard of (serious) possibility. 2 This exposition is more directed at conditional probabilities of the form P(p E). Unconditional probabilities of the form P(p) are explained by Williamson (2000: ch. 10) as measuring the intrinsic plausibility of p. 3 For ease of exposition, I use symbols like p ambiguously both as schematic names for propositions and as propositional schema letters in sentence position. This ambiguity should be harmless as the context will always make clear of which logical type the relevant symbols are. 4 Another reason has to do with the dynamics of belief: assignments of probability 1 cannot be undone by conditionalization, the standard update rule in Bayesian epistemology.

2 THE PRODIGALITY PROBLEM 3 edge based decision theory. She could go contextualist, relativist or sensitive invariantist and hold that higher stakes increase the standards for knowledge, so that one would no longer count as knowing in the problematic scenarios. For instance, Hawthorne and Stanley (2008) adopt sensitive invariantism as part of their defense strategy (but see Greco 2013 for a critical assessment of contextualist and sensitive invariantist solutions of the problem). In a similar spirit, a proponent of an insensitive invariantism about knowledge could explore the possibility that higher stakes require higher-order knowledge (cp. Williamson 2005a, but also Hawthorne and Stanley (2008) consider this as a possible way out). Greco (2013) has objected to this strategy that as things currently stand it is unclear whether such a strategy can be turned into a systematic decision theory. In this paper, I try to close this gap by drawing a more systematic picture of how high stakes decisions might be connected with higher levels of knowledge. If successful, this would make room for a knowledge based decision theory which is compatible with an anti-skeptical insensitive invariantism about knowledge. Here is a brief overview. Section 2 takes a closer look at the prodigality problem. Section 3 introduces the higher-order approach. Section 4 is concerned with how different stakes are reflected in the preferences of an agent. Section 5 presents a problem for the approach so far undertaken. Section 6 develops the idea that symmetry considerations might be the key for solving the prodigality problem. Section 7 turns this suggestion into a precise formulation of a possible decision theory. Section 8 closes with a discussion of a number of anticipated objections. 2 The Prodigality Problem In order to set up the problem for a knowledge based decision theory, a few basic notions have to be introduced. Let us represent the agent s basic preferences by a utility function u : W R from possible worlds into the reals (cp. Lewis 1981). Given two worlds w and w, an inequality like u(w) > u(w ) reveals that w is preferred over w. More generally, the value u(w) is supposed to represent how much the agent would like to see w actualized. To keep things simple, I assume that W is finite. In this case, the expected value (what Jeffrey (1965) calls the news value ) of a proposition A relative to a probability func-

2 THE PRODIGALITY PROBLEM 4 tion P and a body of knowledge K W over the algebra generated by W can be defined as follows: V(A) := P({w} A K)u(w). w W The expected value of A is a weighted average of the values u(w) of worlds in A. It can be seen as an expression of how much a salient agent would welcome the news that A is true. According to evidential decision theory (=EDT), an agent should choose an action A with maximal news value. More precisely, if A 1,..., A n are the available actions, then the agent should perform an action A i such that V(A i ) = max j V(A j ). The presence of K in this formula is due to the fact that P merely represents the evidential probabilities which have not yet been updated by the knowledge the agent has. Although I shall work with this simple version of EDT, let me say that the problem to be discussed does not hinge on this choice. It applies to all broadly Bayesian decision theories including causal decision theory, for it concerns cases in which the available actions are both evidentially and causally independent of the relevant states of affairs. (With hindsight, it will become evident that it applies to any decision theory which allows the possible outcomes of an action to be restricted by the agent s knowledge.) For this reason, I shall often speak of decision theory without any qualifications and I shall, where admissible, ignore the fact that EDT requires to calculate the conditional probabilities P(X i A) and simply work with P(X i ). A consequence of an anti-skeptic insensitive invariantism about knowledge is that even in high stakes decision situations, many contingent propositions are known. If it is further assumed that knowledge is the standard for what should count as a possibility in a decision situation, then even when the stakes are high, many contingent propositions receive epistemic probability 1 in the sense relevant for decision making. Let us call this the prodigality assumption. 5 Intuitively, the prodigality assumption leads to implausible verdicts about certain decision problems. Suppose an agent S knows that she owns a pair of Nikes. She is now offered a bet according to which she stands to gain $100 if she does own a pair of Nikes but is going to lose her life otherwise. Should she 5 The label is derived from Greco s 2013 discussion of the prodigality problem. See below.

3 THE HIGHER-ORDER APPROACH 5 bet? If, for the purposes of this example, S s utilities are identified with money, betting seems superior to not betting. The expected utility of betting would be (with the obvious abbreviations): (1) V(BETTING) = P(NIKES) 100 + P(NO NIKES) u(life) = 1 100 + 0 u(life) = 100. So, the expected value of betting would be 100 (note that it is independent of how much S values her life). Not betting keeps up the status quo: no money is gained or lost. Hence, given that utility was identified with monetary value, not betting has expected utility 0. With the prodigality assumption in place, decision theory recommends taking the bet. Intuitions seem to disagree. One should not be forced to accept low-gain high-loss bets about any piece of ordinary knowledge. Call this, following Greco (2013), the prodigality problem. The quick calculation of the expected value of taking the bet involved a simplification. In order for the expected value of betting to be 100, the agent needs not only to know that she owns a pair of Nikes, but she also has to know that the bookie is fully reliable. For instance, she needs to know that if she takes the bet and owns a pair of Nikes, she will be given a $100 and runs no risk of losing her life. Following Greco (2013), we may call such conditionals bridge principles. In addition to an ordinary piece of knowledge, the set-up of the problem requires knowledge of the bridge principles. But on a robust antiskepticism about knowledge, there seems to be no reason to assume that such knowledge, even if occasionally harder to obtain, is generally unavailable. 3 The Higher-Order Approach Various authors have worried that a knowledge based decision theory cannot be upheld in the light of the prodigality problem (see e.g. Kaplan 2009 and- Greco 2013). If knowledge had probability 1, one would have to bet one s life for a penny on any known proposition. But refraining from doing so does not seem irrational. Is there anything which could be said in response? Suppose I am offered a bet of $100 against my life on a proposition p that I know, say that I own a pair of Nikes. What kind of reasoning may lead me to refuse the bet? Let us set aside potential doubts about the integrity of the

3 THE HIGHER-ORDER APPROACH 6 bookie and a possible aversion against gambling. Then the first question I would probably ask myself is whether I actually own a pair of Nikes. I might answer Yes at this point having a clear memory of my pair of jogging shoes and may even conclude that I know that I own a pair of Nikes. But given that so much is at stake, I might also ask whether this knowledge is afflicted by any possible defects. I might ask myself whether I know that I know that I own a pair of Nikes. I might continue to ask these questions for some iterations of knowledge more. Should I come to the conclusion that for some iteration of knowledge, I probably fail to know, the bet may no longer feel sufficiently safe. Acknowledging the lack of knowledge at some level of iteration is a way of acknowledging the fallibility of our epistemic methods. Mere firstorder knowledge is not an entitlement to absolute certainty, for what looks like knowledge from the inside can sometimes turn out to be a false belief. In the light of this fact, a bet on p need not appear risk-free even if p constitutes knowledge. If this is roughly on the right track, then one may consider the idea that higher stakes require higher iterations of knowledge for a corresponding decision to be rational. Williamson makes a suggestion along such lines in the context of a discussion of insensitive invariantism versus sensitive invariantism or contextualism about knowledge: The insensitive invariantist could try to build variation in the required number of iterations of knowledge into appropriateness itself [...]: in some cases q would be appropriate iff one knew q, in others iff one knew that one knew q, and so on, depending on the stakes. (Williamson 2005a: 232) According to such a view, knowledge would be an appropriate premise for one s practical reasoning only if the stakes are sufficiently low. If the stakes get high, one should be epistemically more careful by resting one s decisions on more secure knowledge, that is on some iteration of knowledge. To make the idea more precise, imagine an order of stakes s 1, s 2,..., s n, so that n iterations of knowledge are required when the stakes are of degree n. There is a certain picture of knowledge required for such a proposal to have any plausibility. Knowing that one knows something should be substantive knowledge which does not come for free. By knowing p, one should not automatically be in a position to know that one knows that p. Thus, one has to assume frequent failures of the so-called KK-principle as put forward, for

3 THE HIGHER-ORDER APPROACH 7 example, in Williamson 2000. As a matter of fact, one may even assume that for human beings the KK-principle fails for any proposition at some level of iteration. For no proposition, all possible iterations of knowledge are known. A second feature of higher-order knowledge assumed here is that it does not merely add a piece of knowledge about one s own mind to the first-order knowledge. Knowing that one knows p requires more than knowledge of p and something like knowledge that one believes p. I shall assume that a certain kind of safety condition is a necessary condition for knowledge: if one knows p, then one does not falsely believe p in appropriately similar situations. Consequently, if one knows that one knows p, one does not falsely believe that one knows p in appropriately similar situations. It follows that one does not falsely believe p in all situations appropriately similar to the original class of appropriately similar situations. Hence, second-order knowledge is safer than first-order knowledge: if one knows that one knows, one does not falsely believe in a larger class of similar situations. Williamson himself is somewhat skeptical about the higher-order approach to decision theory and favors a slightly different solution to the prodigality problem. Part of the reason is a more general skepticism about the prospects of a systematic decision theory: It is important to realize that no decision theory based on expected utility, calculated according to the standard axioms and definitions of mathematical probability theory, will be everywhere consistent with what pretheoretic common sense predicts a sensible person would do. (Williamson 2005b: 480) In particular, he mentions the case of logical theorems and measure-one sets due to non-regular probability distributions. Both are examples of assignments of probability 1 which are in some sense unavoidable. That probability 1 is assigned to logical theorems is already required by the axioms of probability theory. But even contingent propositions are sometimes to be assigned probability 1. Think, for instance, of an infinite lottery where each ticket has the same chance of winning. Then the proposition that the first ticket does not win has probability 1 despite being contingent (for more discussion, see Williamson 2007). But one should at least sometimes be in a position to rationally turn down a bet of one s life on such a proposition or on a (moderately complex) logical theorem. If so, then just about any Bayesian decision theory

3 THE HIGHER-ORDER APPROACH 8 will face a variant of the prodigality problem. A second point Williamson makes is more immediately relevant to a higherorder approach to the prodigality problem. He notes that any decision theory will run occasionally into certain application problems. With respect to a knowledge based decision theory, failures of the KK-principle will lead to cases in which one does not know whether one has correctly applied the theory. Suppose the theory tells us to make a certain decision if we have knowledge that p. Given that knowledge is not always discriminable from false belief we are not always in a position to know that we falsely believe when we happen to do so we are prone to sometimes misapply the theory by acting as if we knew p although in fact we falsely believe p. In a situation of life and death, the consequences of misapplying the theory might be severe. For this reason, Williamson suggests that the risk of misapplying the decision theory needs to be taken into account in making a good decision. One should not bet one s life for a penny on a piece of knowledge, for this would open one up to bet one s life for a penny on a false belief in a relevantly similar situation. To require higher-order knowledge when a lot is at stake is a good cognitive habit aimed at minimizing the risk of misapplying the decision theory. It need not be part of the decision theory itself: Some will be tempted to build an alternative account of rationality around the good cognitive habits. But such habits are too loose and contingent on the accidents of human psychology to provide a systematic decision theory. Nor do they solve the underlying problem, for they are just as vulnerable as anything else to the anti-luminosity argument. Sometimes good cognitive habits would lead one to do something although one is not in a position to know that good cognitive habits would lead one to do it. (Williamson 2005b: 483) According to Williamson, the relation between decisions and higher-order knowledge cannot be expected to be systematic and even if it were, it would still run into the same kind of application problem. The first feature is the reason why Greco (2013) is dissatisfied with such an implementation of a knowledge based decision theory. In the present paper, I wish to show that a higherorder approach to the prodigality problem has more systematic features than one may initially expect. I will leave it open whether the account I develop should be seen as describing a decision theory proper or as revealing some hidden systematicity in the realm of the good cognitive habits. I should per-

4 PREFERENCES AND HIGHER STAKES 9 haps say, though, that I take an association of higher risks with more epistemic care to be a very natural idea, which is why I am inclined to make it part of the decision theory proper. Let me stress that I grant Williamson s point that a higher-order approach to the prodigality problem will still face occasional application problems, but I hope less so. 4 Preferences and Higher Stakes The prodigality problem seems to have two crucial features: (a) particularly valuable goods are at stake and (b) already excluded possibilities have some kind of comeback. If one stood to lose only a $1000, say, one would probably take a bet for a $100 on a proposition one knows to be true. It is the fact that losing the bet would come with a particularly high loss one s life that generates the prodigality problem. Due to the potentially devastating loss of the bet, possibilities which are excluded by one s knowledge become relevant. Given that one generally has less second-order knowledge than firstorder knowledge (KKp implies Kp but not vice versa), considering higher iterations of knowledge does indeed open up previously excluded possibilities. As alluded to earlier, one may suspect that there is some correspondence between the stakes and the number of iterations of knowledge which is required for the decision. Is there any way to model this systematically? Perhaps the first question is how to represent higher stakes. In standard frameworks, preferences are represented by a real-valued utility function u over the possible outcomes of the agent s actions, with higher values representing better outcomes. By itself, this does not allow to discriminate between various stakes, though it is clear that an extreme value for a possible outcome means that a lot is at stake. In order to identify individual stakes, the image of the utility function could be divided into different levels. So we may consider a partition R = E n... E 2 E 1 of the real numbers into a number of intervals ordered by their values, that is if i < j, then the values in E i are smaller than the values in E j. I assume 0 E 1. For simplicity, I also assume that there are only finitely many levels. Note that the higher the levels get, the worse are the outcomes. Outcomes in level E 1 may be seen as the consequences of our everyday decisions. It is the lowest stake. Outcomes at higher levels may concern things like the loss of our job or our house, bankruptcy, our physical

4 PREFERENCES AND HIGHER STAKES 10 health, or the well being of those close to us. 6 It should be stressed that the levels are as subjective as the original preferences. Different people may set the levels differently and may also have different numbers of levels. The levels in the preference structure can be seen as an expression of how much epistemic care the agent wishes to use to protect certain outcomes from happening. It might have been noticed that levels were introduced only in one direction: from good to bad. In principle, one could also consider levels which represent particularly high goods getting rich or winning the Tour de France. Partly because I am less sure about cases of this kind and partly because I would like to keep things simple, I will set this option aside. But let me briefly comment on how one could accommodate such cases. One may extend the partition of levels in the other direction by considering divisions of the form R = E n... E 2 E 1 E 2... E m. On the theory to be developed in more detail below, this would allow one to bet against something one knows but does not know that one knows, say, on the off-chance of becoming a millionaire. Now, following up on the idea that higher stakes require greater epistemic care, let us associate each level with the corresponding iteration of knowledge. So, one would correlate E 1 with knowledge, E 2 with knowledge that one knows, and generally E n with n iterations of knowledge. The claim would then be that a decision which concerns outcomes of level n should be based on n iterations of knowledge. For ordinary decisions, mere first-order knowledge would be enough, but when a lot is at stake, higher iterations of knowledge might be required. Let us write K n for the body of information for which a salient agent possess n iterations of knowledge. More succinctly, the idea can then be expressed thus: if P is a probability function (the evidential probability function on Williamson s account), then a decision concerning actions with possible outcomes of level n (and no higher-level outcomes) should be made by using the probabilities P( K n ) which come from P by conditionalizing on K n. In other words, the standard of possibility relevant to the decision would 6 There is a point of contact with the debate about incommensurability here (see e.g. Broome 1998, 1999b). Although different levels are not incommensurable in the sense of being incomparable one of them is always better than the other, they share with incommensurability the general feature that outcomes may be arranged into decision-theoretically significant classes.

5 HOW DOES ONE KNOW WHAT IS AT STAKE? 11 be K n : what is a possible consequence of one s actions is what is compatible with K n. For instance, when one s life is at stake, one may no longer rely on all one s first-order knowledge but only on those pieces of information for which one possesses sufficiently high iterations of knowledge. 5 How Does One Know what is at Stake? On the picture developed so far, preference orders are given a richer structure by subdividing them into different levels. By considering the possible outcomes of one s actions, one would decide what is at stake in a given decision situation. One would then move to a corresponding iteration of knowledge and calculate the expected utilities of one s options from this perspective, i.e. by using a probability function of the form P( K n ). Unfortunately, this picture is too simple. It presupposes that there is an independently available answer to the question of what is at stake. The problem can be illustrated by considering a few possible resolutions. Let us start with an objective, non-epistmic notion of a stake in a decision situation. Say that what is at stake is at least of level n if an outcome of level n is an objectively possible consequence of one of the agent s available actions. We do not have to decide about how exactly the notion of objective possibility should be understood in order to see that this does not give the right verdict. In an instance of the prodigality problem, the agent knows the proposition p she is offered a bet on and nothing in the prodigality problem requires that the truth value of p is not already settled at the time the decision is made. Hence, there will be instances of the prodigality problem in which the truth value of p is already settled. Consequently, it will not be an objective possibility that the undesired outcome comes about. Objectively speaking, the bet cannot be lost. The problem persist if we consider the first-order epistemic possibilities of the salient agent. Given that the agent knows p, she knows or is at least in a position to know that taking the bet will not lead to the loss of her life. From the perspective of the agent s first-order knowledge, the bet cannot be lost either. It would seem as if there is actually nothing at stake. Of course, the potential risk will show up for some iteration of knowledge at which the agent either does no longer know p or has lost her knowledge of the required bridge principles. For this reason, we might consider the follow-

6 SYMMETRY AS A GUIDELINE 12 ing proposal: an outcome of level n is at stake if at some iteration of knowledge, such an outcome is a possible consequence of one of the actions available to the agent. This might give a more plausible verdict about instances of the prodigality problem, but only at the cost of having implausible consequences for almost all other cases. To see this, note that almost any piece of knowledge is lost at some level of iteration. At very high levels of iteration, one probably knows close to nothing (as mentioned earlier, it might even be that there is to any proposition and agent some level of iteration at which this proposition is no longer known). For this reason, any action has at a high level of iteration possibly very bad consequences. I might get struck by lightning if I step out of the house. I might break my neck by going down the stairs. At very high levels of iteration, one can no longer exclude such possibilities. The present proposal would recommend to evaluate almost any decision problem from an epistemic perspective at which one knows close to nothing. In general, the problem seems to be this. The intuitions about what is at stake do not concern the objective possibilities. They are in some sense about the epistemic possibilities of the agent. However, the first-order knowledge of the agent does not always, as witnessed by the prodigality problem, provide the right standard of possibility. On the other hand, if we consider higher and higher iterations of knowledge, more and more possible consequences of one s available actions open up, so that at some point, almost any action will possibly lead to disaster. In sum, the right iteration of knowledge for determining what is at stake cannot always be the lowest nor always be a very high one. Is there some non-arbitrary way of picking something in-between? 6 Symmetry as a Guideline We can take from the previous section that the challenge for the higher-order approach is to pair a decision problem with an appropriate iteration of knowledge in a non-circular and non-arbitrary way. The iteration of knowledge from the perspective of which the decision problem is evaluated should not be too high and should vary with the problem at hand. In ordinary decision situations, no further iterations of knowledge should have to be appealed to. Only when the stakes get high, iterations of knowledge should become relevant. The approach I shall explore suggests that symmetry considerations may help

6 SYMMETRY AS A GUIDELINE 13 to find an appropriate iteration of knowledge. Low stake situations may be distinguished from high stake situations by recognizing that high stake situations have an asymmetric risk profile: some options come with risks visible from a corresponding level of knowledge which are unmatched by some of the alternatives. Making this notion of symmetry precise will be the main task of the present section. Some terminology will be helpful. In the current framework, possible outcomes of an action are modeled as possible worlds (or small sets thereof which do not differ in any respect the agent cares about; cp. Lewis 1981). The preferences of an agent are defined over possible worlds, which are further subdivided into levels E 1, E 2,.... Call a possible outcome an n-level outcome or an outcome of level n if it is situated in the preference order at level E n. Given the agent s knowledge and its iterations K = K 1, K 2,..., we can define a graded notion of possibility. Call an outcome o an n-possible consequence of an available action a if the conjunction of a and o is compatible with K n, i.e. if performing the action a and yielding the outcome o is compatible with the body of information for which the agent possesses n iterations of knowledge. For n = 1, this simply means that it is an epistemic possibility for the agent that the action will result in outcome o. Derivatively, let us say that an outcome o is n-possible if it is an n-possible consequence of at least one of the available actions. It is important not to conflate the notions of an n-level outcome and an n-possible outcome. The former notion deals with the level of goodness or badness of the outcome, whereas the latter concerns the question what the agent knows about the possible consequences of her available options. To get a first grip on how symmetry considerations might come into play, let us take a closer look at the initial example giving rise to the prodigality problem. I know that I own a pair of Nikes and I also know the relevant bridge principles, i.e. that if I bet and win, I will be given $100 and if I I bet and lose, I will lose my life. To fix ideas, suppose that protecting my life is very important to me, so I might assign it something like level 5 intending to base decisions where my life is at stake only on information for which I have five iterations of knowledge. In the terminology introduced above, losing my life is a 5-level outcome. As far as possibilities are concerned, losing my life is not 1-possible, for it is incompatible with my knowledge that I am going to

6 SYMMETRY AS A GUIDELINE 14 take the bet and lose. But things might be different when it comes to five iterations of knowledge. Perhaps I do not have five iterations of knowledge for owning a pair of Nikes. More likely even is that I do not have five iterations of knowledge for the relevant bridge principles. Be that as it may, let us assume that losing my life, an outcome of level five, is a 5-possible consequence of taking the bet. What about not taking the bet? Here nothing happens and we can assume that even at five iterations of knowledge it does not become possible that I will lose my life. Thus, from the perspective of five iterations of knowledge, the number of iterations which is associated with the outcome of losing my life, the risk profile of the two available actions is asymmetric: taking the bet comes with a high level risk whereas not taking the bet does not. Compare this with a case in which I have to choose between two paths taking me through dangerous territory. I have been told by reliable sources that both of them are safe, so we may assume that I know that they are safe. But we may also assume that my knowledge is not particularly robust: already after a few iterations, I no longer know that the two routes are safe. As a result, losing my life is an 5-possible outcome. In this case, however, the risk profile of the two available actions is symmetric: a 5-level outcome is a 5-possible consequence of both of the actions. Thus, even when we consider iterations of knowledge, none of the two actions reveals a hidden risk not shared by the other. It is tempting to conclude that this might be the reason why I can make my decision in this case from the perspective of my first-order knowledge, while I should move to a higher iteration of knowledge in the former case illustrating the prodigality problem due to the fact that it reveals an asymmetry in the risks the available actions come with. Let us see how the idea of risk symmetry can be made more precise. This is a somewhat more complicated task than one might initially expect, for one has to simultaneously keep track of the levels of the possible outcomes in the preference order and of the unfolding hierarchy of iterations of knowledge. To cope with the interplay of these two dimensions, it is helpful to start by focussing on a fixed level m of outcomes and a fixed iteration n of knowledge (the previous examples assumed m = n = 5). If we take a clue from the example of the two paths, the risk profile of the two available actions seems symmetric from the fifth iteration of knowledge because both actions have a level-5

6 SYMMETRY AS A GUIDELINE 15 outcome which is an equally probable consequence of these actions. There is a one-to-one mapping between the level-5 outcomes of the two actions which preserves probability. In general, the situation may be a little more complex, though. The available actions may have more than one possible outcome of the relevant level, the number of such outcomes may also be different for various actions, and these outcomes may not always be equally likely consequences of the actions. Still, if actions with more problematic outcomes lead to them with a lower probability, they could be counted as coming essentially with the same risk. If this is correct, the notion we need is one of expected risk. We may define such a notion for an action A, relative to a level of outcome m and a number n of iterations of knowledge, thus: EV(A, m, n) := w E m P(w A K n ) u(w). This mirrors the definition of expected utility with the difference that it considers only outcomes of a certain level and evaluates the probabilities based on possibly higher-order knowledge. With this definition in place, we can then say that a family of available actions A i are symmetrical relative to outcomes of level m and evaluated from the perspective of n iterations of knowledge, for short (m, n)-symmetrical, if EV(A i, m, n) = EV(A j, m, n) for all i, j. For instance, the paths example is (5, 5)-symmetrical, whereas the problematic case illustrating the prodigality problem is one where the available actions are not (5, 5)-symmetrical: as I described the case, losing my life becomes possible from the higher-order perspective and so EV(BET, 5, 5) will be negative while EV(NOT BET, 5, 5) remains zero. The asymmetry in the latter case only shows up for a higher iteration of knowledge: the two available actions are still (5, 1)-symmetrical. From the perspective of first-order knowledge, none of them comes with any high-level risk. In order to make progress in finding the right epistemic perspective from which a given decision should be made, we should also focus on the interplay between the level of an outcome and higher degrees of knowledge. Consider, for instance, a hypothetical situation in which the available actions are not (2, 3)-symmetrical. Should this be a reason to count three iterations of knowledge as being relevant for making the decision? If we go back to the initial proposal of how levels of outcomes might be connected with iterations of knowl-

6 SYMMETRY AS A GUIDELINE 16 edge, the suggested answer is negative, because the levels of outcomes were introduced with the idea that they indicate how much epistemic care should be invested into decisions where corresponding outcomes are at risk. A level 2 outcome is one for which the agent has decided that decisions where it is at stake should be epistemically protected by two layers of knowledge. But this means that an asymmetry showing up at three iterations of knowledge should not count as relevant, for this is more epistemic protection than the agent takes to be necessary. Consider, for instance, a decision situation where the agent does not have three iterations of knowledge guaranteeing that one of the actions will not lead to a loss of her job, while from the perspective of two levels of knowledge losing her job is not a possibility. If job loss is a level 2 outcome for her, then according to the present line of thought, the asymmetry for three iterations of knowledge should not force her to move to this level of knowledge in order to make her decision. Generalizing, one may consider the thought that (m, n)-asymmetry is only a factor if m n. The idea would be that an asymmetry at a certain layer of knowledge should only be taken into account if it concerns outcomes the layer of knowledge is supposed to protect. More precisely, we may call a layer n of knowledge symmetrical if it is (m, n)-symmetrical for all m n. Going back to the examples, it may be the fact that the fifth layer of knowledge is not symmetrical in the prodigality case which makes us steer away from betting, while in the paths example the symmetry at higher iterations is the reason why we should base the decision on our first-order knowledge. As defined, symmetry from the perspective of a certain layer of knowledge is a restricted kind of indifference. In the case of first-order knowledge, it actually coincides with indifference. To see this, recall that the first level is a symmetrical layer of knowledge if it is (m, 1)-symmetrical for all m 1. But the latter condition is satisfied by all m, and so the first level being symmetrical means that it is symmetrical regarding all levels of outcomes. But now note that the expected value of an action is just the sum of the expected values for the various levels of outcomes, which derives from the fact that they form a partition: (2) V(A) = w P(w A K 1 ) u(w) = m w Em P(w A K 1 ) u(w) = m EV(A, m, 1).

6 SYMMETRY AS A GUIDELINE 17 Hence, if two actions A 1 and A 2 are such that EV(A 1, m, 1) = EV(A 2, m, 1) for all m, they will have the same utility, i.e. V(A 1 ) = V(A 2 ). Symmetry for higher iterations of knowledge will no longer imply this, however, for symmetry is defined so that it only looks up but not down. Still, a symmetrical layer of knowledge comes with a limited recommendation of indifference: as far as higher-level goods are concerned, there is no reason to prefer one of the actions over the others. Each of the actions comes with the same amount of risk as far as higher-level goods are concerned (to be precise: goods which belong to a level at least as high as the layer of knowledge one is looking at). This observation might help to narrow down the situations in which a decision should be made from a higher-order perspective of knowledge. This is an important constraint, for a decision made from a higher-order perspective is a decision which avails itself of less information than is available at a lower level of knowledge, for one generally has more first-order knowledge than second-order knowledge, and so on. Moving to a higher layer of knowledge thus comes with a certain price: one is going to make a decision which will, objectively speaking, tend to be suboptimal. Of course, one also gains something in return: more epistemic care for goods which are particularly valuable. But one should only pay the price of moving to a higher layer of knowledge when it is really necessary. Here is where symmetry might come in: a symmetrical layer of knowledge tells us that the goods it is supposed to protect are unaffected by the choice we make, which means that we can go down to a lower iteration of knowledge and make our decision on the basis of a larger, though somewhat less secure information base. As a consequence, this suggests that the right epistemic perspective from which a decision should be approached is never a symmetrical layer of knowledge unless all levels are symmetrical including the first one, in which case we can be indifferent. For one can always improve one s decision by going down at least one level without ignoring any kind of hidden risks. Hence, an intermediate conclusion is this: a decision should be made from an asymmetrical level of knowledge. The question thus simplifies to: which asymmetrical layer of knowledge should guide our decisions? If the levels in one s preference order are finite as assumed, then there will

7 A DECISION THEORY 18 always be a highest asymmetrical layer of knowledge. Take m to be the highest preference level and suppose n > m. Then a layer of n iterations of knowledge is trivially symmetrical, for there are no outcomes of a level at least as high as n, for which an asymmetry in risks among the actions could then be revealed. The highest asymmetrical layer reveals an asymmetry in risks for one s available actions concerning a level of outcome in need of more epistemic care than any of the outcomes at lower levels. It is therefore a natural candidate for being the perspective from which one s decision should be made. Were one to make it at a lower level, one might ignore the risk revealed at the higher layer of knowledge which was chosen to protect the goods of the corresponding level. On the other hand, it does not seem necessary to go higher than the highest asymmetrical level, for as we have seen, symmetrical layers of knowledge always suggest to go down a layer. Thus, this is the proposal I would like to make: the right epistemic perspective from which to evaluate a decision problem is the highest asymmetrical level of knowledge. 7 A Decision Theory The previous section contains a sketch of a possible decision theory which connects iterations of knowledge with higher-level goods. What is still missing is a clear statement of the formal structure of the theory. To this end, it proves helpful to start with a description of a decision problem as it is conceived of within such a theory. A decision problem can be identified with an ordered tuple D = (W, P, (K n ) n N, u, (E m ) m=1,2,...,k, (A i ) i=1,2,...,j ). W is a set of possible worlds the possible outcomes, which is assumed to be finite for the sake of simplicity. P is a probability function defined over W (to be fully precise: it is defined over the algebra generated by the power set of W). The function P represents the evidential probabilities. As far as its role in the theory is concerned, it can be compared to an initial probability function in standard frameworks of Bayesianism. (K n ) n N is a family of nested non-empty sets of possible worlds K 1 K 2..., where K n represents the information for which the salient agent possesses n iterations of knowledge. As defined earlier, u is a utility function from the set W of possible worlds into the reals and (E m ) m N is the ordered (and finite) partition of W into different levels. Finally, the family A 1, A 2,..., A j is a finite partition of W representing the actions available to the

7 A DECISION THEORY 19 agent. The evidence an agent has for a proposition will be taken to be the probability conditional on the agent s knowledge, i.e. as given by P 1 ( ) := P( K 1 ). More generally, let us define P n ( ) := P( K n ). The probabilities P n represent the likelihoods of propositions from the perspective of n iterations of knowledge. Moreover, given a decision problem D, let f (D) be the highest asymmetrical layer of knowledge. If all iterations are symmetrical, set f (D) := 1 (as explained earlier, this means that the agent can be indifferent). With this in place, we can now calculate the expected utility from the perspective of the asymmetrical layer of knowledge, to which we may refer as the level adjusted expected utility: V (A) := P f (D) (w A) u(w). w W The claim would then be that the agent should choose an action A i for which V (A) is maximal. The form of the definition for expected utility is the same as in standard explications of (evidential) decision theory. So rational agents are still taken to be expected utility maximizers. What is added to standard decision theory is a claim about what the standard for possibility is relative to a given decision problem: sometimes it is simply knowledge, at other times a layer of higherorder knowledge. This means that the probabilities figuring in the calculation of expected utility may vary from decision problem to decision problem depending on how much is at stake. This is the crucial (but only) difference to standard decision theories. In ordinary situations in which all higher levels are symmetrical, so that the highest asymmetrical level of knowledge is simply 1, the proposal converges to the standard theory as outlined earlier, i.e. one finds V (A) = V(A). There is a special case for which the theory can be improved a little further. Suppose f (D) > 1, i.e. the decision problem is approached from a higher layer of knowledge. Assume in addition that there is more than one action A for which V (A) is maximal. Typically, decision theory recommends in such cases to be indifferent between these options. But in the present kind of case, there might still be a reason to prefer one of the options over the others, for one can avail oneself of more information by going down a layer of knowledge. With more information at hand, there could be a reason to favor one of the options.

8 OBJECTIONS AND REPLIES 20 So, the idea would be that one calculates V 1 (A) := P f (D) 1 (w A) u(w) w W for those actions with maximal V (A) and choose one with maximal V 1 (A). This might filter out some actions, but it could still leave one with more than one. If one has not already reached the level of first-order knowledge, one may simply repeat the procedure. So one would successively avail oneself of more and more knowledge to see whether one can find a reason to prefer one of the options with maximal expected utility at the highest asymmetrical layer of knowledge. The procedure terminates if either one has reached level 1 or there is only one option left with maximal expected utility. If level 1 is reached with more than one action having maximal utility, we have a real case of indifference. Starting with the set of actions A0 of those actions A for which V (A) is maximal, this defines a finite nested sequence A0 A 1 A 2... A j. The more refined proposal then reads: an agent should pick an action out of A j. This will be an action A for which V (A) is maximal, but it may have the additional advantage of having greater expected utility from the perspective of a more informative layer of knowledge. 8 Objections and Replies There are still a number of questions open and a lot of further issues to be addressed. I conclude with a discussion of some anticipated objections, with a focus on those which strike me to be the most immediate ones. 1. Does it really solve the prodigality problem? The worry here might be that the theory does not predict that one should turn down the bet under all circumstances. Suppose we have sufficiently many iterations of knowledge for the proposition to be bet upon. Then the expected value of betting will be the same as in the original problem for first-order knowledge and the theory recommends taking the bet. And even if we do not have enough iterations of knowledge, there will still be cases in which the probability of the proposition is so extremely high that the expected value of betting is higher than the expected value of not betting. So, the theory seems to solve the prodigality problem only for a limited range of cases.

8 OBJECTIONS AND REPLIES 21 I disagree. To begin with, the prodigality problem was not that a bet of $100 against our life should never be taken. The problem was that one should not always be forced to take the bet if the target proposition is known (Greco (2013) makes the same point). This latter problem is solved by the theory. But perhaps the former still is a problem even if it is not to be identified with the prodigality problem. This is an interesting question, but I am inclined to think that it has a negative answer. If one has enough iterations of knowledge, the bet will be very safe, much safer than if one possessed only first-order knowledge. Taking this observation from the realm of betting to more ordinary decision situations, this means that one can ignore certain potential risks if one has strong enough evidence to suggest that they are not actual. Ordinarily, we do not take into consideration outlandish possibilities ruled out by information for which we have sufficiently many iterations of knowledge, like our partner being the much wanted ax murderer or our house falling apart by tomorrow. But even a stronger conclusion seems to be justified. In practice, we sometimes (not always) seem to tolerate a very small probability of a fairly big loss for a moderate gain. On the present theory, this kind of tolerance is appropriate if the probability is sufficiently small when evaluated from a sufficiently high level of knowledge. This means that the low probability will be particularly resilient because it is based on a body of information for which we possess a high number of iterations of knowledge. To further defend this consequence, it should also be noted that any Bayesian decision theory according to which agents should maximize expected utility has the consequence that an action with a particular bad possible outcome should sometimes be favored when the probability of the outcome is sufficiently small. 7 If one is unhappy about this result, one would probably have to give up utility maximization and add an element to decision theory which ranks actions without particularly bad consequences over actions with such consequences independently of the respective probabilities. In principle, it seems possible to add such an element to the present theory by placing those 7 The reason is that Bayesian decision theories typically validate a version of the continuity axiom. This axiom states: If A is preferred to B and B to C, then there are probabilities p, q (0, 1) such that ApC is preferred to B and B to AqC, where for any probability r, ArC means the prospect of getting A with probability r and C with probability 1 r. So, the axiom requires that our preferences for prospects do not change if one adds possible outcomes with very small probabilities, even if the outcome is arbitrarily bad like the loss of one s life.