Self-recommending decision theories
A PDF of this blogpost is available here.
Once again, I find myself stumbling upon a philosophical thought that seems so natural that I feel reasonably confident it must have been explored before, but I can't find where. So, in this blogpost, I'll set it out in the hope that a kind reader will know where to find a proper version already written up fully.*
I'd like to develop a type of objection that might be raised against a theory of rational decision-making. Here, I'll raise it against Lara Buchak's risk-weighted expected utility theory, in particular, but there will be many other theories to which it applies.
In brief, the objection applies to decision theories that are not self-recommending. That is, it applies to a decision theory if there is a particular instance of that theory that recommends that you use some alternative decision theory to make your decision; if you were to use this decision theory to choose which decision theory to use to make your choices, it would tell you to choose a different one, and not itself. We might naturally say that a decision theory that is not self-recommending in this sense is not a coherent means by which to make decisions, and that seems to be a strong strike against it.
A self-recommending Timothy Dalton |
On what basis might we criticize a theory of rational decision making? One popular way is to show that any individual who adopts the theory is exploitable; that is, there are decisions they might face in response to which the theory will lead them to choose certain options when there are alternative options that are guaranteed to be better; that is, in the jargon of decision theory, there are alternatives that dominate the recommendations of the decision theory in question. This is the sort of objection that money pump arguments raise against their targets. For instance, suppose my decision theory permits cyclical preferences, so that I may prefer
This sort of objection is also often raised against decision theories that permit sensitivity to risk. For instance, take the most extreme risk-averse decision theory available, namely Abraham Wald's Maximin. This doesn't just permit sensitivity to risk---it demands it. It says that, in any decision problem, you should choose the option whose worst-case outcome is best. So, suppose you are faced with the choice between
Then Maximin says that you should choose
Now, after facing that first decision, and choosing
You choose
So Maximin is exploitable.
I've argued in various places that I don't find exploitability arguments compelling.** They show only that the decision rule will lead to a bad outcome when the decision maker is faced with quite a specific series of decision problems. But that tells me little about the performance of the decision rule over the vast array of possible decisions I might face. Perhaps an exploitable rule compensates for its poor performance in those particular cases by performing extremely well in other cases. For all the exploitability objection tells me, that could well be the case.
Recognising this problem, you might instead ask: how does this decision rule perform on average over all decision problems you might face? And indeed it's easy to show that decision theories that disagree with expected utility theory will perform worse on average than expected utility theory itself. But that's partly because we've stacked the deck in favour of expected utility theory. After all, looking at the average performance over all decision problems is just looking at the expected performance from the point of view of a credence function that assigns equal probability to all possible decision problems. And, as we'll show explicitly below, expected utility theory judges itself to be the best decision theory to use; that is, it does best in expectation; that is, it does best on average.
But while this argument begs the question against non-expected utility theories, it does suggest a different way to test a decision theory: ask not whether it does best on average, and thus by the lights of expected utility theory; ask rather whether it does best by its own lights; ask whether it judges itself to be the best decision theory. Of course, this is a coherence test, and like all coherence tests, passing it is not sufficient for rationality. But it does seem that failing is sufficient for irrationality. It is surely irrational to use a method for selecting the best means to your ends that does not think it is the best method for selecting the best means to your ends.
Let's begin by seeing a couple of theories that pass the test. Expected utility theory is the obvious example, and running through that will allow us to set up the formal framework. We begin with the space of possible states. There are two components to these states:
- Let
be the set of possible worlds grained finely enough to determine the utilities of all the options between which you will pick; - Let
be the set of decision problems you might face.
Then a state is a pair
Then expected utility theory says that, faced with a decision problem
Now, let's ask how expected utility theory judges itself. Given a decision theory
Now, suppose
with strict inequality if
So
with strict inequality if
Maximin, which we met above, is another self-recommending decision theory. What is the value or choiceworthiness that Maximin assigns to a decision theory
Now, suppose
Now, Maximin is usually rejected as a reasonable decision rule for other reasons. For one thing, without further supplementary principles, it permits choices that are weakly dominated---that is, in some decision problems, it will declare one option permissible when there is another that is at least as good at all worlds and better at some. And since it pays no attention to the probabilities of the outcomes, it also permits choices that are stochastically dominated---that is, in some decision problems, it will declare one option permissible when there is another with the same possible outcomes, but higher probabilities for the better of those outcomes and lower probabilities for the worse. For another thing, Maximin just seems too extreme. It demands that you to take £1 for sure instead of a 1% chance of 99p and 99% chance of £10trillion.
An alternative theory of rational decision making that attempts to accommodate less extreme attitudes to risk is Lara Buchak's risk-weighted expected utility theory. This theory encodes your attitudes to risk in a function
Now, suppose
So the risk-weighted expected utility of
Here is an example to illustrate. The decision is between
And suppose
while
So, while the expected utility of
Now we're ready to ask the central question of this post: does risk-weighted utility theory recommend itself? And we're ready to give our answer, which is that it doesn't.
It's tempting to think it does, and for the same reason that expected utility theory does. After all, if you're certain that you'll face a particular decision problem, risk-weighted expected utility theory recommends using it to make the decision. How could it not? After all, it recommends picking a particular option, and therefore recommends any theory that will pick that option, since using that theory will have the same utility as picking the option at every world. So, you might expect, it will also recommend itself when you're uncertain which decision you'll face. But risk-weighted expected utility theory doesn't work like that.
Let me begin by noting the simplest case in which it recommends something else. This is the case in which there are two decision problems,
You think each is equally likely, you think each world is equally likely, and you think the worlds and decision problems are independent. So,
So REU will tell you to choose
and
So risk-weighted utility theory does not recommend itself in this situation. Yet it doesn't seem fair to criticize it on this basis. After all, perhaps it redeems itself by its performance in the face of other decision problems. In exploitability arguments, we only consider one series of decisions. Here, we only consider a pair of possible decisions. What happens when we have much much more limited information about the decision problems we'll face?
Let's suppose that there is a finite set of utilities,
Here are some results: Set
and
So, for many natural risk-averse and risk-seeking risk functions, risk-weighted utility theory isn't self-recommending. And this, it seems to me, is a problem for these versions of the theory.
Now, for all my current results say, it's possible that there is a risk function other than
* The work of which I'm aware that comes closest to what interests me here is Catrin Campbell-Moore and Bernhard Salow's exploration of proper scoring rules for risk-sensitive agents in Buchak's theory. But it's not quite the same issue. And the idea of judging some part or whole of our decision-making apparatus by looking at its performance over all decision problems we might face I draw from Mark Schervish's and Ben Levinstein's work. But again, they are interested in using decision theories to judge credences, not using decision theories to judge themselves.
** In Section 13.7 of my Choosing for Changing Selves and Chapter 6 of my Dutch Book Arguments.
*** I make this restriction because it's the one for which I have some calculations; there's no deeper motivation. From fiddling with the calculations, it looks to me as if this restriction is inessential.
Comments
Post a Comment