More on self-recommending decision theories

A PDF of this blogpost can be found here.

Last week, I wrote about how we might judge a decision theory by its own lights. I suggested that we might ask the decision theory whether it would choose to adopt itself as a decision procedure if it were uncertain about which decisions it would face. And I noted that many instances of Lara Buchak's risk-weighted expected utility theory (REU) do not recommend themselves when asked this question. In this post, I want to give a little more detail about that case, and also note a second decision theory that doesn't recommend itself, namely, Γ-Maximin (MM), a decision theory designed to be used when uncertainty is modeled by imprecise probabilities.

A cat judging you...harshly
 

The framework is this. For the sake of the calculations I'll present here, we assume that you'll face a decision problem with the following features:

  • there will be two available options, a and b;
  • each option is defined for two exhaustive and exclusive possibilities, which we'll call worlds, w1 and w2; so, each decision problem is determined by a quadruple (a1,a2,b1,b2), where ai is the utility of option a at wi, and bi is the utility of b at wi;
  • all of the utilities will be drawn from the set {0,1,2,,20}; so, there are 214=194,481 possible decision problems.

This is all you know about the decision problem you'll face. You place a uniform distribution over the possible decision problems you'll face.  You take each to have probability 1194,481.

You also assign probabilities to w1 and w2

  • In the case of REU, you have a credence function (p,1p) over w1 and w2, so that p is your credence in w1 and 1p is your credence in w2
  • In the case of MM, you represent your uncertainty by a set of such credence functions {(x,1x):pxq}.
In both cases, you take the worlds to be independent from the decision problem you'll face.

In the case of REU, you also have a risk function r. In this post, I'll only consider risk functions of the form r(x)=rk. I'll write rk for that function.

With all of that in place, we can ask our question about REU. Fix your credence function (p,1p). Now, given a particular risk function r, does REU-with-r judge itself to be the best decision theory available? Or is there an alternative decision theory---possibly REU-with-a-different-risk-function, but not necessarily---such that the risk-weighted expected utility from the point of view of r of REU-with-r is less that the risk-weighted expected utility from the point of view of r of this alternative? And the answer is that, for many natural choices of r, there is.

Let's see this in action. Let p=12=1p. Let's consider the risk functions rk(x)=xk for 0.5k2. Then the following table gives some results. A particular entry, (row k, column k), gives the risk-weighted expected utility from the point of view of rk of using risk-weighted expected utility from the point of view of rk to make your decisions. In each column k, the entry in green is what the risk function rk thinks of itself; the entries in blue indicate the risk functions rk that rk judges to be better than itself; and the entry in red is the risk function rk that rk judges to be best.

How risk-weighted expected utility theories judge each other

There are a few trends to pick out:

  • Risk-inclined risk functions (rk(x)=xk for 0.5k<1) judge less risk-inclined ones to the better than themselves;
  • The more risk-inclined, the further away the risk functions can be and still count as better, so r0.5 judges r0.9 to be better than itself, but r0.7 doesn't judge r1 to be better;
  • And similarly, mutatis mutandis, for risk-averse risk functions (rk(x)=xk for 1<k2). Each judges less risk-averse risk functions to be better than themselves;
  • And the more risk-averse, the further away a risk function can be and still be judged better.
  • It might look like r0.9 and r1.1 are self-recommending, but that's just because we haven't consider more fine-grained possibilities between them and r1. When we do, we find they follow the pattern above.
  • The risk-neutral risk function r1 is genuinely self-recommending. REU with this risk function is just expected utility theory.

So much for REU. Let's turn now to MM. First, let me describe this decision rule. Suppose you face a decision problem between a=(a1,a2) and b=(b1,b2). Then, first, you decide how much you value each option. You take this to be the minimum expected utility it gets from credence functions in the set of credence functions that represent your uncertainty. That is, you take each (x,1x) from your set of credence functions, so that pxq, you calculate the expected utility of a relative to that credence function, and you value a by the minimum expectation you come across. Then you pick whichever of the two options you value most. That is, you pick the one whose minimum expected utility is greatest.

Let's turn to asking how the theory judges itself. Here, we don't have different versions of the theory specified by different risk-functions. But let me consider different sets of credences that might represent our uncertainty. I'll ask how the theory judges itself, and also how it judges the version of expected utility theory (EU) where you use the precise credence function that sits at the midpoint of the credence functions in the set that represents your uncertainty. So, for instance, if p=0.3 and q=0.4 and the representor is {(x,1x):0.3x0.4}, I'll be comparing how MM thinks of itself and how it thinks of the version of expected utility theory that uses the precise credence function (0.35,0.65). In the first column here, we have the values of p and q; in the second, we have the minimum expected utility for MM relative to each of these pairs of values; in the final column, we have the minimum expected utility for EU relative to those pairs.

MMEU(0.3,0.4)12.48612.490(0.3,0.6)12.34712.377(0.3,0.9)12.69012.817(0.1,0.2)12.87012.873

Again, some notable features:

  • in each case, MM judges EU to be better than itself (I suspect this is connected to the fact that ther is no strictly proper scores for imprecise credences, but I'm not sure quite how yet! For treatments of that, see Seidenfeld, Schervish, & Kadane, Schoenfield, Mayo-Wilson & Wheeler, and Konek.)
  • greater uncertainty (which is represented by a broader range of credence functions) leads to a bigger difference between MM and EU;
  • having a midpoint that lies further from the centre also seems to lead to a bigger difference.

At some point, I'll try to write up some thoughts about the consequences of these facts. Could a decision theory that does not recommend itself be rationally adopted? But frankly it's far too hot to think about that today.

Comments

  1. Hvala vam što ste podijelili svoja razmišljanja. Zaista cijenim vaš trud i čekat ću vaš daljnji post, hvala vam još jednom.

    ReplyDelete

Post a Comment