Ordinary and unordinary decision theory

https://www.lesswrong.com/posts/uS6vdQH8zpHyMAsxR/ordinary-and-unordinary-decision-theory

Contents

I have three guiding principles.

Ordinary decision theory

We’re dealing with a rational actor, presumably you, an action space \mathcal{A}, and a set of random utilities U_{a},a\in\mathcal{A}. I take the action set seriously – it is possible to choose any of the actions in the set. We know the distribution of U_{a} for every a\in\mathcal{A}. Let’s say that an acta^{} is rational if it maximizes the expected utility EU_{a}, i.e.,a^{}\in\textrm{argmax}{a}EU{a}.We’ll assume that \textrm{argmax}{a}EU{a} exists. (Defining a rational act is hard when a maximum does not exist. For instance when the act a gives deterministic utility U_{a}=\sum_{i=1}^{a}i^{-2} when \mathcal{A}=\mathbb{N}, and even harder when the utility function is discontinuous at infinity. See "Satan’s apple".) You may want to pause for a second or two to figure out if you agree with this notion of a rational act. Is it the usual definition? Is it strange in any way? Let’s take a look at a simple example. Suppose you’re given the option to flip either (A) a biased coin with probability 0.7 for heads, (B) a biased coin with probability 0.6 for heads. In either case, you can bet $1 and on either tails or heads, winning $2 if you’re right and losing your $1 if you’re wrong. Your action space is \mathcal{A}=\left{ \begin{array}{ccc}(\text{coin A},\text{heads}), & (\text{coin B},\text{heads}), & \text{don't play}.\(\text{coin A},\text{tails}), & (\text{coin B},\text{tails}),\end{array}\right}The expected utilities are EU_{a}=\left{ \begin{array}{ccc}0.4, & 0.2, & 0.\-0.4,- & 0.2,\end{array}\right}.Unsurprisingly, choosing coin A and heads is the rational act. So far, so good. Ordinary decision theory wants you to choose between different random variables U_{a}, not just between different acts a and a fixed source of randomness U. Usually, you would be restricted to acting within the context of a single random variable, as in the von Neumann–Morgenstern utility theorem, where we maximize Eu(a,X). This distinction usually doesn’t matter mathematically, as it’s mathematically trivial to transform a EU_{a} problem to a Eu(a,X) problem. It does matter for interpretation though! Now take note of the following:

Newcomb’s problem and incompatibility between descriptions

Consider the following decision problem. You have two acts, a=1 and a=2. The utilities U_{a} are deterministic and equal to U_{1}=$1,000,000 and U_{2}=$10,000. Which act do you choose? Using the definition of a rational act above, you choose a=1. This is utterly uncontroversial, isn’t it? It turns out it isn’t, as adding mathematically irrelevant information to the problem makes the majority of philosophers choose a=2! Now, let’s add Newcomb’s description to the problem so that U_{1} corresponds to one-boxing and U_{2} to two-boxing. The mathematical formulation of Newcomb’s problem coincides with the setup above: "You have two acts, a=1 and a=2. The utilities U_{a} are deterministic, and are equal to U_{1}=$1,000,000 and U_{2}=$10,000." This information was sufficient for U_{1} to be rational – adding more information can’t change that. Newcomb’s problem feels like a significant problem because it involves a clash between intuitions; on one hand, you wish to reason causally, which leads to the incorrect application of dominance analysis. On the other hand, you want to do the rational act. I can see two resolutions to these

Counterfactual game-theory

It’s a consequence of Jensen’s inequality that, for any random variable Z \max_{a\in A}E[U_{a}]\leq E(\max_{a\in A}E[U_{a}\mid Z]).In other words, you cannot expect to lose expected utility by gaining information. A couple of decision-theoretic paradoxes ostensibly violate Jensen’s inequality. That is a big deal since it implies that we are either (i) dealing with an inconsistent decision problem, or (ii) the problem cannot be formulated as an ordinary decision problem. After all, Jensen’s inequality is a theorem. Recall counterfactual mugging

Omega, a perfect predictor, flips a coin. If it comes up tails Omega asks you for $100. If it comes up heads, Omega pays you $10,000 if it predicts that you would have paid if it had come up tails. Here’s an intuitive analysis of this problem. If you could force yourself to the strategy of paying when X=\textrm{heads}, then \begin{eqnarray*} EU_{\textrm{pay}} & = & \frac{1}{2}E[U_{\textrm{pay}}\mid X=\text{heads})]+\frac{1}{2}E[U_{\textrm{pay}}\mid X=\text{tails}],\ & = & \frac{1}{2}[10000-100]=4950. \end{eqnarray*}On the other hand, refusing to pay is the optimal action, conditioned on X=\text{tails}, which yields a utility of 0. But using this strategy makes \begin{equation} 4950=\max_{a\in A}EU_{a}>E(\max_{a\in A}E[U_{a}\mid X])=0, \end{equation}which appears to imply that you lose money, in expectation, by gaining information. As Jensen’s inequality is violated, counterfactual mugging isn’t an ordinary decision-theoretic problem. What is it then? The most natural solution (one-shot) solution to this problem uses game theory, by letting you and your counterfactual self be the two players. The Nash equilibrium is "not pay", so it does not optimize expected utility across both players (i.e., it’s not Pareto optimal), but that’s not exactly uncommon in game theory. Just as defection is rational in the Prisoner’s dilemma, I see no problem with not paying in counterfactual mugging; your counterfactual self isn’t you. (The counterfactual player does not make an action in counterfactual mugging, but he does in the counterfactual Prisoner’s dilemma.) Psy-kosh’s non-anthropic problem can also be analyzed with a game-theoretic lens to explain why coordinating has the highest payoff and is thus rational, provided it is an option. There is simply no reason to believe game-theoretic reasoning to be Pareto-optimal, so it should be avoided if possible. I haven’t seen any detailed analyses of Psy-kosh’s problem and similar "counterfactual" game-theoretic problems yet, and I don’t believe they are fully solved. But I do believe counterfactual game-theory forms an interesting subset of unordinary decision problems. Developing a theory for counterfactual game theory could be worthwhile. Please tell me if you want to work on it (and like the approach of this post).

When the size of the action space affects U_{a}

Finally, we have paradoxical decision problems about precommitment. In transparent Omega, both boxes are transparent, the $10,000 box and the $1,000,000 box. Omega perfectly predicts your choice and puts $1,000,000 in the transparent box if he predicts one-boxing. If you could remove the strategy a_{2} from the action set, then Omega would correctly predict a_{2}, and put $1,000,000 in the box, so that \max_{a\in\mathcal{A}}EU_{a}=U_{a_{1}}=$1,000,000. Since \mathcal{A}={a_{2}}, the decision problem is trivial. Now consider the case when the action space isn’t reduced, i.e., \mathcal{A}={a_{1},a_{2}}. Since Omega predicts perfectly, he can’t possibly place $1,000,000 in the box, as it would make it possible for you to two-box. (Mere possibility is enough. If you’re given a choice, you’re given a choice.). Thus we’re left with two separate decision problems: \begin{eqnarray*}\mathcal{A}^{\textrm{small}}={a_{1}}, & & U_{a_{1}}^{\textrm{small}}=$1,000,000,\\mathcal{A}={a_{1},a_{2}}, & & U_{a_{1}}=$0,\quad U_{a_{2}}=$10,000.\end{eqnarray*}You can combine both problems into one:

Viewcomb: The set-up is the same as Newcomb, only now you can look inside the opaque box before making your decision. However, if you wish to make your decision straight away, without first looking in the box, you must pay a small fee. As before, the predictor predicted only whether you will one-box or two-box. Here you’re asked to choose between the small decision problem, yielding a guaranteed $1,000,000, and the big decision problem, yielding a guaranteed $10,000. In conclusion, you should make your decision straight away if the fee is smaller $990,000. A standard precommitment version of Viewcomb won’t work. If you write a contract pledging to pay $11,000 when two-boxing, it’s always rational to one-box. But the (irrational) possibility of one-boxing is still there, and Omega, being a perfect predictor, cannot allow you to make that choice. Thus you’re stuck in the large decision problem.

Comment

https://www.lesswrong.com/posts/uS6vdQH8zpHyMAsxR/ordinary-and-unordinary-decision-theory?commentId=vZmnagqWSnEHnaG9F

I guess I don’t see what’s so wrong with "you can’t do all the actions in your action space," which just seems like another way of saying "this action space isn’t the right formalization of the problem."

Comment

https://www.lesswrong.com/posts/uS6vdQH8zpHyMAsxR/ordinary-and-unordinary-decision-theory?commentId=FntCFo3vBgbBns8qH

I disagree. Sometimes your entire payoffs also change when you change your action space (in the informal description of the problem). That is the point of the last example, where precommitment changes the possible payoffs, not only restricts the action space.

https://www.lesswrong.com/posts/uS6vdQH8zpHyMAsxR/ordinary-and-unordinary-decision-theory?commentId=RhxRDhtSNSfmmFqe9

I remain confused about Newcomb-style problems in general, and counterfactual mugging in particular. They seem to typically be fairly trivially self-contradictory when you look at them from the lens of computability theory.

Omega, a perfect predictor, flips a coin. If it comes up tails Omega asks you for $100. If it comes up heads, Omega pays you $10,000 if it predicts that you would have paid if it had come up tails. Either the agent must be running a decidable algorithm, or there is no such Omega satisfying the requirements[1].In the former case, there is no contradiction[2].In the latter case, the problem is a moot point. Suppose you had an algorithm[3] for Omega. Then I feed your Omega an agent that does this:

  • Run Omega on myself[4].

  • If the prediction from 1 is that I would pay if the coinflip comes up tails:

  • If the coinflip came up tails:

  • Don’t pay.

  • Otherwise (that is, if the prediction from 1 is that I would not pay if the coinflip comes up tails):

  • If the coinflip came up tails:

  • Pay.

Unfortunately, it’s fairly simple to show that Omega cannot avoid a contraction here. (About the best that Omega can do is loop forever /​ fail to halt… which still doesn’t fit the problem.) This is a contradiction, so Omega cannot exist.

  • ^Or Omega is super-Turing, in which case all bets are off.

  • ^Jensen’s inequality as used in this context only applies if the agent can do unbounded computation, for one. (Rough sketch of an example with finite computation: suppose I have an agent that does at most C operations. If I ask it for the best outcome over [1] (that is, one outcome with a value of 1), the answer is 1, fairly straightforwardly. If I instead ask it for the best outcome over C outcomes each with a value of \epsilon and one outcome with a value of 1, the agent has an expected value of no more than \frac{C}{C+1} , as it doesn’t have enough operations to scan all possible outcomes looking for the good one.)

  • ^Strictly speaking, Omega is non-deterministic as stated, but you can recover determinism by passing the coinflip result into Omega as an argument.

  • ^This is non-trivial, but doable. It requires a vaguely quine-like construct.

Comment

https://www.lesswrong.com/posts/uS6vdQH8zpHyMAsxR/ordinary-and-unordinary-decision-theory?commentId=pB8DZ4BTBCWjeKBwK

The usual formulations typically have Omega being "bigger" than your agent, such that Omega can predict your agent, but your agent cannot predict Omega. This is generally taken to be unproblematic because physical agents are always bounded by time + memory constraints, which makes them analogous to finite-state machines rather than full Turing machines. (If unbounded computation were permitted, on the other hand, then you would indeed need a super-Turing version of Omega to pull off the trick.)

Comment

...in which case much of game theory does not apply and the paradoxes, by and large, aren’t actually paradoxical, no?

Comment

Paradoxical decision problems are paradoxical in the colloquial sense (such as Hilbert’s hotel or Bertrand’s paradox), not the literal sense (such as "this sentence is false"). Paradoxicality is in the eye of the beholder. Some people think Newcomb’s problem is paradoxical, some don’t. I agree with you and don’t find it paradoxical.

Comment

In that case why are people spending so much effort on it[1]? Ditto, why is there so much argumentation based around applying game theory to Newcomb’s problem[2] (or variants) when much of game theory does not apply to it?


Paradoxical decision problems are paradoxical in the colloquial sense (such as Hilbert’s hotel or Bertrand’s paradox), not the literal sense (such as "this sentence is false"). I think that there is a lot of effort being wasted due to not clearly distinguishing counterintuative results and self-contradictory results in general, and this is a prime example.

  • ^The LessWrong "Newcomb’s Problem" tag has 41 entries with a combined total of over 1,000 comments, for instance.

  • ^See also any argument[3] that says that X is invalid because it provides ‘the wrong solution’ to Newcomb’s Problem.

  • ^Such as "Evidential decision theories are inadequate rational decision theories. For either they provide wrong solutions to Newcomb’s problem [...]", which is a direct quote from the overview of Newcomb’s problem linked in the description of the tag ‘Newcomb’s Problem’* on this site*.