Update: If you are interested in understanding my thoughts on Newcomb's problem, I would recommend starting with Why 1-boxing doesn't imply backwards causation and then Deconfusing Logical Counterfactuals. The later doesn't quite represent my most recent views, but I still think it's worth reading.

I no longer endorse the Prediction Problem as being informative about which decision theory is better, but rather only as a useful intuition pump for why you should care about meta-theoretic uncertainty.

When trying to understand a problem, it is often helpful to reduce it to something simpler. Even if the problem seems as simple as possible, it may still be able to be simplified further. This post will demystify Newcomb's Problem by reducing it to the Prediction Problem, which works as follows:

  • Step 1: An AI called Alpha scans you and makes a prediction about what action you are going to take. A thousand people have already played this game and Alpha was correct every time.
  • Step 2: You pull either the left or the right lever. Neither does anything.
  • Step 3: Alpha gives you $1 million if it predicted that you'd pull the left lever, otherwise it gives you $1000.

The empirical answer seems to be that you ought to pull the left lever. On the other hand, some strictly following Causal Decision Theory ought to be indifferent to the two solutions. After all, the reasoning goes, Alpha has already made their prediction nothing you do now can change this.

At this point someone who thinks they are smarter than they actually are might decide that pulling the left lever may have an upside, but doesn't have a downside, so you may as well pull it and then go about their lives without thinking about this problem any more. This is the way to win if you were actually thrust into such a situation, but a losing strategy if your goal is to actually understand decision theory. I've argued before that practise problems don't need to be realistic, it's also fine if they are trivial. If we can answer why exactly you ought to pull the left lever, then we should also be able to justify one-boxing for Newcomb's Problem and also Timeless Decision Theory.

"Decision Theory" is misleading

The name "decision theory" seems to suggest a focus on making an optimal decision, which then causes the optimal outcome. For the Prediction Problem, the actual decision does absolutely nothing in and of itself, while if I'm correct, the person who pulls the left lever gains 1 million extra dollars. However this is purely as a result of the kind of agent that they are; all the agent has to do in order to trigger this is exist. The decision doesn't actually have any impact apart from the fact that it would be impossible to be the kind of agent that always pulls the left lever without actually pulling the left lever.

The question then arises: do you (roughly) wish to be the kind of agent that gets good outcomes or the kind of agent that makes good decisions? I need to clarify this before it can be answered. "Good outcomes" is evaluated by the expected utility that an agent receives, with the counterfactual being that an agent with a different decision making apparatus encountered this scenario instead. To avoid confusion, we'll refer to these counterfactuals as timeless-counterfactuals and the outcomes as holistic-outcomes and the optimal such counterfactual as holistically-optimal. I'm using "good decisions" to refer to the casual impact of a decision on the outcome. The counterfactuals are the agent "magically" making a different decision at that point, with everything else that happened before being held static, even the decision making faculties of the agent itself. To avoid confusion, we'll refer to these counterfactuals as point-counterfactuals and the decisions over these as point-decisions and the optimal such counterfactual as point-optimal.

I will argue that we should choose good outcomes as the method by which this is obtained is irrelevant. In fact, I would almost suggest using the term Winning Theory instead of Decision Theory. Eliezer made a similar case very elegantly in Newcomb's Problem and Regret of Rationality, but this post aims to identify the exact flaw in the two-boxing argument. Since one-boxing obtains the holistically-optimal outcome, but two-boxing produces the point-optimal decision, I need to show why the former deserves preference.

At this point, we can make two interesting observations. Firstly, two-boxers gain $1000 extra dollars as a result of their decision, but miss out on $1 million dollars as a result of who they are. Why are these two figures accounted for differently? Secondly, both of these approaches are self-affirming after the prediction. At this point, the point-optimal decision is to choose point-optimality and the holistically-optimal decision being to choose holistic-optimality. This might appear a stalemate, but we can resolve this conflict by investigating why point-optimality is usually considered important.

Why do we care about point-optimality anyway?

Both the Prediction Problem and Newcomb's Problem assume that agents don't have libertarian free will; that is, the ability to make decisions unconstrained by the past. For if they did, Alpha wouldn't be able to perfectly or near perfectly predict the agent's future actions from their past state without some kind of backwards causation which would then make one-boxing the obvious choice. So we can assume a deterministic or probabilistically deterministic universe. For simplicity, we'll just work with the former and assume that agents are deterministic.

The absence of free will is important because it affects what exactly we mean by making a decision. Here's what a decision is not: choosing from a variety of options all of which were (in the strictest sense) possible at the time given the past. Technically, only one choice was possible and that was the choice taken. The other choices only become strictly possible when we imagine the agent counter-factually having a different brain state.

The following example may help: Suppose a student has a test on Friday. Reasoning that determinism means that the outcome is already fixed, the student figures that they may as well not bother to study. What's wrong with this reasoning?

The answer is that the outcome is only known to be fixed because whether or not the student studies is fixed. When making a decision, you don't loop over all of the strictly possible options, because there is only one of them and that is whatever you actually choose. Instead, you loop over a set of counterfactuals (and the one actual factual, though you don't know it at the time). While the outcome of the test is fixed in reality, the counterfactuals can have a different outcome as they aren't reality.

So why do we actually care about the point-optimal decision if it can't actually strictly change what you choose as this was fixed from the beginning of time? Well even if you can't strictly change your choice, you can still be fortunate enough to be an agent that always was going to try to calculate the best point-decision and then carry it out (this is effective for standard decision theory problems). If such an agent can't figure out the best point-decision itself, it would choose to pay a trivial amount (say 1 cent) to an oracle find out this out, assuming that the differences in the payoffs aren't similarly trivial. And over a wide class of problems, so long as this process is conducted properly, the agent ends up in the world with the highest expected utility.

So what about the Prediction Problem?

The process described for point-optimality assumes that outcomes are purely a result of actions. But for the Prediction Problem, the outcome isn't dependent on actions at all, but instead on the internal algorithm at time of prediction. Even if our decision doesn't cause the past state that is analysed by Alpha to create its prediction, these are clearly linked in some kind of manner. But point-optimality assumes outcomes are fixed independently of our decision algorithm. The outcomes are fixed for a given agent, but it is empty to say fixed for a given agent whatever its choice as each agent can only make one choice. So allowing any meaningful variation over choices requires allowing variation over agents in which case we can no longer assume that the outcomes are fixed. At this point, whatever the specific relationship, we are outside the intended scope of point-optimal decision making.

Taking this even further, asking, "What choice ought I make?" is misleading because given who you are, you can only make a single choice. Indeed, it seems strange that we care about point-optimality, even in regular decision theory problems, given that point-counterfactuals indicate impossible situations. An agent cannot be such that it would choose X, but then magically choose Y instead, with no casual reason. In fact, I'd suggest that only reason why we care about point-counterfactuals is that they are equivalent to the actually consistent timeless-counterfactuals in normal decision theory problems. After all, in most decision theory problems, we can alter an agent to carry out a particular action at a particular point of time without affecting any other elements of the problem.

Getting more concrete, for the version of the Prediction Problem where we assume Alpha is perfect, you simply cannot pull the right lever and have Alpha predict the left lever. This counterfactual doesn't correspond to anything real, let alone anything that we care about. Instead, it makes much more sense to consider the timeless-counterfactuals, which are the most logical way of producing consistent counterfactuals from point-counterfactual. In this example, the timeless-counterfactuals are pulling the left lever and having Alpha predicting left; or pulling the right lever and having Alpha predicting right.

In the probabilistic version where Alpha correctly identifies you pulling the right lever 90% of the time and the left lever 100% of the time, we will imagine that a ten-sided dice is rolled and Alpha correctly identifies you pulling the right lever as long as the dice doesn't show a ten. You simply cannot pull the right lever with the dice showing a number that is not ten and have Alpha predict you will pull the left lever. Similarly, you cannot pull the right level with the dice showing a ten and have Alpha predict the correct result. The point counterfactuals allow this, but these situations are inconsistent. In contrast, the timeless counterfactuals insist on consistency between the dice-roll and your decision, so actually correspond to something meaningful.

If you are persuaded to reject point-optimality, I would suggest a switch to a metric built upon a notion of good outcomes instead for two reasons. Firstly, point-optimality is ultimately motivated by the fact that it provides good outcomes within a particular scope. Secondly, both the one-boxers and two-boxers see their strategy as producing better outcomes.

In order to make this work, we just need to formulate good outcomes in a way that accounts for agents being predestined to perform strategies as opposed to agents exercising some kind of libertarian free will. The natural way to do this is to work with holistic-counterfactuals instead of point-counterfactuals.

But doesn't this require backwards causation?

How can a decision affect a prediction at an earlier time? Surely this should be impossible. If human adopts the timeless approach in the moment it's because either:

a) They were fooled into it by reasoning that sounded convincing, but was actually flawed

b) They realised that the timeless approach best achieves their intrinsic objectives, even accounting for the past being fixed. For example, they value whatever currency is offered in the experiment and they ultimately value achieving the best outcome in these terms, then they realise that Timeless Decision Theory delivers this.

Remembering that agent's "choice" of what decision theory to adopt is already predestined, even if the agent only figured this out when faced with this situation. You don't really make a decision in the sense we usually think about it; instead you are just following inevitable process. For an individual who ultimately values outcomes as per b), the only question is whether the individual will carry out this process of producing a decision theory that matches their intrinsic objectives correctly or incorrectly. An individual who adopts the timeless approach wins because Alpha knew that they were going to carry out this process correctly, while an individual who adopts point-optimality loses because Alpha knew they were always going to make a mistake in this process.

The two-boxers are right that you can only be assured of gaining the million if you are pre-committed in some kind of manner, although they don't realise that determinism means that we are all pre-committed in a general sense to whatever action we end up taking. That is, in addition to explicit pre-commitments, we can also talk about implicit pre-commitments. An inevitable flaw in reasoning as per a) is equivalent to pre-commitment, although from the inside it will feel as though you could have avoided it. So are unarticulated intrinsic objectives that are only identified and clarified at the point of the decision as per b); clarifying these objectives doesn't cause you to become pre-committed, it merely reveals what you were pre-committed to. Of course, this only works with super-human predictors. Normal people can't be relied upon to pick up on these deep aspects of personality and so require more explicit pre-commitment in order to be convinced (I expanded this into a full article here).

What about agents that are almost pre-committed to a particular action? Suppose 9/10 times you follow the timeless approach, but 1/10 you decided to do the opposite. More specifically, we'll assume that when a ten-sided dice roll shows a 10, you experience a mood that convinces you to take the later course of action. Since we're assuming determinism, Alpha will be aware of this before they make their prediction. When the dice shows a ten, you feel really strongly that you have exercised free will as you would have acted differently in the counterfactual where your mood was slightly different. However, given that the dice did show a ten, your action was inevitable. Again, you've discovered your decision rather than made it. For example, if you decide to be irrational, the predictor knew that you were in that mood at the start, even if you did not.

Or going further, a completely rational agent that wants to end up in the world with the most dollars doesn't make that decision in the Prediction Problem so that anything happens, it makes that decision because it can make no other. If you make another decision, you either have different objectives or you have an error in your reasoning, so you weren't the agent that you thought you were.

When you learn arguments in favour of one side or another, it changes what your choice would have been in the counterfactual where you were forced to make a decision just before you made that realisation, but what happens in reality is fixed. It doesn't change the past either, but it does change your estimation of what Alpha would have predicted. When you lock in your choice, you've finalised your estimate of the past, but this looks a lot like changing the past, especially if you had switched to favouring a different decision at the last minute. Additionally, when you lock in your choice it isn't like the future was just locked in at that time as it were already fixed. Actually, making a decision can be seen as a process that makes the present line up with the past predictions and again this can easily be mistaken for changing the past.

But further than this, I want to challenge the question: "How does my decision affect a past prediction?" Just like, "What choice ought I make?", if we contemplate a fixed individual, then we must fix the decision as well. If instead we consider a variety of individuals taking a variety of actions, then the question becomes, "How does a individual/decision pair affect a prediction prior to the decision?", which isn't exactly a paradox.

Further Reading:

Anna Salamon started writing an incomplete sequence on this problem. I only read them after finishing the first version of this post, but she provides a better explanation that I do of why we need to figure out what kind of counterfactual we are talking about, what exactly "should", "would" and "could" mean, and what the alternatives are.

New Comment
11 comments, sorted by Click to highlight new comments since: Today at 7:54 PM

I usually just think about which decision theory we'd want to program into an AI which might get copied, its source code inspected, etc. That lets you get past the basic stuff, like Newcomb's Problem, and move on to more interesting things. Then you can see which intuitions can be transferred back to problems involving humans.

It turns out that many of the complications (multiple players, amnesia, copying, predictors, counterfactuals) lead to the same idea: that we should model things game-theoretically and play the global optimal strategy no matter what, instead of trying to find the optimal decision locally at each node. That idea summarizes a large part of UDT (Wei's original proposals of UDT also included dealing with logical uncertainty, but that turned out to be much harder.) Hence my recent posts on how to model anthropic updates and predictors game-theoretically.

"I usually just think about which decision theory we'd want to program into an AI which might get copied, its source code inspected" - well everyone agrees that if you can pre-commit to one-box you ought to. The question is what about if you're in the situation and you haven't pre-committed. My answer is that if you take a choice, then you were implicitly pre-committed.

People who follow UDT don't need to precommit, they have a perfectly local decision procedure: think back, figure out the best strategy, and play a part in it. The question of precommitment only arises if you follow CDT, but why would you follow CDT?

Correct, but you're justifying UDT by arguing what you should do if you had pre-committed. A two-boxer would argue that this is incorrect because you haven't pre-committed.

The idea of playing the best strategy can stand on its own, it doesn't need to be justified by precommitment. I'd say the idea of myopically choosing the next move needs justification more.

For example, when you're dealt a weak hand in poker, the temptation to fold is strong. But all good players know you must play aggressively on your weakest hands, because if you fold, you might as well light up a neon sign saying "I have a strong hand" whenever you do play aggressively, allowing your opponent to fold and cut their losses. In this case it's clear that playing the best strategy is right, and myopically choosing the next move is wrong. You don't need precommitment to figure it out. Sure, it's a repeated game where your opponent can learn about you, but Newcomb's Problem has a predictor which amounts to the same thing.

Thank you for posting this! I'm posting here for the first time, although I've spent a significant amount of time reading the Sequences already (I just finished Seeing with Fresh Eyes). The comments on determinism cleared up a few uncertainties about Newcomb's Problem for me.

When I have explained the problem to others, I have usually used the phrasing where Alpha is significantly better than average at predicting what you will choose, but not perfect. (This helps reduce incredulity on the part of the average listener.) I have also used the assumption that Alpha does this by examining your mental state, rather than by drawing causal arrows backward in time. One of my friends suggested precommitting to a strategy that one-boxes 51% of the time and two-boxes 49% of the time, chosen at the time you receive the boxes by some source that is agreed to be random such as rolling two d10's. His logic is that Alpha would probably read your mind accurately, and that if he did, he would decide based on your mental state to put the money in the box, since you are more likely to one-box than not.

This seemed like a very good strategy (assuming the logic and the model of the problem are correct, which is far from certain), and I wondered why this strategy wasn't at least being discussed more. It seems that most other people were assuming determinism while I was assuming libertarian free will.

What do all of you think of my friend's strategy?

Is the assumption of determinism a comment on the actual state of the universe, or simply a necessary assumption to make the problem interesting?

Well that would work for a predictor that 100% predicts your most likely strategy. If the predictor has even a slight chance of predicting the 49% strategy instead of the 51% strategy, you'll lose out as you're risking a million to gain a thousand. But yes, the discussion in my post assumes that the predictor can predict any sources of randomness that you have access to.

I upvoted because I like having new perspectives on things, even if I have read others previously. It seems the harder the concept, the more beneficial are additional presentations.

I therefore expect it will be a long time until we stop benefiting from new presentations of Newcomb.

I often think of "determinism" as too strong a word for what's going on. The past is fixed, and the past influences the present, but that doesn't exactly mean that the present is determined wholly by the past, but the past as we see it from the present can be no other way than what it is and can have no other effect than what it has. This doesn't mean the present and future are fixed unless we want to commit to a particular metaphysical claim about the universe; instead it just means that the past is "perfect" or complete and we move forward from there. We can then reasonably admit all sorts of ways the past need not determine the future while also acknowledging that the future is causally linked the a fixed past.

Well, I also explained how libertarian free will breaks the scenario and once we've excluded that we should be able to assume a form of determinism or probabilistic determinism.

To me this is not really a matter of whether or not we have libertarian free will. In fact I think we don't and need not posit it to explain anything. My point is perhaps more that when we talk about "determinism" it's often mixed up with ideas about a clock-work universe that flows forward in a particular way that we can calculate in advance, but the computation is so complex that the only way to do it is to actually let time advance so the computation plays out in the real world, and thus although the present and future may be linked to the past it can't be known as well as we can possibly know it until we get there.