Counterfactual Mugging

In Two Types of Updatelessness, makes a distinction between all-upside updatelessness and mixed-upside updatelessness. In all-upside case, utilising an updateless decision theory provides a better result in the current situation, while in a mixed-upside case the benefits go to other possible selves. Unlike Newcomb's Problem or Parfait'Parfit's Hitchhiker, Counterfactual Mugging is a mixed-upside case.

Depending on how the problem inis phrased, intuition calls for different answers. For example, Eliezer Yudkowsky has argued that framing the problem in a way Omega is a regular aspect of the environment which regularly asks such types of questions makes most people answer 'Yes'. However, Vladimir Nesov points out that Rationalists Should Win could be interpreted as suggesting that we should not pay. After all, even though paying in the tails case would cause you to do worse in the counterfactual where the coin came up heads, you already know the counterfactual didn't happen, so it's not obvious that you should pay. This issue has been discussed in this question.

Formal decision theories also diverge. For Causal Decision Theory, you can only affect those probabilities that you are causalcausally linked to. Hence, the answer should be 'No'. In Evidential Decision Theory any kind of connection is accounted, then the answer should be 'No'. Timeless Decision Theory answer seems undefined, however Yudkowsky has argued that if the problem is recurrently presented, one should answer 'Yes' on the basis of enhancing its probability of gaining $10000 in the next round. This seems to be Causal Decision Theory prescription as well. Updateless decision theory1 prescribes giving the $100, on the basis your decision can influence both the 'heads branch' and 'tails branch' of the universe.

Omega, a perfect predictor, flips a coin. If ifit comes up tails Omega asks you for $100. If it comes up heads, Omega pays you $10,000 if it predicts that you would have paid if it had come up tails.

I was careful to avoid endorsing any conclusion in the post in order to foment discussion. If we add in my own position (that paying up is winning), the point is that the conception of winning that only cares about actuality is wrong. This does rest on the possibility of the unfortunate interpretation of winning as only caring about actuality, and therefore suggesting that we shouldn't pay, though saying that I was pointing that out seems like a stretch. So maybe it's kinda correct to say that I was pointing it out, but seriously misleading, to the point where it misled even me...

I'd guess that line was referencing this:

And so I ask you all: is the decision to give up $100 when you have no real benefit from it, only counterfactual benefit, an example of winning?

From your Counterfactual Mugging post.

However, Vladimir Nesov points out that Rationalists Should Win could be interpreted as suggesting that we should not pay.

I'm not familiar with this position. Citation needed?

4ESRogs3y
I'd guess that line was referencing this: From your Counterfactual Mugging post.
2Vladimir_Nesov3y
I was careful to avoid endorsing any conclusion in the post in order to foment discussion. If we add in my own position (that paying up is winning), the point is that the conception of winning that only cares about actuality is wrong. This does rest on the possibility of the unfortunate interpretation of winning as only caring about actuality, and therefore suggesting that we shouldn't pay, though saying that I was pointing that out seems like a stretch. So maybe it's kinda correct to say that I was pointing it out, but seriously misleading, to the point where it misled even me...

Regardless of the particular decision theory, it is generally agreed that if you can pre-commit in advance that you should do so. The dispute is purely over what you should do if you didn't pre-commit.