Related to: Can Counterfactuals Be True?, Newcomb's Problem and Regret of Rationality.

Imagine that one day, Omega comes to you and says that it has just tossed a fair coin, and given that the coin came up tails, it decided to ask you to give it $100. Whatever you do in this situation, nothing else will happen differently in reality as a result. Naturally you don't want to give up your $100. But see, Omega tells you that if the coin came up heads instead of tails, it'd give you $10000, but only if you'd agree to give it $100 if the coin came up tails.

Omega can predict your decision in case it asked you to give it $100, even if that hasn't actually happened, it can compute the counterfactual truth. Omega is also known to be absolutely honest and trustworthy, no word-twisting, so the facts are really as it says, it really tossed a coin and really would've given you $10000.

From your current position, it seems absurd to give up your $100. Nothing good happens if you do that, the coin has already landed tails up, you'll never see the counterfactual $10000. But look at this situation from your point of view before Omega tossed the coin. There, you have two possible branches ahead of you, of equal probability. On one branch, you are asked to part with $100, and on the other branch, you are conditionally given $10000. If you decide to keep $100, the expected gain from this decision is $0: there is no exchange of money, you don't give Omega anything on the first branch, and as a result Omega doesn't give you anything on the second branch. If you decide to give $100 on the first branch, then Omega gives you $10000 on the second branch, so the expected gain from this decision is

-$100 * 0.5 + $10000 * 0.5 = $4950

So, this straightforward calculation tells that you ought to give up your $100. It looks like a good idea before the coin toss, but it starts to look like a bad idea after the coin came up tails. Had you known about the deal in advance, one possible course of action would be to set up a precommitment. You contract a third party, agreeing that you'll lose $1000 if you don't give $100 to Omega, in case it asks for that. In this case, you leave yourself no other choice.

But in this game, explicit precommitment is not an option: you didn't know about Omega's little game until the coin was already tossed and the outcome of the toss was given to you. The only thing that stands between Omega and your 100$ is your ritual of cognition. And so I ask you all: is the decision to give up $100 when you have no real benefit from it, only counterfactual benefit, an example of winning?

P.S. Let's assume that the coin is deterministic, that in the overwhelming measure of the MWI worlds it gives the same outcome. You don't care about a fraction that sees a different result, in all reality the result is that Omega won't even consider giving you $10000, it only asks for your $100. Also, the deal is unique, you won't see Omega ever again.

Counterfactual Mugging
New Comment
296 comments, sorted by Click to highlight new comments since:
Some comments are truncated due to high volume. (⌘F to expand all)Change truncation settings
[-]mwaser360

Imagine that one day you come home to see your neighbors milling about your house and the Publisher's Clearinghouse (PHC) van just pulling away. You know that PHC has been running a new schtick recently of selling $100 lottery tickets to win $10,000 instead of just giving money away. In fact, you've used that very contest as a teachable moment with your kids to explain how once the first ticket of the 100 printed was sold, scratched, and determined not to be the winner -- that the average expected value of the remaining tickets was greater than their cost and they were therefore increasingly worth buying. Now, it's weeks later, most of the tickets have been sold, scratched, and not winners and they came to your house. In fact, there were only two tickets remaining. And you weren't home. Fortunately, your neighbor and best friend Bob asked if he could buy the ticket for you. Sensing a great human interest story (and lots of publicity), PHC said yes. Unfortunately, Bob picked the wrong ticket. After all your neighbors disperse and Bob and you are alone, Bob says that he'd really appreciate it if he could get his hundred dollars back. Is he mugging you? Or, do you give it to him?

8pnrjulius
Yes, I think you still owe him the $100. But I like how you made it into a relatively realistic scenario.
4lsparrish
Considering the ticket was worth $5,000 when he bought it, sure.
7mwaser
Did you give the same answer to Omega? The cases are exactly analogous. (Or do you argue that they are not?)

The disanalogy here is that you have a long term social relationship with Bob that you don't have with Omega, and the $100 are an investment into that relationship.

2thrawnca
Also, there is the possibility of future scenarios arising in which Bob could choose to take comparable actions, and we want to encourage him in doing so. I agree that the cases are not exactly analogous.

The outcomes don't seem to be tied together as they were in the original problem; is it true that if had he won, he would only then have given you the money if, had he not won, you would have given him the $100 back? That isn't clear.

The counterfactual anti-mugging: One day No-mega appears. No-mega is completely trustworthy etc. No-mega describes the counterfactual mugging to you, and predicts what you would have done in that situation not having met No-mega, if Omega had asked you for $100.

If you would have given Omega the $100, No-mega gives you nothing. If you would not have given Omega $100, No-mega gives you $10000. No-mega doesn't ask you any questions or offer you any choices. Do you get the money? Would an ideal rationalist get the money?

Okay, next scenario: you have a magic box with a number p inscribed on it. When you open it, either No-mega comes out (probability p) and performs a counterfactual anti-mugging, or Omega comes out (probability 1-p), flips a fair coin and proceeds to either ask for $100, give you $10000, or give you nothing, as in the counterfactual mugging.

Before you open the box, you have a chance to precommit. What do you do?

9Eliezer Yudkowsky
I would have no actionable suspicion that I should give Omega the $100 unless I knew about No-mega. So I get the $10000 only if No-mega asks the question "What would Eliezer do knowing about No-mega?" and not if No-mega asks the question "What would Eliezer do not knowing about No-mega?"
3wafflepudding
You forgot about MetaOmega, who gives you $10,000 if and only if No-mega wouldn't have given you anything, and O-mega, who kills your family unless you're an Alphabetic Decision Theorist. This comment doesn't seem specifically anti-UDT -- after all, Omega and No-mega are approximately equally likely to exist; a ratio of 1:1 if not an actual p of .5 -- but it still has the ring of Just Cheating. Admittedly, I don't have any formal way of telling the difference between decision problems that feel more or less legitimate, but I think part of the answer might be that the Counterfactual Mugging isn't really about how to act around superintelligences: It illustrates a more general need to condition our decisions based on counterfactuals, and as EY pointed out, UDT still wins the No-mega problem if you know about No-mega, so whether or not we should subscribe to some decision theory isn't all that dependent on which superintelligences we encounter. I'm necroing pretty hard and might be assuming too much about what Caspian originally meant, so the above is more me working this out for myself than anything else. But if anyone can explain why the No-mega problem feels like cheating to me, that would be appreciated.
1Vladimir_Nesov
Do you have a point?

Yes, that there can just as easily be a superintelligence that rewards people predicted to act one way as one that rewards people predicted to act the other. Which precommitment is most rational depends depends on the which type you expect to encounter.

I don't expect to encounter either, and on the other hand I can't rule out fallible human analogues of either. So for now I'm not precommitting either way.

You don't precommit to "give away the $100, to anyone who asks". You precommit to give away the $100 in exactly the situation I described. Or, generalizing such precommitments, you just compute your decisions on the spot, in a reflectively consistent fashion. If that's what you want do to with your future self, that is.

[-]Jonii120

there can just as easily be a superintelligence that rewards people predicted to act one way as one that rewards people predicted to act the other.

Yeah, now. But after Omega really, really, appears in front of you, chance of Omega existing is about 1. Chance of No-Mega is still almost non-existent. In this problem, existence of Omega is given. It's not something you are expecting to encounter now, just as we're not expecting to encounter eccentric Kavkan billionaires that will give you money for toxicating yourself. The Kavka's Toxin and the counterfactual mugging present a scenario that is given, and ask you how would you act then.

2David Scott Krueger (formerly: capybaralet)
But you aren't supposed to be updating... the essence of UDT, I believe, is that your policy should be set NOW, and NEVER UPDATED. So... either: 1. You consider the choice of policy based on the prior where you DIDN'T KNOW whether you'd face Nomega or Omega, and NEVER UPDATE IT (this seems obviously wrong to me: why are you using your old prior instead of your current posterior?). or 2. You consider the choice of policy based on the prior where you KNOW that you are facing Omega AND that the coin is tails, in which case paying Omega only loses you money.
2TheWakalix
It doesn't prevent doing different actions in different circumstances, though. That's not what "updateless" means. It means that you should act as your past self would have precommitted to doing in your situation. Your probability estimate for "I see Omega" should be significantly greater than "I see Omega, and also Nomega is watching and deciding how to act", so your decision should be mostly determined by Omega, not Nomega. (The Metanomega also applies - there's a roughly equal chance of Metanomega or Nomega waiting and watching. [Metanomega = Nomega reversed; gives payoff iff predicts you paying.])
2TheWakalix
I see where I went wrong. I assumed that the impact of one's response to Omega is limited to the number of worlds in which Omega exists. That is, my reasoning is invalid if ("what I do in scenario X" is meaningful and affects the world even if scenario X never happens). In other words, when one is being counterfactually modeled, which is exactly the topic of discussion.
0David Scott Krueger (formerly: capybaralet)
Thanks for pointing that out. The answer is, as expected, a function of p. So I now find explanations of why UDT gets mugged incomplete and misleading. Here's my analysis: The action set is {give, don't give}, which I'll identify with {1, 0}. Now, the possible deterministic policies are simply every mapping from {N,O} --> {1,0}, of which there are 4. We can disregard the policies for which pi(N) = 1, since giving money to Nomega serves no purpose. So we're left with pi_give and pi_don't, which give/don't, respectively, to Omega. ---------------------------------------- Now, we can easily compute expected value, as follows: r (pi_give(N)) = 0 r (pi_give(O, heads)) = 10 r (pi_give(0, tails)) = -1 r (pi_don't(N)) = 10 r (pi_don't(0)) = 0 So now: Eg := E_give(r) = 0 p + .5 (10-1) * (1-p) Ed := E_don't(r) = 10 p + 0 (1-p) Eg > Ed whenever 4.5 (1-p) > 10 p, i.e. whenever 4.5 > 14.5 p i.e. whenever 9/29 > p So, whether you should precommit to being mugged depends on how likely you are to encounter N vs. O, which is intuitively obvious.

Philosopher Kenny Easwaran reported in 2007 that:

Josh von Korff, a physics grad student here at Berkeley, and versions of Newcomb’s problem. He shared my general intuition that one should choose only one box in the standard version of Newcomb’s problem, but that one should smoke in the smoking lesion example. However, he took this intuition seriously enough that he was able to come up with a decision-theoretic protocol that actually seems to make these recommendations. It ends up making some other really strange predictions, but it seems interesting to consider, and also ends up resembling something Kantian!

The basic idea is that right now, I should plan all my future decisions in such a way that they maximize my expected utility right now, and stick to those decisions. In some sense, this policy obviously has the highest expectation overall, because of how it’s designed.

Korff also reinvents counterfactual mugging:

Here’s another situation that Josh described that started to make things seem a little more weird. In Ancient Greece, while wandering on the road, every day one either encounters a beggar or a god. If one encounters a beggar, then one can choose to either give the b

... (read more)

In Ancient Greece, while wandering on the road, every day one either encounters a beggar or a god.

If it's an iterated game, then the decision to pay is a lot less unintuitive.

My two bits: Omega's request is unreasonable.

Precommitting is something that you can only do before the coin is flipped. That's what the "pre" means. Omega's game rewards a precommitment, but Omega is asking for a commitment.

Precommitting is a rational thing to do because before the coin toss, the result is unknown and unknowable, even by Omega (I assume that's what "fair coin" means). This is a completely different course of action than committing after the coin toss is known! The utility computation for precommitment is not and should not be the same as the one for commitment.

In the example, you have access to information that pre-you doesn't (the outcome of the flip). If rationalists are supposed to update on new information, then it is irrational for you to behave like pre-you.

0[anonymous]
Precommittment does make one boxing on Newcomblike problems a whole lot easier. But it isn't necessarily required. That's why Vladimir made an effort to exculde precommitment. I don't agree. I suggest that pre-you has exactly the same information that you have. The pre-you must be considered to have been given exactly the same inputs as you to the extent that they influence the decision. That is implied by the ability of the Omega to make the accurate prediction that we have been assured he made.
2Sideways
By definition, pre-you only has access to the coin's probability distribution, while you have access to the result of the coin flip. Surely you don't mean to say that's the same thing? From the perspective of a non-superintelligence, Omega's prediction abilities are indistinguishable from magic. Human beings can't tell what they "imply." Trying to figure out the implications with a primate brain will only get you into a paradox like claiming a fact is the same as a probability distribution. All we can reasonably do is stipulate Omega's abilities needed to make the problem work and no further.

We're assuming Omega is trustworthy? I'd give it the $100, of course.

[-]MBlume490

Had the coin come up differently, Omega might have explained the secrets of friendly artificial general intelligence. However, he now asks that you murder 15 people.

Omega remains completely trustworthy, if a bit sick.

Ha, I'll re-raise: Had the coin come up differently, Omega would have filled ten Hubble volumes with CEV-output. However, he now asks that you blow up this Hubble volume.

(Not only do you blow up the universe (ending humanity for eternity) you're glad that Omega showed to offer this transparently excellent deal. Morbid, ne?)

8AndySimpson
Ouch.
[-]MBlume200

For some reason, raising the stakes in these hypotheticals to the point of actual pain has become reflex for me. I'm not sure if it's to help train my emotions to be able to make the right choices in horrible circumstances, or just my years in the Bardic Conspiracy looking for an outlet.

6jimrandomh
Raising the stakes in this way does not work, because of the issue described in Ethical Injunctions: it is less likely that Omega has presented you with this choice, than that you have gone insane.

So imagine yourself in the most inconvenient possible world where Omega is a known feature of the environment and has long been seen to follow through on promises of this type; it does not particularly occur to you or anyone that believing this fact makes you insane.

When I phrase it that way - imagine myself in a world full of other people confronted by similar Omega-induced dilemmas - I suddenly find that I feel substantially less uncomfortable; indicating that some of what I thought was pure ethical constraint is actually social ethical constraint. Still, it may function to the same self-protective effect as ethical constraint.

To add to the comments below, if you're going to take this route, you might as well have already decided that encountering Omega at all is less likely than that you have gone insane.

[-]jimmy110

That may be true, but it's still a dodge. Conditional on not being insane, what's your answer?

Additionally, I don't see why Omega asking you to give it 100 dollars vs 15 human lives necessarily crosses the threshold of "more likely that I'm just a nutbar". I don't expect to talk to Omega anytime soon...

4[anonymous]
We're assuming Omega is trustworthy? I'd murder 15 people, of course. I'll note that the assumption that I trust the Omega up to stakes this high is a big one. I imagine that the alterations being done to my brain in the counterfactualisation process would have rather widespread implications on many of my thought processes and beliefs once I had time to process it.
3MBlume
Completely agreed, a major problem in any realistic application of such scenarios. I'm afraid I don't follow.
[-]kurige110

Can you please explain the reasoning behind this? Given all of the restrictions mentioned (no iterations, no possible benefit to this self) I can't see any reason to part with my hard earned cash. My "gut" says "Hell no!" but I'm curious to see if I'm missing something.

[-]MBlume320

There are various intuition pumps to explain the answer.

The simplest is to imagine that a moment from now, Omega walks up to you and says "I'm sorry, I would have given you $10000, except I simulated what would happen if I asked you for $100 and you refused". In that case, you would certainly wish you had been the sort of person to give up the $100.

Which means that right now, with both scenarios equally probable, you should want to be the sort of person who will give up the $100, since if you are that sort of person, there's half a chance you'll get $10000.

If you want to be the sort of person who'll do X given Y, then when Y turns up, you'd better bloody well do X.

If you want to be the sort of person who'll do X given Y, then when Y turns up, you'd better bloody well do X.

Well said. That's a lot of the motivation behind my choice of decision theory in a nutshell.

[-]MBlume110

Thanks, it's good to know I'm on the right track =)

I think this core insight is one of the clearest changes in my thought process since starting to read OB/LW -- I can't imagine myself leaping to "well, I'd hand him $100, of course" a couple years ago.

9thomblake
I think this describes one of the core principles of virtue theory under any ethical system. I wonder how much it depends upon accidents of human psychology, like our tendency to form habits, and how much of it is definitional (if you don't X when Y, then you're simply not the sort of person who Xes when Y)
8kurige
That's not the situation in question. The scenario laid out by Vladimir_Nesov does not allow for an equal probability of getting $10000 and paying $100. Omega has already flipped the coin, and it's already been decided that I'm on the "losing" side. Join that with the fact that me giving $100 now does not increase the chance of me getting $10000 in the future because there is no repetition. Perhaps there's something fundamental I'm missing here, but the linearity of events seems pretty clear. If Omega really did calculate that I would give him the $100 then either he miscalculated, or this situation cannot actually occur. -- EDIT -- There is a third possibility after reading Cameron's reply... If Omega is correct and honest, then I am indeed going to give up the money. But it's a bit of a trick question, isn't it? I'm going to give up the money because Omega says I'm going to give up the money and everything Omega says is gospel truth. However, if Omega hadn't said that I would give up the money, then I wouldn't of given up the money. Which makes this a bit of an impossible situation. Assuming the existence of Omega, his intelligence, and his honesty, this scenario is an impossibility.
[-]MBlume250

I feel like a man in an Escher painting, with all these recursive hypothetical mes, hypothetical kuriges, and hypothetical omegas.

I'm saying, go ahead and start by imagining a situation like the one in the problem, except it's all happening in the future -- you don't yet know how the coin will land.

You would want to decide in advance that if the coin came up against you, you would cough up $100.

The ability to precommit in this way gives you an advantage. It gives you half a chance at $10000 you would not otherwise have had.

So it's a shame that in the problem as stated, you don't get to precommit.

But the fact that you don't get advance knowledge shouldn't change anything. You can just decide for yourself, right now, to follow this simple rule:

If there is an action to which my past self would have precommited, given perfect knowledge, and my current preferences, I will take that action.

By adopting this rule, in any problem in which the oppurtunity for precommiting would have given you an advantage, you wind up gaining that advantage anyway.

0[anonymous]
That one sums it all up nicely!
[-]MBlume220

I'm actually not quite satisfied with it. Probability is in the mind, which makes it difficult to know what I mean by "perfect knowledge". Perfect knowledge would mean I also knew in advance that the coin would come up tails.

I know giving up the $100 is right, I'm just having a hard time figuring out what worlds the agent is summing over, and by what rules.

ETA: I think "if there was a true fact which my past self could have learned, which would have caused him to precommit etc." should do the trick. Gonna have to sleep on that.

ETA2: "What would you do in situation X?" and "What would you like to pre-commit to doing, should you ever encounter situation X?" should, to a rational agent, be one and the same question.

ETA2: "What would you do in situation X?" and "What would you like to pre-commit to doing, should you ever encounter situation X?" should, to a rational agent, be one and the same question.

...and that's an even better way of putting it.

2[anonymous]
Note that this doesn't apply here. It's "What would you do if you were counterfactually mugged?" versus "What would you like to pre-commit to doing, should you ever be told about the coin flip before you knew the result?". X isn't the same.
9Vladimir_Nesov
MBlume: This phrasing sounds about right. Whatever decision-making algorithm you have drawing your decision D when it's in situation X, should also come to the same conditional decision before the situation X appeared, "if(X) then D". If you actually don't give away $100 in situation X, you should also plan to not give away $100 in case of X, before (or irrespective of whether) X happens. Whichever decision is the right one, there should be no inconsistency of this form. This grows harder if you must preserve the whole preference order.
7conchis
"Perfect knowledge would mean I also knew in advance that the coin would come up tails." This seems crucial to me. Given what I know when asked to hand over the $100, I would want to have pre-committed to not pre-committing to hand over the $100 if offered the original bet. Given what I would know if I were offered the bet before discovering the outcome of the flip I would wish to pre-commit to handing it over. From which information set I should evaluate this? The information set I am actually at seems the most natural choice, and it also seems to be the one that WINS (at least in this world). What am I missing?
-5fractalman
1The_Duck
Not if precommiting potentially has other negative consequences. As Caspian suggested elsewhere in the thread, you should also consider the possibility that the universe contains No-megas who punish people who would cooperate with Omega.
1MBlume
...why should you also consider that possibility?
7The_Duck
Because if that possibility exists, you should not necessarily precommit to cooperate with Omega, since that risks being punished by No-mega. In a universe of No-megas, precommiting to cooperate with Omega loses. This seems to me to create a distinction between the questions "what would you do upon encountering Omega?" and "what will you now precommit to doing upon encountering Omega?" I suppose my real objection is that some people seem to have concluded in this thread that the correct thing to do is to, in advance, make some blanket precommitment to do the equivalent of cooperating with Omega should they ever find themselves in any similar problem. But I feel like these people have implicitly made some assumptions about what kind of Omega-like entities they are likely to encounter: for instance that they are much more likely to encounter Omega than No-mega.
2pengvado
But No-mega also punishes people who didn't precommit but would have chosen to cooperate after meeting Omega. If you think No-mega is more likely than Omega, then you shouldn't be that kind of person either. So it still doesn't distinguish between the two questions.
-5fractalman
5Nebu
I don't see this situation is impossible, but I think it's because I've interpreted it differently from you. First of all, I'll assume that everyone agrees that given a 50/50 bet to win $10'000 versus losing $100, everyone would take the bet. That's a straightforward application of utilitarianism + probability theory = expected utility, right? So Omega correctly predicts that you would have taken the bet if he had offered it to you (a real no brainer; I too can predict that you would have taken the bet had he offered it). But he didn't offer it to you. He comes up now, telling you that he predicted that you would accept the bet, and then carried out the bet without asking you (since he already knew you would accept the bet), and it turns out you lost. Now he's asking you to give him $100. He's not predicting that you will give him that number, nor is he demanding or commanding you to give it. He's merely asking. So the question is, do you do it? I don't think there's any inconsistency in this scenario regardless of whether you decide to give him the money or not, since Omega hasn't told you what his prediction would be (though if we accept that Omega is infallible, then his prediction is obviously exactly whatever you would actually do in that situation).
5MBlume
Omega hasn't told you his predictions in the given scenario.
1[anonymous]
That's absolutely true. In exactly the same way, if the Omega really did calculate that I wouldn't give him the $100 then either he miscalculated, or this situation cannot actually occur. The difference between your counterfactual instance and my counterfactual instance is that yours just has a weird guy hassling you with deal you want to reject while my counterfactual is logically inconsistent for all values of 'me' that I identify as 'me'.
4kurige
Thank you. Now I grok. So, if this scenario is logically inconsistent for all values of 'me' then there really is nothing that I can learn about 'me' from this problem. I wish I hadn't thought about it so hard.
0[anonymous]
Logically inconsistent for all values of '' that would hand over the $100. For all values of '' that would keep the $100 it is logically consistent but rather obfuscated. It is difficult to answer a multiple choice question when considering the correct answer throws null.
4ArisKatsaris
I liked this position -- insightful, so I'm definitely upvoting. But I'm not altogether convinced it's a completely compelling argument. With the amounts reversed, Omega could have walked up to you and said "I would have given you $100 except if I asked you for $10.000 you would have refused." You'd then certainly wish to have been the sort of person to counterfactually have given up the $10000, because in the real world it'd mean you'd get $100, even though you'd certainly REJECT that bet if you had a choice for it in advance.
3endoself
Not necessarily; it depends on relative frequency. If Omega has a 10^-9 chance of asking me for $10000 and otherwise will simulate my response to judge whether to give me $100, and if I know that (perhaps Omega earlier warned me of this), I would want to be the type of person who gives the money.
2John_Maxwell
Is that an acceptable correction?
[-]MBlume120

Well, with a being like Omega running around, the two become more or less identical.

2John_Maxwell
If we're going to invent someone who can read thoughts perfectly, we may as well invent someone who can conceal thoughts perfectly. Anyway, there aren't any beings like Omega running around to my knowledge. If you think that concealing motivations is harder than I think, and that the only way to make another human think you're a certain way is to be that way, say that.
1swestrup
And if Omega comes up to me and says "I was going to kill you if you gave me $100. But since I've worked out that you won't, I'll leave you alone." then I'll be damn glad I wouldn't agree. This really does seem like pointless speculation. Of course, I live in a world where there is no being like Omega that I know of. If I knew otherwise, and knew something of their properties, I might govern myself differently.
7MBlume
We're not talking Pascal's Wager here, you're not guessing at the behaviour of capricious omnipotent beings. Omega has told you his properties, and is assumed to be trustworthy.
4swestrup
You are stating that. But as far as I can tell Omega is telling me its a capricious omnipotent being. If there is a distinction, I'm not seeing it. Let me break it down for you: 1) Capricious -> I am completely unable to predict its actions. Yes. 2) Omnipotent -> Can do the seemingly impossible. Yes. So, what's the difference?
6bogdanb
It's not capricious in the sense you give: you are capable of predicting some of its actions: because it's assumed Omega is perfectly trustworthy, you can predict with certainty what it will do if it tells you what it will do. So, if it says it'll give you 10k$ in some condition (say, if you one-box its challenge), you can predict that it'll give it the money if that condition arises. If it were capricious in the sense of complete inability of being predicted, it might amputate three of your toes and give you a flower garland. Note that the problem supposes you do have certainty that Omega is trustworthy; I see no way of reaching that epistemological state, but then again I see no way Omega could be omnipotent, either. ---------------------------------------- On an somewhat unrelated note, why would Omega ask you for 100$ if it had simulated you wouldn't give it the money? Also, why would it do the same if it had simulated you would give it the money? What possible use would an omnipotent agent have for 100$?
0fractalman
Omega is assumed to be mildly bored and mildly anthropic. And his asking you for 100$ could always be PART of the simulation.
1bogdanb
Yes, it’s quite reasonable that if it was curious about you it would simulate you and ask the simulation a question. But once it did that, since the simulation was perfect, why would it waste the time to ask the real you? After all, in the time it takes you to understand Omega’s question it could probably simulate you many times over. So I’m starting to think that encountering Omega is actually pretty strong evidence for the fact that you’re simulated.
0Jiro
Maybe Omega recognizes in advance that you might think this way, doesn't want it to happen, and so precommits to asking the real you. With the existence of this precommitment, you may not properly make this reasoning. Moreover, you should be able to figure out that Omega would precommit, thus making it unnecessary for him to explicitlyy tell you he's doing so.
0bogdanb
(Emphasis mine.) I don’t think, given the usual problem formulation, that one can figure out what Omega wants without Omega explicitly saying it, and maybe not even in that case. It’s a bit like a deal with a not-necessarily-evil devil. Even if it tells you something and you’re sure it’s not lying and you think you the wording is perfectly clear, you should still assign a very high probability that you have no idea what’s really going on and why.
0swestrup
If we assume I'm rational, then I'm not going to assume anything about Omega. I'll base my decisions on the given evidence. So far, that appears to be described as being no more and no less than what Omega cares to tell us.
0fractalman
Fine, then interchange "assume omega is honest" with, say, "i've played a billiion rounds of one-box two-box with him" ...It should be close enough.
-2bogdanb
I realize this is fighting the problem, but: If I remember playing a billion rounds of the game with Omega, that is pretty strong evidence that I’m a (slightly altered) simulation. An average human takes about a ten million breaths each year... OK, so assume that I’m a transhuman and can actually do something a billion times. But if Omega can simulate me perfectly, why would it actually waste the time to ask you a question, once it simulated you answering it? Let alone do that a billion times... This also seems like evidence that I’m actually simulated. (I notice that in most statements of the problem, the wording is such that it is implied but not clearly stated that the non-simulated version of you is ever involved.)

I work on AI. In particular, on decision systems stable under self-modification. Any agent who does not give the $100 in situations like this will self-modify to give $100 in situations like this. I don't spend a whole lot of time thinking about decision theories that are unstable under reflection. QED.

1thomblake
Even considering situations like this and having special cases for them sounds like it would add a bit much cruft to the system. Do you have a working AI that I could look at to see how this would work?

If you need special cases, your decision theory is not consistent under reflection. In other words, it should simply always do the thing that it would precommit to doing, because, as MBlume put it, the decision theory is formulated in such fashion that "What would you precommit to?" and "What will you do?" work out to be one and the same question.

1pjeby
But this is precisely what humans don't do, because we respond to a "near" situation differently than a "far" one. Your advance prediction of your decision is untrustworthy unless you can successfully simulate the real future environment in your mind with sufficient sensory detail to invoke "near" reasoning. Otherwise, you will fail to reach a consistent decision in the actual situation. Unless of course, In the actual situation, you're projecting back, "What would I have decided in advance to do had I thought about this in advance?" -- and you successfully mitigate all priming effects and situationally-motivated reasoning. Or to put all of the above in short, common-wisdom form: "that's easy for you to say NOW..." ;-)
6[anonymous]
Here is one intuitive way of looking at it: Before tossing the coin, the Omega perfectly emulates my decision making process. In this emulation he tells me that I lost the coin toss, explains the deal and asks me to give him $100. If this emulated me gives up the $100 then he has a good chance of getting $10,000. I have absolutely no way of knowing whether I am the 'emulated me' or the real me. Vladmir's specification is quite unambiguous. I, me, the one doing the deciding right now in this real world, am the same me as the one inside the Omega's head. If the emulation is in any way different to me then the Omega isn't the Omega. The guy in the Omega's head has been offered a deal that any rational man would accept, and I am that man. So, it may sound stupid that I'm giving up $100 with no hope of getting anything back. But that's because the counterfactual is stupid, not me.
5Nebu
(Disclaimer: I'm going to use the exact language you used, which means I will call you "stupid" in this post. I apologize if this comes off as trollish. I will admit that I am also quite torn about this decision, and I feel quite stupid too.) No offense, but assuming free will, you are the one who is deciding to actually hand over the $100. The conterfactual isn't the one making the decision. You are. You are in a situation, and there are two possible actions (lose $100 or don't lose $100), and you are choosing to lose $100. So again, are you sure you are not stupid?
1[anonymous]
And now I try to calculate what you should treat as being the probability that you're being emulated. Assume that Omega only emulates you if the coin comes up heads. Suppose you decide beforehand that you are going to give Omega the $100, as you ought to. The expected value of this is $4950, as has been calculated. Suppose that instead, you decide beforehand that E is the probability you're being emulated assuming you hear that came up tails. You'll still decide to give Omega the $100; therefore, your expected value if you hear that it came up heads is $10,000. Your expected value if you hear that the coin came up tails is -$100(1-E) + $10,000E. The probability that you hear that the coin comes up tails should be given by P(H) + P(T and ~E) + P(T and E) = 0, P(H) = P(T and ~E), P(T and ~E) = P(T) - P(T and E), P(T and E) = P(E|T) * P(T). Solving these equations, I get P(E|T) = 2, which probably means I've made a mistake somewhere. If not, c'est l'Omega?
-5fractalman
8MichaelVassar
So from my and Omega's perspective this coin is random and my behavior is predictable. Amusing. My question: What if Omega says "due to quirks in your neurology, had I requested it, you would have pre-committed to bet $100 against $46.32. As it happens, you lost anyway, but you would have taken an unfavorable deal. Would you pay then?

Nope. I don't care what quirks in my neurology do - I don't care what answer the material calculator returns, only the answer to 2 + 2 = ?

-5fractalman
6Vladimir_Nesov
The coin toss may be known to Omega and predicted in advance, it only needs to initially have 50/50 odds to you for the expected gain calculation to hold. When Omega tells you about the coin, it communicates to you its knowledge about the toss, about an independent variable of initial 50/50 odds. For example, Omega may tell you that it hasn't tossed the coin yet, it'll do so only a thousand years from now, but it predicted that the coin will come up tails, so it asks you for your $100.

This requires though that Omega have decided to make the bet in a fashion which exhibited no dependency on its advance knowledge of the coin.

0Nebu
This is a big issue which I unsucessfully tried to address in my non-existing 6+ paragraph explanation. Why the heck is Omega making bets if he can already predict everything anyway? That said, it's not clear that when Omega offers you a bet, you should automatically refuse it under the assumption that Omega is trying to "beat" you. It seems like Omega doesn't really mind giving away money (pretty reasonable for an omniscient entity), since he seems to be willing to leave boxes with millions of dollars in them just lying around. What is Omega's purpose is entirely unknown. Maybe he wants you to win these bets. If you're a rational person who "wants to win", I think you can just "not worry" about what Omega's intents are, and figure out what sequence of actions maximizes your utility (which in these examples always seems to directly translate into maximizing the amount of money you get).
0fractalman
Quantum Coins. seriously. they're easy enough to predict if you accept many worlds. as for the rest... entertainment? Could be a case of "even though I can predict these humans so well, it's fascinating as to just how many of them two-box no matter how obvious i make it." It's not impossible-we know that we exist, it is not impossible that some race resembling our own figured out a sufficient solution to the lob problem and became a race of omegas...
3jimmy
That's just like playing "Eeny, meeny, miny, moe" to determine who's 'it'. Once you figure out if there's an even or odd number of words, you know the answer, and it isn't random to you anymore. This may be great as a kid choosing who gets a cookie (wow! I win again!), but you're no longer talking about something that can go either way. For a random output of a known function, you still need a random input.
0fractalman
The trick with eeny-meeny-miney-moe is that it's long enough for us to not consciously and quickly identify whether the saying is odd or even, gives a 0, 1, or 2 on modulo 3, etc, unless we TRY to remember what it produces, or TRY to remember if it's odd or even before pointing it out. Knowing that doing so consciously ruins its capacity, we can turn to memory decay to restore some of the pseudo-random quality. basically, by sufficiently decoupling "point at A" from "choose A" to our internal cognitive algorithms...we change the way we route visual input and spit out a "point at X". THAT"S where the randomness of eeny-meeny-miney-moe comes in...though I've probably got only one use left of it when it comes to situations with 2 items thanks to writing this up...
-7fractalman
[-]Omega140

Hi,

My name is Omega. You may have heard of me.

Anyway, I have just tossed a fair coin, and given that the coin came up tails, I'm gonna have to ask each of you to give me $100. Whatever you do in this situation, nothing else will happen differently in reality as a result. Naturally you don't want to give up your $100. But see, if the coin came up heads instead of tails, I'd have given you each $10000, but only to those that would agree to give me $100 if the coin came up tails.

You forgot to add that we have sufficient reason to believe everything you say.

I don't believe you.

[-]MBlume130

You know, if Omega is truly doing a full simulation of my cognitive algorithm, then it seems my interactions with him should be dominated by my desire for him to stop it, since he is effectively creating and murdering copies of me.

The decision doesn't need to be read off from a straightforward simulation, it can be an on-demand, so to say, reconstruction of the outcome from the counterfactual. I believe it should be possible to calculate just your decision, without constructing a morally significant computation. Knowing your decision may be as simple as checking whether you adhere a certain decision theory.

0[anonymous]
There is no rule that says I need to care what the Omega does in his own head. If you object to being temporarily emulated then I can certainly see why you would be adverse to that. But I don't happen to object and nor do I feel in any way oblidged to. Even if I'm the emulated me.
0MBlume
but...if you're the emulated you...you're going to die after you give him/don't give him the money.
5[anonymous]
I can see why that bothers you and it is perfectly reasonable to include a penalty for being temporarily emulated in your utility function. The way I see it is that I started off as one me, had an extra emulated me for a bit and then was back to me again. There is just as much utility left over as there was at the start, with a bit extra thrown in for the favourable dice game. I have lost nothing. The emulated me gains nothing to gain be caring whether he is the real me or not. If he had the option of breaking free and being Cam II then by all means he would do so. But as it happens emulated me exists for a finite term and I have no way of exploiting that through my decision making. I just make whatever decision benefits the real me, whether that be this me or the other me. This is the way I see things: I am an AI running on a installation of Pen Drive Linux, an operating system that runs off a USB stick. Omega tells me that he has purchesed 10,000 identical USB sticks and duplicated me onto each of them. He tells me that in 1 hour he is going to destroy 10,000 of the USB sticks but double the processing power and RAM on the computer that runs me. He plans to repeat the same procedure every day. I have one hour. Do I use that time to exploit my net connection, appropriate funds and hire an assassain to kill Omega so he ceases his evil mass murder? As it happens, I don't. It bothers me not at all which of the 10,001 clones I am. There's still going to be a me that is more or less the same as me. If it happens that the copy of me currently running is destroyed I am quite willing to accept that. I don't consider myself 'killed' or 'dead'. I consider that I lost the memory of one conversation with some crazy Omega but gained a bunch of processing power and ram. Whatever. Go ahead, keep at it big O. In summary: I just don't think my instinctive aversion do death applies reasonably to situations where clones of me are being created and destroyed all willy nilly. In
3MBlume
It's not just about the USB sticks -- to me that seems inert. But if he's running you off those USB sticks for (let's say) a few hours every day, then you could (in fact there is a 1000/1001 chance that you will) wake up tomorrow morning and find yourself running from one of those drives, and know that there is a clear horizon of a few hours on the subjective experiences you can anticipate. This is a prospect which I, at least, would find terrifying.
-2[anonymous]
Maybe Omega exists in a higher spatial dimension and just takes an instantaneous snapshot of the universal finite state automata you exist in (as a p-zombie).

I guess I'm a bit tired of "God was unable to make the show today so the part of Omniscient being will be played by Omega" puzzles, even if in my mind Omega looks amusingly like the Flying Spaghetti Monster.

Particularly in this case where Omega is being explicitly dishonest - Omega is claiming to be either be sufficiently omniscient to predict my actions, or insufficiently omniscient to predict the result of a 'fair' coin, except that the 'fair' coin is explicitly predetermined to always give the same result . . . except . . .

What's the point of using rationalism to think things through logically if you keep placing yourself into illogical philosophical worlds to test the logic?

4Jonii
Coin is not predetermined, and it doesn't matter if Omega has hand-selected every result of the coin toss, as long as we don't have any reason to slide the probability of the result to either direction.
-1[anonymous]
Could be a quantum coin, which is unpredictable under current laws of physics. Anyway, this stuff actually does have applications in decision theory. Quibbling over the practical implementations of the thought experiment is not actually useful to you or anybody else.
0wedrifid
More precisely it is exactly predictable but for most practical purposes can be treated as equivalent to an unpredictable coin.
-2[anonymous]
By 'unpredictable' I mean 'under current formalisms of physics it is not possible for us to accumulate enough information to predict it'.
0wedrifid
By 'more precisely' I mean... no. The way you have phrased it makes your statement false. You can predict what the future outcome of a quantum coin will be (along the lines of branches with heads and tails their respective amplitudes). A related prediction you cannot make - when the quantum event has already occurred but you have not yet observed it you cannot predict what your observation will be (now that 'your' refers only to the 'you' in the specific branch). Again, for practical purposes - for most people's way of valuing most future outcomes - the future coin can be treated as though it is an unpredictable coin.
0[anonymous]
I was using 'you' and 'us' in the colloquial sense of the subjective experiences of a specific, arbitrary continuity chosen at random from of the set of Everette branches in the hypothetical branch of the world tree that this counterfactual occurs in. Now, I CAN start listing my precise definitions for every potentially ambiguous term I use, or we could simply agree not to pick improbable and inconsistent interpretations of the other's words. Frankly, I'd much prefer the latter, as I cannot abide pedants. EDIT: Or you could downvote all my posts. That's cool too.
1wedrifid
Since the distinction is of decision theoretical relevance and the source of much confusion I choose to clarify incorrect usages of 'unpredictable' in this particular environment. By phrasing it as 'more precisely' I leave plenty of scope for the original speaker to be assumed to be just speaking loosely. Unfortunately you chose to fortify and defend an incorrect position instead of allowing the additional detail. Now you have given a very nice definition of 'you' but even with that definition both of your claims are just as incorrect as when they started. Fixing 'you' misses the point. You are probably too entrenched in your position to work with but for anyone else who wants to talk about 'unpredictable' quantum coins, qualifiers like ("for most intents and purposes", "effectively") are awesome!
-2Luke_A_Somers
By reading the quantum coin flip, you definitely entangle yourself with it, and there's no way you're going to stay coherent. As a hard-core Everettian, I find the original usage and the followup totally unobjectionable in principle. Your clarification was good except for the part where it said Ati's statement was wrong. There exists a reading of the terms which leaves those wrong, yes. So don't use that one.
-3wedrifid
It should be noted that 'all my posts' does not refer to karma-assassaination here. Rather, that three comments here were downvoted. This is correct (and in accord with my downvoting policy).
1[anonymous]
And I perceived you as being needlessly pedantic and choosing implausible interpretations of my words so that you could correct me. You'll note that your comment karma stands. I am, in fact, aware of quantum mechanics, and you are, of course, entirely correct. Coins behave in precisely deterministic ways, even if they rely on, say, radioactive decay. The causality just occurs in many Everette branches. That said, there is no way that before 'you' 'flip the coin' you can make a prediction about its subjective future state, and have more than half of your future selves be right. If that's not 'unpredictable' by the word's colloquial definition, then I'm not sure the word has any meaning. You will notice that when I said that the coin is unpredictable, I did not claim, or even imply that the world was undeterministic, or that quantum mechanics was wrong. If I had said such a thing, you would have right to correct me. As it is, you took the opportunity to jump on my phrasing to correct me of a misconception that I did not, in fact, possess. That is being pedantic, it is pointless, and above all it is annoying. I apologize for rudeness, but trying to catch up others on their phrasing is a shocking waste of intellect and time. EDIT: Again, I can totally discard every word that's entrenched in good, old-fashioned single-universe connotations, and spell out all the fascinating multiverse implications of everything I say, if that will make you happy -- but it will make my posts about five times longer, and it will make it a good deal more difficult to figure out what the hell I'm saying, which rather defeats the purpose of using language.
0wedrifid
I'll note that I reject your 'implausible' claim, object to all insinuations regarding motive, stand by my previous statements and will maintain my policy of making mild clarifications when the subject happens to come up. There seems to be little else to be said here.
-3[anonymous]
As you like. Though I do hope you apply your strident policy of technical correctness in your home life, for consistency's sake. For example: someone (clearly wrong) like I would merely say, in our archaic and hoplessly monocosmological phrasing 'I am going to lunch.' This is clearly nonsense. You will, over the set of multiverse branches, do a great many things, many of them having nothing to do with food, or survival. The concept of 'I' and 'lunch' are not even particularly well defined. In contrast, someone held to your standard of correctness would have to say 'The computation function implemented in the cluster of mass from which these encoded pressure waves are emanating will execute a series of action for which they predict that in the majority of future Everette branches of this fork of the world tree, the aforementioned cluster of mass will accumulate new amplitude and potential energy through the process of digestion within the next hour and fifteen minutes.' Clearly this is more efficient and less confusing to the reader.
1wedrifid
I consider the selection of analogies made in the parent to constitute a misrepresentation (and fundamental misunderstanding) of the preceding conversation.
[-]bill80

I convinced myself to one-box in Newcomb by simply treating it as if the contents of the boxes magically change when I made my decision. Simply draw the decision tree and maximize u-value.

I convinced myself to cooperate in the Prisoner's Dilemma by treating it as if whatever decision I made the other person would magically make too. Simply draw the decision tree and maximize u-value.

It seems that Omega is different because I actually have the information, where in the others I don't.

For example, In Newcomb, if we could see the contents of both boxes, then... (read more)

8Vladimir_Nesov
Yes, the objective in designing this puzzle was to construct an example where according to my understanding of the correct way to make decision, the correct decision looks like losing. In other cases you may say that you close your eyes, pretend that your decision determines the past or other agents' actions, and just make the decision that gives the best outcome. In this case, you choose the worst outcome. The argument is that on reflection it still looks like the best outcome, and you are given an opportunity to think about what's the correct perspective from which it's the best outcome. It binds the state of reality to your subjective perspective, where in many other thought experiments you may dispense with this connection and focus solely on the reality, without paying any special attention to the decision-maker.
2bill
In Newcomb, before knowing the box contents, you should one-box. If you know the contents, you should two-box (or am I wrong?) In Prisoner, before knowing the opponent's choice, you should cooperate. After knowing the opponent's choice, you should defect (or am I wrong?). If I'm right in the above two cases, doesn't Omega look more like the "after knowing" situations above? If so, then I must be wrong about the above two cases... I want to be someone who in situation Y does X, but when Y&Z happens, I don't necessarily want to do X. Here, Z is the extra information that I lost (in Omega), the opponent has chosen (in Prisoner) or that both boxes have money in them (in Newcomb). What am I missing?
1Larks
No - in the prisoners' dilemma, you should always defect (presuming the payoff matrix represents utility), unless you can somehow collectively pre-commit to co-operating, or it is iterative. This distinction you're thinking of only applies when reverse causation comes into play.
[-]taw80

I really fail to see why you're all so fascinated by Newcomb-like problems. When you break causality, all logic based on causality doesn't function any more. If you try to model it mathematically, you will get inconsistent model always.