Speaking of problems I don't know how to solve, here's one that's been gnawing at me for years.

The operation of splitting a subjective worldline seems obvious enough - the skeptical initiate can consider the Ebborians, creatures whose brains come in flat sheets and who can symmetrically divide down their thickness.  The more sophisticated need merely consider a sentient computer program: stop, copy, paste, start, and what was one person has now continued on in two places.  If one of your future selves will see red, and one of your future selves will see green, then (it seems) you should anticipate seeing red or green when you wake up with 50% probability.  That is, it's a known fact that different versions of you will see red, or alternatively green, and you should weight the two anticipated possibilities equally.  (Consider what happens when you're flipping a quantum coin: half your measure will continue into either branch, and subjective probability will follow quantum measure for unknown reasons.)

But if I make two copies of the same computer program, is there twice as much experience, or only the same experience?  Does someone who runs redundantly on three processors, get three times as much weight as someone who runs on one processor?

Let's suppose that three copies get three times as much experience.  (If not, then, in a Big universe, large enough that at least one copy of anything exists somewhere, you run into the Boltzmann Brain problem.)

Just as computer programs or brains can split, they ought to be able to merge.  If we imagine a version of the Ebborian species that computes digitally, so that the brains remain synchronized so long as they go on getting the same sensory inputs, then we ought to be able to put two brains back together along the thickness, after dividing them.  In the case of computer programs, we should be able to perform an operation where we compare each two bits in the program, and if they are the same, copy them, and if they are different, delete the whole program.  (This seems to establish an equal causal dependency of the final program on the two original programs that went into it.  E.g., if you test the causal dependency via counterfactuals, then disturbing any bit of the two originals, results in the final program being completely different (namely deleted).)

So here's a simple algorithm for winning the lottery:

Buy a ticket.  Suspend your computer program just before the lottery drawing - which should of course be a quantum lottery, so that every ticket wins somewhere.  Program your computational environment to, if you win, make a trillion copies of yourself, and wake them up for ten seconds, long enough to experience winning the lottery.  Then suspend the programs, merge them again, and start the result.  If you don't win the lottery, then just wake up automatically.

The odds of winning the lottery are ordinarily a billion to one.  But now the branch in which you win has your "measure", your "amount of experience", temporarily multiplied by a trillion.  So with the brief expenditure of a little extra computing power, you can subjectively win the lottery - be reasonably sure that when next you open your eyes, you will see a computer screen flashing "You won!"  As for what happens ten seconds after that, you have no way of knowing how many processors you run on, so you shouldn't feel a thing.

Now you could just bite this bullet.  You could say, "Sounds to me like it should work fine."  You could say, "There's no reason why you shouldn't be able to exert anthropic psychic powers."  You could say, "I have no problem with the idea that no one else could see you exerting your anthropic psychic powers, and I have no problem with the idea that different people can send different portions of their subjective futures into different realities."

I find myself somewhat reluctant to bite that bullet, personally.

Nick Bostrom, when I proposed this problem to him, offered that you should anticipate winning the lottery after five seconds, but anticipate losing the lottery after fifteen seconds.

To bite this bullet, you have to throw away the idea that your joint subjective probabilities are the product of your conditional subjective probabilities.  If you win the lottery, the subjective probability of having still won the lottery, ten seconds later, is ~1.  And if you lose the lottery, the subjective probability of having lost the lottery, ten seconds later, is ~1.  But we don't have p("experience win after 15s") = p("experience win after 15s"|"experience win after 5s")*p("experience win after 5s") + p("experience win after 15s"|"experience not-win after 5s")*p("experience not-win after 5s").

I'm reluctant to bite that bullet too.

And the third horn of the trilemma is to reject the idea of the personal future - that there's any meaningful sense in which I can anticipate waking up as myself tomorrow, rather than Britney Spears.  Or, for that matter, that there's any meaningful sense in which I can anticipate being myself in five seconds, rather than Britney Spears.  In five seconds there will be an Eliezer Yudkowsky, and there will be a Britney Spears, but it is meaningless to speak of the current Eliezer "continuing on" as Eliezer+5 rather than Britney+5; these are simply three different people we are talking about.

There are no threads connecting subjective experiences.  There are simply different subjective experiences.  Even if some subjective experiences are highly similar to, and causally computed from, other subjective experiences, they are not connected.

I still have trouble biting that bullet for some reason.  Maybe I'm naive, I know, but there's a sense in which I just can't seem to let go of the question, "What will I see happen next?"  I strive for altruism, but I'm not sure I can believe that subjective selfishness - caring about your own future experiences - is an incoherent utility function; that we are forced to be Buddhists who dare not cheat a neighbor, not because we are kind, but because we anticipate experiencing their consequences just as much as we anticipate experiencing our own.  I don't think that, if I were really selfish, I could jump off a cliff knowing smugly that a different person would experience the consequence of hitting the ground.

Bound to my naive intuitions that can be explained away by obvious evolutionary instincts, you say?  It's plausible that I could be forced down this path, but I don't feel forced down it quite yet.  It would feel like a fake reduction.  I have rather the sense that my confusion here is tied up with my confusion over what sort of physical configurations, or cascades of cause and effect, "exist" in any sense and "experience" anything in any sense, and flatly denying the existence of subjective continuity would not make me feel any less confused about that.

The fourth horn of the trilemma (as 'twere) would be denying that two copies of the same computation had any more "weight of experience" than one; but in addition to the Boltzmann Brain problem in large universes, you might develop similar anthropic psychic powers if you could split a trillion times, have each computation view a slightly different scene in some small detail, forget that detail, and converge the computations so they could be reunified afterward - then you were temporarily a trillion different people who all happened to develop into the same future self.  So it's not clear that the fourth horn actually changes anything, which is why I call it a trilemma.

I should mention, in this connection, a truly remarkable observation: quantum measure seems to behave in a way that would avoid this trilemma completely, if you tried the analogue using quantum branching within a large coherent superposition (e.g. a quantum computer).  If you quantum-split into a trillion copies, those trillion copies would have the same total quantum measure after being merged or converged.

It's a remarkable fact that the one sort of branching we do have extensive actual experience with - though we don't know why it behaves the way it does - seems to behave in a very strange way that is exactly right to avoid anthropic superpowers and goes on obeying the standard axioms for conditional probability.

In quantum copying and merging, every "branch" operation preserves the total measure of the original branch, and every "merge" operation (which you could theoretically do in large coherent superpositions) likewise preserves the total measure of the incoming branches.

Great for QM.  But it's not clear to me at all how to set up an analogous set of rules for making copies of sentient beings, in which the total number of processors can go up or down and you can transfer processors from one set of minds to another.

To sum up:

  • The first horn of the anthropic trilemma is to confess that there are simple algorithms whereby you can, indetectably to anyone but yourself, exert the subjective equivalent of psychic powers - use a temporary expenditure of computing power to permanently send your subjective future into particular branches of reality.
  • The second horn of the anthropic trilemma is to deny that subjective joint probabilities behave like probabilities - you can coherently anticipate winning the lottery after five seconds, anticipate the experience of having lost the lottery after fifteen seconds, and anticipate that once you experience winning the lottery you will experience having still won it ten seconds later.
  • The third horn of the anthropic trilemma is to deny that there is any meaningful sense whatsoever in which you can anticipate being yourself in five seconds, rather than Britney Spears; to deny that selfishness is coherently possible; to assert that you can hurl yourself off a cliff without fear, because whoever hits the ground will be another person not particularly connected to you by any such ridiculous thing as a "thread of subjective experience".
  • The fourth horn of the anthropic trilemma is to deny that increasing the number of physical copies increases the weight of an experience, which leads into Boltzmann brain problems, and may not help much (because alternatively designed brains may be able to diverge and then converge as different experiences have their details forgotten).
  • The fifth horn of the anthropic trilemma is to observe that the only form of splitting we have accumulated experience with, the mysterious Born probabilities of quantum mechanics, would seem to avoid the trilemma; but it's not clear how to have analogous rules could possibly govern information flows in computer processors.

I will be extremely impressed if Less Wrong solves this one.

New Comment
232 comments, sorted by Click to highlight new comments since: Today at 3:20 AM
Some comments are truncated due to high volume. (⌘F to expand all)Change truncation settings

I've been thinking about this topic, off and on, at least since September 1997, when I joined the Extropians mailing list, and sent off a "copying related probability question" (which is still in my "sent" folder but apparently no longer archived anywhere that Google can find). Both Eliezer and Nick were also participants in that discussion. What are the chances that we're still trying to figure this out 12 years later?

My current position, for what it's worth, is that anticipation and continuity of experience are both evolutionary adaptations that will turn maladaptive when mind copying/merging becomes possible. Theoretically, evolution could have programmed us to use UDT, in which case this dilemma wouldn't exist now, because anticipation and continuity of experience is not part of UDT.

So why don't we just switch over to UDT, and consider the problem solved (assuming this kind of self-modification is feasible)? The problem with that is that much of our preferences are specified in terms of anticipation of experience, and there is no obvious way how to map those onto UDT preferences. For example, suppose you’re about to be tortured in an hour. Should you make as... (read more)

12y later

I just happened to see this whilst it happens to be 12 years later. I wonder what your sense of this puzzle is now (object-level as well as meta-level).

I'm not really aware of any significant progress since 12 years ago. I've mostly given up working on this problem, or most object-level philosophical problems, due to slow pace of progress and perceived opportunity costs. (Spending time on ensuring a future where progress on such problems can continue to be made, e.g., fighting against x-risk and value/philosophical lock-in or drift, seems a better bet even for the part of me that really wants to solve philosophical problems.) It seems like there's a decline in other LWer's interest in the problem, maybe for similar reasons?

Seeing that others here are trying to figure out how to make probabilities of anticipated subjective experiences work, I should perhaps mention that I spent quite a bit of time near the beginning of those 12 years trying to do the same thing. As you can see, I eventually gave up and decided that such probabilities shouldn't play a role in a decision theory for agents who can copy and merge themselves.

This isn't to discourage others from exploring this approach. There could easily be something that I overlooked, that a fresh pair of eyes can find. Or maybe someone can give a conclusive argument that explains why it can't work.

BTW, notice that UDT not only doesn't involve anticipatory probabilities, it doesn't even involve indexical probabilities (i.e. answers to "where am I likely to be, given my memories and observations?" as opposed to "what should I expect to see later?"). It seems fairly obvious that if you don't have indexical probabilities, then you can't have anticipatory probabilities. (See ETA below.) I tried to give an argument against indexical probabilities, which apparently nobody (except maybe Nesov) liked. Can anyone do better?

ETA: In the Absent-M... (read more)

My thread of subjective experience is a fundamental part of how I feel from the inside. Exchanging it for something else would be pretty much equivalent to death - death in the human, subjective sense. I would not wish to exchange it unless the alternative was torture for a googol years or something of that ilk.

Why would you wish to switch to UDT?

6Wei Dai14y
That's a good point. I probably wouldn't want to give up my thread of subjective experience either. But unless I switch (or someone comes up with a better solution than UDT), when mind copying/merging becomes possible I'm probably going to start making some crazy decisions. I'm not sure what the solution is, but here's one idea. An UDT agent doesn't use anticipation or continuity of experience to make decisions, but perhaps it can run some computations on the side to generate the qualia of anticipation and continuity. Another idea, which may be more intuitively acceptable, is don't make the switch yourself. Create a copy, and have the copy switch to UDT (before it starts running). Then give most of your resources to the copy and live a single-threaded life under its protection. (I guess the copy in this case isn't so much a copy but more of a personal FAI.)
1DanArmak14y
That's what I was thinking, too. You make tools the best way you can. The distinction between tools that are or aren't part of you will ultimately become meaningless anyway. We're going to populate the galaxy with huge Jupiter brains that are incredibly smart and powerful but whose only supergoal is to protect a tiny human-nugget inside.
9Eliezer Yudkowsky14y
Because we can't interpret UDT's decision algorithm as providing epistemic advice. It says to never update our priors and even to go on putting weight on logical impossibilities after they're known to be impossible. UDT tells us what to do - but not what to anticipate seeing happen next.

This presumably places anticipation together with excitement and fear -- an aspect of human experience, but not a useful concept for decision theory.

4Eliezer Yudkowsky14y
I'm not convinced that "It turns out that pi is in fact greater than three" is a mere aspect of human experience.
0Vladimir_Nesov14y
If you appeal to intuitions about rigor, it's not so much an outlier since fear and excitement must be aspects of rigorously reconstructed preference as well.
3UnholySmoke14y
I find myself simultaneously convinced and unconvinced by this! Anticipation (dependent, of course, on your definition) is surely a vital tool in any agent that wants to steer the future? Or do you mean 'human anticipation' as differentiated from other kinds? In which case, what demarcates that from whatever an AI would do in thinking about the future? However, Dai, your top level comment sums up my eventual thoughts on this problem very well. I've been trying for a long time to resign myself to the idea that a notion of discrete personal experience is incompatible with what we know about the world. Doesn't make it any easier though. My two cents - the answer to this trilemma will come from thinking about the system as a whole rather than personal experience. Can we taboo 'personal experience' and find a less anthropocentric way to think about this?
3Wei Dai14y
Ok, we can count that as a disadvantage when comparing UDT with alternative solutions, but why is it a deal-killer for you, especially since you're mainly interested in decision theory as a tool for programming FAI? As long as the FAI knows what to do, why do you care so much that it doesn't anticipate seeing what happen next?
6Eliezer Yudkowsky14y
Because I care about what I see next. Therefore the FAI has to care about what I see next - or whatever it is that I should be caring about.
2Wei Dai14y
Ok, but that appears to be the same reason that I gave (right after I asked the question) for why we can't switch over to UDT yet. So why did you give a another answer without reference to mine? That seems to be needlessly confusing. Here's how I put it: There's more in that comment where I explored one possible approach to this problem. Do you have any thoughts on that? Also, do you agree (or think it's a possibility) that specifying preferences in terms of anticipation (instead of, say, world histories) was an evolutionary "mistake", because evolution couldn't anticipate that one day there would be mind copying/merging technology? If so, that doesn't necessarily mean we should discard such preferences, but I think it does mean that there is no need to treat it as somehow more fundamental than other kinds of preferences, such as, for example, the fear of stepping into a teleporter that uses destructive scanning, or the desire not to be consigned to a tiny portion of Reality due to "mistaken" preferences.
2Eliezer Yudkowsky14y
I can't switch over to UDT because it doesn't tell me what I'll see next, except to the extent it tells me to expect to see pi < 3 with some measure. It's not that it doesn't map. It's that UDT goes on assigning measure to 2 + 2 = 5, but I'll never see that happen. UDT is not what I want to map my preferences onto, it's not a difficulty of mapping.
1Wei Dai14y
That's not what happens in my conception of UDT. Maybe in Nesov's, but he hasn't gotten it worked out, and I'm not sure it's really going to work. My current position on this is still that you should update on your own internal computations, but not on input from the outside. ETA: Is that the same point that Dan Armak made, which I responded to, or a different one?
0Vladimir_Nesov14y
It's not you who should use UDT, it's the world. This is a salient point of departure between FAI and humanity. FAI is not in the business of saying in words what you should expect. People are stuff of the world, not rules of the world or strategies to play by those rules. Rules and strategies don't depend on particular moves, they specify how to handle them, but plays consist of moves, of evidence. This very distinction between plays and strategies is the true origin of updatelessness. It is the fault to make this distinction that causes the confusion UDT resolves.
7Wei Dai14y
Nesov, your writings are so hard to understand sometimes. Let me take this as an example and give you some detailed feedback. I hope it's useful to you to determine in the future where you might have to explain in more detail or use more precise language. Do you mean "it's not only you", or "it's the world except you"? If it's the latter, it doesn't seem to make any sense. If it's the former, it doesn't seem to answer Eliezer's objection. Do you mean FAI should use UDT, and humanity shouldn't? Ok, this seems clear. (Although why not, if that would make me feel better?) By "stuff", do you mean "part of the state of the world"? And people do in some sense embody strategies (what they would do in different situations), so what do you mean by "people are not strategies"? This part makes sense, but I don't see the connection to what Eliezer wrote.
1Vladimir_Nesov14y
I mean the world as substrate, with "you" being implemented on the substrate of FAI. FAI runs UDT, you consist of FAI's decisions (even if in the sense of "influenced by", there seems to be no formal difference). The decisions are output of the strategy optimized for by UDT, two levels removed from running UDT themselves. Yes, in the sense that humanity runs on the FAI-substrate that uses UDT or something on the level of strategy-optimization anyway, but humanity itself is not about optimization. I suspect that people should be found in plays (what actually happens given the state of the world), not strategies (plans for every eventuality).
0Vladimir_Nesov14y
There is no problem with FAI looking at both past and future you -- intuition only breaks down when you speak of first-person anticipation. You don't care what FAI anticipates to see for itself and whether it does. The dynamic of past->future you should be good with respect to anticipation, just as it should be good with respect to excitement.
1Nick_Tarleton14y
But part of the question is: must past/future me be causally connected to me?
0Vladimir_Nesov14y
Part of which question? And whatever you call "causally connected" past/future persons is a property of the stuff-in-general that FAI puts into place in the right way.
0SilasBarta14y
Unless I'm misunderstanding UDT, isn't speed another issue? An FAI must know what's likely to be happening in the near future in order to prioritize its computational resources so they're handling the most likely problems. You wouldn't want it churning through the implications of the Loch Ness monster being real while a mega-asteroid is headed for the earth.
3Eliezer Yudkowsky14y
Wei Dai should not be worrying about matters of mere efficiency at this point. First we need to know what to compute via a fast approximation. (There are all sorts of exceptions to this principle, and they mostly have to do with "efficient" choices of representation that affect the underlying epistemology. You can view a Bayesian network as efficiently compressing a raw probability distribution, but it can also be seen as committing to an ontology that includes primitive causality.)
1SilasBarta14y
But that path is not viable here. If UDT claims to make decisions independently of any anticipation, then it seems it must be optimal on average over all the impossibilities it's prepared to compute an output for. That means it must be sacrificing optimality in this world-state (by No Free Lunch), even given infinite computing time, so having a quick approximation doesn't help. If an AI running UDT is just as prepared to find Nessie as to find out how to stop the incoming asteroid, it will be inferior to a program designed just to find out how to stop asteroids. Expand the Nessie possibility to improbable world-states, and the asteroid possibility to probable ones, and you see the problem. Though I freely admit I may be completely lost on this.
3MichaelHoward14y
Not small. I read that list and similar forums in the early 90s before becoming an AGI relinquishmentarian till about 2 years ago. When coming back to the discussions, I was astonished how most of the topics on discussion were essentially the same ones I remembered from 15 years earlier.
2Johnicholas14y
Note - there is a difference between investigating "what would evolution do?", as a jumping-off point for other strategies, and recommending "we should do what evolution does". Why is it that if I set up a little grid-world on my computer and evolve little agents, I seem to get answers to the question "what does evolution do"? Am I encoding "where to look" into the grid-world somehow?

Whatever the correct answer is, the first step towards it has to be to taboo words like "experience" in sentences like, "But if I make two copies of the same computer program, is there twice as much experience, or only the same experience?"

What making copies is, is creating multiple instances of the same pattern. If you make two copies of a pattern, there are twice as many instances but only one pattern, obviously.

Are there, then, two of 'you'? Depends what you mean by 'you'. Has the weight of experience increased? Depends what you mean by 'experience'. Think in terms of patterns and instances of patterns, and these questions become trivial.

I feel a bit strange having to explain this to Eliezer Yudkowsky, of all people.

8Eliezer Yudkowsky14y
Can I redefine what I mean by "me" and thereby expect that I will win the lottery? Can I anticipate seeing "You Win" when I open my eyes? It still seems to me that expectation exists at a level where I cannot control it quite so freely, even by modifying my utility function. Perhaps I am mistaken.

I think the conflict is resolved by backing up to the point where you say that multiple copies of yourself count as more subjective experience weight (and therefore a higher chance of experiencing).

But if I make two copies of the same computer program, is there twice as much experience, or only the same experience? Does someone who runs redundantly on three processors, get three times as much weight as someone who runs on one processor?

Let's suppose that three copies get three times as much experience. (If not, then, in a Big universe, large enough that at least one copy of anything exists somewhere, you run into the Boltzmann Brain problem.)

I have a top-level post partly written up where I attempt to reduce "subjective experience" and show why your reductio about the Boltzmann Brain doesn't follow, but here's a summary of my reasoning:

Subjective experience appears to require a few components: first, forming mutual information with its space/time environment. Second, forming M/I with its past states, though of course not perfectly.

Now, look at the third trilemma horn: Britney Spears's mind does not have M/I with your past memories. So it is flat-out incoherent to ... (read more)

7Eliezer Yudkowsky14y
I see! So you're identifying the number of copies with the number of causally distinct copies - distinct in the causality of a physical process. So copying on a computer does not produce distinct people, but spontaneous production in a distant galaxy does. Thus real people would outweigh Boltzmann brains. But what about causally distinct processes that split, see different tiny details, and then merge via forgetting? (Still, this idea does seem to me like progress! Like we could get a bit closer to the "magical rightness" of the Born rules this way.)

Actually, let me revise that: I made it more complicated than it needs to be. Unless I'm missing something (and this does seem too simple), you can easily resolve the dilemma this way:

Copying your upload self does multiply your identities but adds nothing to your anticipated probabilities that stem from quantum branching.

So here's what you should expect:

-There's still a 1 in a billion chance of experiencing winning the lottery.

-In the event you win the lottery, you will also experience being among a trillion copies of yourself, each of whom also have this experience. Note the critical point: since they all wake up in the same Everett branch, their subjective experience does not get counted in at the same "level" as the experience of the lottery loser.

-If you merge after winning the lottery you should expect, after the merge, to remember winning the lottery, and some random additional data that came from the different experiences the different copies had.

-This sums to: ~100% chance of losing the lottery, 1 in a billion chance of winning the lottery plus forgetting a few details.

-Regarding the implications of self-copying in general: Each copy (or original or instantiatio... (read more)

3UnholySmoke14y
Yeah I get into trouble there. It feels as though two identical copies of a person = 1 pattern = no more people than before copying. But flip one bit and do you suddenly have two people? Can't be right. That said, the reason we value each person is because of their individuality. The more different two minds, the closer they are to two separate people? Erk. Silas, looking forward to that post.
6gwern14y
Why not? Imagine that bit is the memory/knowledge of which copy they are. After the copying, each copy naturally is curious what happened, and recall that bit. Now, if you had 1 person appearing in 2 places, it should be that every thought would be identical, right? Yet one copy will think '1!'; the other will think '0!'. As 1 != 0, this is a contradiction. Not enough of a contradiction? Imagine further that the original had resolved to start thinking about hot sexy Playboy pinups if it was 1, but to think about all his childhood sins if 0. Or he decides quite arbitrarily to become a Sufi Muslim if 0, and a Mennonite if 1. Or... (insert arbitrarily complex mental processes contingent on that bit). At some point you will surely admit that we now have 2 people and not just 1; but the only justifiable step at which to say they are 2 and not 1 is the first difference.

At some point you will surely admit that we now have 2 people and not just 1

Actually I won't. While I grok your approach completely, I'd rather say my concept of 'an individual' breaks down once I have two minds with one bit's difference, or two identical minds, or any of these borderline cases we're so fond of.

Say I have two optimisers with one bit's difference. If that bit means one copy converts to Sufism and the other to Mennonism, then sure, two different people. If that one bit is swallowed up in later neural computations due to the coarse-grained-ness of the wetware, then we're back to one person since the two are, once again, functionally identical. Faced with contradictions like that, I'm expecting our idea of personal identity to go out the window pretty fast once tech like this actually arrives. Greg Egan's Diaspora pretty much nails this for me, have a look.

All your 'contradictions' go out the window once you let go of the idea of a mind as an indivisible unit. If our concept of identity is to have any value (and it really has to) then we need to learn to think more like reality, which doesn't care about things like 'one bit's difference'.

0gwern14y
Ack. So if I understand you right, your alternative to bit-for-bit identity is to loosen it to some sort of future similarity, which can depend on future actions and outcomes; or in other words, there's a radical indeterminacy about even the minds in our example: are they same or are they different, who knows, it depends on whether the Sufism comes out in the wash! Ask me later; but then again, even then I won't be sure whether those 2 were the same when we started them running (always in motion the future is). That seems like quite a bullet to bite, and I wonder whether you can hold to any meaningful 'individual', whether the difference be bit-wise or no. Even 2 distant non-borderline mindsmight grow into each other.
2UnholySmoke14y
Indeed, that's what I'm driving at. Harking back to my earlier comment, changing a single bit and suddenly having a whole new person is where my problem arises. If you change that bit back, are you back to one person? I might not be thinking hard enough, but my intuition doesn't accept that. With that in mind, I prefer to bite that bullet than talk about degrees of person-hood.
2gwern14y
Here's an intuition for you: you take the number 5 and add 1 to it; then you subtract 1 from it; don't you have what you started with? Well, I can't really argue with that. As long as you realize you're biting that bullet, I think we're still in a situation where it's just dueling intuitions. (Your intuition says one thing, mine another.)
1Psy-Kosh14y
The downside is that it's not really that reductionistic. What if you flip a bit in part of an offline memory store that you're not consciously thinking about at the time or such?
2gwern14y
What if I hack & remove $100 from your bank account. Are you just as wealthy as you were before, because you haven't looked? If the 2 copies simply haven't looked or otherwise are still unaware, that doesn't mean they are the same. Their possible futures diverge. And, sure, it's possible they might never realize - we could merge them back before they notice, just as I could restore the money before the next time you checked, but I think we would agree that I still committed a crime (theft) with your money; why couldn't we feel that there was a crime (murder) in the merging?
1Psy-Kosh14y
Huh? My point is a bitflip in a non conscious part, before it influences any of the conscious processing, well, if prior to that bit flip you would have said there was only one being, then I'd say after that they'd still not yet diverged. Or at least, not entirely. As far as a merging, well, in that case who, precisely, is the one that's being killed?
1gwern14y
So only anything in immediate consciousness counts? Fine, let's remove all of the long-term memories of one of the copies - after all, he's not thinking about his childhood... Obviously whichever one isn't there afterwards; if the bit is 1, then 0 got killed off & vice versa. If we erase both copies and replace with the original, then both were killed.
1Psy-Kosh14y
I'd have to say that IF two (equivalent) instances of a mind count as "one mind", then removing an unaccessed data store does not change that for the duration that the effect of the removal doesn't propagate directly or indirectly to the conscious bits. If one then restores that data store before anything was noticed regarding it being missing, then, conditional on the assumption that IF the two instances originally only counted as one being, then.... so they remain. EDIT: to clarify, though... my overall issue here is that I think we may be effectively implicitly treating conscious agents as irreducible entities. If we're ever going to find an actual proper reduction of consciousness, well, we probably need to ask ourselves stuff like "what if two agents are bit for bit identical... except for these couple of bits here? What if they were completely identical? Is the couple bit difference enough that they might as well be completely different?" etc...
0Eliezer Yudkowsky14y
And if we restore a different long-term memory instead?
1Psy-Kosh14y
I think I'd have to say still "Nothing of significance happened until memory access occurs" Until then, well, how's it any different then stealing your books... and then replacing them before you notice? Now, as I said, we probably ought be asking questions like "what if in the actual "conscious processing" part of the agent, a few bits were changed in one instance... but just that... so initially, before it propagates enough to completely diverge... what should we say? To say it completely changes everything instantly, well... that seems too much like saying "conscious agents are irreducible", so... (just to clarify: I'm more laying out a bit of my confusion here rather than anything else, plus noting that we seem to have been, in our quest to find reductions for aspects of consciousness, implicitly treating agents as irreducible in certain ways)
0gwern14y
Indeed. It's not obvious what we can reduce agents down further into without losing agents entirely; bit-for-bit identity is at least clear in a few situations. (To continue the example - if we see the unaccessed memory as being part of the agent, then obviously we can't mess with it without changing the agent; but if we intuitively see it as like the agent having Internet access and the memory being a webpage, then we wouldn't regard it as part of its identity.)
0UnholySmoke14y
Standard Dispute. If wealthy = same amount of money in the account, no. If wealthy = how rich you judge yourself to be. The fact that 'futures diverge' is irrelevant up until the moment those two different pieces of information have causal contact with the brain. Until that point, yes, they are 'the same
1SilasBarta14y
I don't know; I'm still working through the formalism and drawing causal networks. And I just realized I should probably re-assimilate all the material in your Timeless Identity post, to see the relationship between identity and subjective experience. My brain hurts. For now, let me just mention that I was trying to do something similar to what you did when identifying what d-connects the output of a calculator on Mars and Venus doing the same calculation. There's an (imperfect) analog to that, if you imagine a program "causing" its two copies, which each then get different input. They can still make inferences about each other despite being d-separated by knowledge of their pre-fork state. The next step is to see how this mutual information relates to the kind between one sentient program's subsequent states. And, for bonus points, make sure to eliminate time by using the thermodynamic arrow and watch the entropy gain from copying a program.
0Eliezer Yudkowsky14y
...okay, that part didn't make any particular sense to me.
0SilasBarta14y
Heh, maybe you just had read more insight into my other comment than there actually was. Let me try to rephrase the last: I'm starting from the perspective of viewing subjective experience as something that forms mutual information with its space/time surroundings, and with its past states (and has some other attributes I'll add later). This means that identifying which experience you will have in the future is a matter of finding which bodies have mutual information with which. M/I can be identified by spotting inferences in a Bayesian causal network. So what would a network look like that has a sentient program being copied? You'd show the initial program as being the parent of two identical programs. But, as sentient programs with subjective experience, they remember (most of) their state before the split. This knowledge has implications for what inferences one of them can make about the other, and therefore how much mutual information they will have, which in turn has implications for how their subjective experiences are linked. My final sentence was noting the importance of checking the thermodynamic constraints on the processes going on, and the related issue, of making time removable from the model. So, I suggested that instead of phrasing questions about "previous/future times", you should phrase such questions as being about "when the universe had lower/higher total entropy". This will have implications for what the sentience will regard as "its past". Furthermore, the entropy calculation is affected by copy (and merge) operations. Copying involves deleting to make room for the new copies, whereas merging throws away information if the copies aren't identical. Now, does that make it any clearer, or does it just make it look like you overestimated my first post?
3Nominull14y
Yes? Obviously? You can go around redefining anything as anything. You can redefine a ham sandwich as a steel I-beam and thereby expect that a ham sandwich can support hundreds of pounds of force. The problem is that in that case you lose the property of ham sandwiches that says they are delicious. In the case of redefining you as someone who wins the lottery, the property you are likely to lose is the property of generating warm fuzzy feelings of identification inside Eliezer Yudkowsky.
7Alicorn14y
"If you call a tail a leg, how many legs does a dog have...? Four. Calling a tail a leg doesn't make it one."
2Furcas14y
That was said by someone who didn't realize that words are just labels.

Words are just labels, but in order to be able to converse at all, we have to hold at least most of them in one place while we play with the remainder. We should try to avoid emulating Humpty Dumpty. Someone who calls a tail a leg is either trying to add to the category originally described by "leg" (turning it into the category now identified with "extremity" or something like that), or is appropriating a word ("leg") for a category that already has a word ("tail"). The first exercise can be useful in some contexts, but typically these contexts start with somebody saying "Let's evaluate the content of the word "leg" and maybe revise it for consistency." The second is juvenile code invention.

2SilasBarta14y
What about if evolution repurposed some genus's tail to function as a leg? The question wouldn't be so juvenile or academic then. And before you roll your eyes, I can imagine someone saying, "How many limbs does a mammal have, if you count the nose as a limb? Four. Calling a nose a limb doesn't make it one." And then realizing they forgot about elephants, whose trunks have muscles that allow it to grip things as if it had a hand.
4Alicorn14y
That looks like category reevaluation, not code-making, to me. If you think an elephant's trunk should be called a limb, and you think that elephants have five limbs, that's category reevaluation; if you think that elephant trunks should be called limbs and elephants have one limb, that's code.
9RichardChappell14y
Speakers Use Their Actual Language, so someone who uses 'leg' to mean leg or tail speaks truly when they say 'dogs have five legs.' But it remains the case that dogs have only four legs, and nobody can reasonably expect a ham sandwich to support hundreds of pounds of force. This is because the previous sentence uses English, not the counterfactual language we've been invited to imagine.

To condense my response to a number of comments here:

It seems to me that there's some level on which, even if I say very firmly, "I now resolve to care only about future versions of myself who win the lottery! Only those people are defined as Eliezer Yudkowskys!", and plan only for futures where I win the lottery, then, come the next day, I wake up, look at the losing numbers, and say, "Damnit! What went wrong? I thought personal continuity was strictly subjective, and I could redefine it however I wanted!"

You reply, "But that's just because you're defining 'I' the old way in evaluating the anticipated results of the experiment."

And I reply, "...I still sorta think there's more to it than that."

To look at it another way, consider the Born probabilities. In this case, Nature seems to have very definite opinions about how much of yourself flows where, even though both copies exist. Now suppose you try to redefine your utility function so you only care about copies of yourself that see the quantum coin land heads up. Then you are trying to send all of your measure to the branch where the coin lands up heads, by exercising your right to red... (read more)

I'm sorry, I don't think I can help. It's not that I don't believe in personal continuity, it's that I can't even conceive of it.

At t=x there's an Eliezer pattern and there's a Bill Gates pattern. At t=x+1 there's an Eliezer+1 pattern and a Bill Gates+1 pattern. A few of the instances of those patterns live in worlds in which they won the lottery, but most don't. There's nothing more to it than that. How could there be?

Some Eliezer instances might have decided to only care about Eliezer+1 instances that won the lottery, but that wouldn't change anything. Why would it?

I can't be the only one who sees this discussion as parallel to the argument over free will, right down to the existence of people who proudly complain that they can't see the problem.

Do you see how this is the same as saying "Of course there's no such thing as free will; physical causality rules over the brain"? Not false, but missing completely that which actually needs to be explained: what it is that our brain does when we 'make a choice', and why we have a deeply ingrained aversion to the first question being answered by some kind of causality.

8Furcas14y
There's a strong similarity, all right. In both cases, the bullet-biters describe reality as we have every reason to believe it is, and ask the deniers how reality would be different if free will / personal continuity existed. The deniers don't have an answer, but they're very insistent about this feeling they have that this undefined free will or continuity thing exists. Explaining this feeling could be interesting, but it has very little to do with the question of whether what the feeling is about, is real.
5CronoDAS14y
Wouldn't that just mean that, there was someone who was very much like Eliezer Yudkowsky and who remembered being Eliezer Yudkowsky, but woke up and discovered they were no longer Eliezer Yudkowsky? /me suspects that he just wrote a very confused sentence It seems to me as though our experience of personal continuity has an awful lot to do with memory... I remember being me, so I'm still the same me. I think. It feels like there's a wrong question in here somewhere, but I don't know what it is!
5Eliezer Yudkowsky14y
Okay, let me try another tack. One of the last greatest open questions in quantum mechanics, and the only one that seems genuinely mysterious, is where the Born statistics come from - why our probability of seeing a particular result of a quantum experiment, ending up in a particular decoherent blob of the wavefunction, goes as the squared modulus of the complex amplitude. Is it the case that the Born probabilities are necessarily explained - can only be explained - by some hidden component of our brain which says that we care about the alternatives in proportion to their squared modulus? Since (after all) if we only cared about results that went a particular way, then, from our perspective, we would always anticipate seeing the results go that way? And so what we anticipate seeing, is entirely and only dependent on what we care about? Or is there a sense in which we end up seeing results with a certain probability, a certain measure of ourselves going into those worlds, regardless of what we care about? If you look at it closely, this is really about an instantaneous measure of the weight of experience, not about continuity between experiences. But why don't the same arguments on continuity work on measure in general?
8Christian_Szegedy14y
I have been thinking about quite a bit in the last few days and I have to say, I find this close to impossible. The solution must be much more fundamental: Assumptions like the above ignore that the Born rule is also necessary for almost everything to work: For example the working of our most basic building blocks are tied to this rule. It is much more then just our psychological "caring". Everything in our "hardware" and environment would immediately cease to exist if the rule would be different Based on this, I think that attempts (like that of David Wallace, even if it would be correct) trying to prove the Born rule based on rationality and decision theory have no chance to be conclusive or convincing. A good theory to explain the rule should also explain why we see the reality as we see it, even if never really make conscious measurements on particles. In our lives, we (may) see different type of apparent randomnesses: * incomplete information * inherent (quantum) randomness To some extent these two type of randomness are connected and look isomorphic on surface (in the macro-world). The real question is: "Why are they connected?" Or more specifically: "Why does the amplitude of the wave function result in (measured) probabilities that resembles to those of random geometric perturbations of the wave function?" If you flip a real coin, for you it does not look very different from flipping a quantum coin. However the 50/50 chance of heads and tails can be explained purely by considering the geometric symmetry of the object. If you assume that the random perturbing events are distributed in a geometrically uniform way, you will immediately deduce the necessity of even chance. I think the clue of the Born rule will be to relate similar geometric considerations to relate perturbation based probability to quantum probability.
5Vladimir_Nesov14y
Quantum probability is only "inherent" because by default you are looking at it from the system that only includes one world. With a coin, the probability is merely "epistemic" because there is a definite answer (heads or tails) in the system that includes one world, but this same probability is as inherent for the system that only includes you, the person who is uncertain, and doesn't include the coin. The difference between epistemic and inherent randomness is mainly in the choice of the system for which the statement is made, with epistemic probability meaning the same thing as inherent probability with respect to the system that doesn't include the fact in question. (Of course, this doesn't take into account the specifics of QM, but is right for the way "quantum randomness" is usually used in thought experiments.)
0Christian_Szegedy14y
I don't dispute this. Still, my posting implicitly assumed the MWI. My argument is that the brain as an information processing unit has a generic way of estimating probabilities based on a single-worldline of the Multiverse. This world both contains randomness stemming from missing information and quantum branching, but our brain does not differentiate between these two kind of randomnesses. The question is how to calibrate our brain's expectation of the quantum branch it will end up. What I speculate is that the quantum randomness to some extent approximates an "incomplete information" type of randomness on the large scale. I don't know the math (if I'd knew I'd be writing a paper :)), but I have a very specific intuitive idea, that could be turned into a concrete mathematical argument: I expect the calibration to be performed based on geometric symmetries of our 3 dimensional space: if we construct a sufficiently symmetric but unstable physical process (e.g. throwing a coin) than we can deduce a probability for the outcome to be 50/50 assuming a uniform geometric distribution of possible perturbations. Such a process must somehow be related to the magnitudes of wave function and has to be shown to behave similarly on the macro level. Admitted, this is just a speculation, but it is not really philosophical in nature, rather an intuitive starting point on what I think has a fair chance ending up in a concrete mathematical explanation of the Born probabilities in a formal setting.
1Eliezer Yudkowsky14y
Does your notion of "incomplete information" take into account Bell's Theorem? It seems pretty hard to make the Born probabilities represent some other form of uncertainty than indexical uncertainty.
1Christian_Szegedy14y
I don't suggest hidden variables. The idea is that quantum randomness should resemble incomplete information type of randomness on the large scale and the reason that we perceive the world according to the Born rule is that our brain can't distinguish between the two kind of randomnesses.
3Wei Dai14y
There are beings out there in other parts of Reality, who either anticipate seeing results with non-Born probabilities, or care about future alternatives in non-Born proportions. But (as I speculated earlier) those beings have much less measure under a complexity-based measure than us. In other words, what you're asking is, is there is an objective measure over Reality, or is it just a matter of how much we care about about each part of it. I've switched positions on this several times, and I'm still undecided now. But here are my current thoughts. First, considerations from algorithmic complexity suggest that the measure we use can't be completely arbitrary. For example, we certainly can't use one that takes an infinite amount of information to describe, since that wouldn't fit into our brain. Next, it doesn't seem to make sense to assign zero measure to any part of Reality. Why should there be a part of it that we don't care about at all? So that seems to narrow down the possibilities quite a bit, even if there is no objective measure. Maybe we can find other considerations to further narrow down the list of possibilities? I'd say that "continuity between experiences" is a separate problem. Even if the measure problem is solved, I might still be afraid to step into a transporter based on destructive scanning and reconstruction, and need to figure out whether I should edit that fear away, tell the FAI to avoid transporting me that way, or do something else. I don't understand this one. What "arguments on continuity" are you referring to?
2Johnicholas14y
QM has to add up to normality. We know it is a dumb idea to attempt (quantum) suicide. We're pretty confident it is a dumb idea to do simple algorithms increasing one's redundancy before pleasant realizations and reducing it afterward. It sounds as if you are refusing to draw inferences from normal experience regarding (the correct interpretation of) QM. There is no "Central Dogma" that inferences can only go from micro-scale to macro-scale. From the macro-scale values that we do hold (e.g. we care about macro-scale probable outcomes), we can derive the micro-scale values that we should hold (e.g. care about Born weights). I don't have an explanation why Born weights are nonlinear - but the science is almost completely irrelevant to the decision theory and the ethics. The mysterious, nonintuitive nature of QM doesn't percolate up that much. That is why we have different fields called "physics", "decision theory", and "ethics".
0Psy-Kosh14y
I read that part several times, and I'm still not quite following. Mind elaborating or rephrasing that bit? Thanks.
3Nominull14y
I would be careful about declaring future yous who don't win the lottery to not be you; if Many-Worlds happens not to be true, you've just committed a different sort of quantum suicide.
2Z_M_Davis14y
But all the resulting observers who see the coin come up tails aren't you. You just specified that they weren't. Who cares what they think?

If I jumped off a cliff and decided not to care about hitting the ground, I would still hit the ground. If I played a quantum lottery and decided not to care about copies who lost, almost all of me would still see a screen saying "You lose". It seems to me that there is a rule governing what I see happen next, which does not care what I care about. I am asking how that rule works, because it does so happen that I care about it.

You-now doesn't want to jump off the cliff because, among all the blobs of protoplasm that will exist in 5 minutes, you-now cares especially about one of them: the one that is causally connected in a certain way to the blob that is you-now. You-now evidently doesn't get to choose the nature of the causal connection that induces this concern. That nature was probably fixed by natural selection. That is why all talk about "determining to be the person who doesn't jump off the cliff" is ineffectual.

The question for doubters is this. Suppose, contrary to your intuition, that matters were just as I described. How would what-you-are-experiencing be any different? If you concede that there would be no difference, perhaps your confusion is just with how to talk about "your" future experiences. So then, what precisely is lost if all such talk is in terms of the experiences of future systems causally connected to you-now in certain ways?

Of course, committing to think this way highlights our ignorance about which causal connections are among these "certain ways". But our ignorance about this question doesn't mean that there isn't a determinate answer. There most likely is a determinate answer, fixed by natural selection and other determinants of what you-now cares about.

When you wake up, you will almost certainly have won (a trillionth of the prize). The subsequent destruction of winners (sort of - see below) reduces your probability of being the surviving winner back to one in a billion.

Merging N people into 1 is the destruction of N-1 people - the process may be symmetrical but each of the N can only contribute 1/N of themself to the outcome.

The idea of being (N-1)/N th killed may seem a little odd at first, but less so if you compare it to the case where half of one person's brain is merged with half of a different person's (and the leftovers discarded).

EDIT: Note that when the trillion were told they won, they were actually being lied to - they had won a trillionth part of the prize, one way or another.

2PlatypusNinja14y
Suppose that, instead of winning the lottery, you want your friend to win the lottery. (Or you want your random number generator to crack someone's encryption key, or you want a meteor to fall on your hated enemy, etc.) Then each of the trillion people would experience the full satisfaction from whatever random result happened.
1snarles13y
This. How does Yudkosky's careless statement "Just as computer programs or brains can split, they ought to be able to merge" not immediately light up as the weakest link of the entire post? If you think merging ought to work, then why not also think that quantum suicide ought to work?
1JamesAndrix14y
In the case where the people are computer programs, none of that works.
0Nubulous14y
If you mean that a quantitative merge on a digital computer is generally impossible, you may be right. But the example I gave suggests that merging is death in the general case, and is presumably so even for identical merges, which can be done on a computer.
1JamesAndrix14y
I fail to see why that is the general case. For that matter, I fail to see why losing some(many, most) of my atoms and having them be quickly replaced by atoms doing the exact same job should be viewed as me dying at all.
3Nubulous14y
If you have two people to start with, and one when you've finished, without any further stipulation about which people they are, then you have lost a person somewhere. To come to a different conclusion would require an additional rule, which is why it's the general case. That additional rule would have to specify that a duplicate doesn't count as a second person. But since that duplicate could subsequently go on to have a separate different life of its own, the grounds for denying it personhood seem quite weak. It's not dying in the sense of there no longer being a you, but it is still dying in the sense of there being fewer of you. To take the example of you being merged with someone, those atoms you lose, together with the ones you don't take from the other person, make enough atoms, doing the right jobs, to make a whole new person. In the symmetrical case, a second "you". That "you" could have gone on to live its own life, but now won't. Hence a "you" has died in the process. In other words, merge is equivalent to "swap pieces then kill". The above looks as though it will work just as well with bits, or the physical representation of bits, rather than atoms (for the symmetrical case).
3JamesAndrix14y
If a person were running on a inefficiently designed computer with transistors and wires much larger than they needed to be, it would be possible to peel away and discard (perhaps) half of the atoms in the computer without affecting it's operation or the person. This would be much like ebborian reproduction, but merely a shedding of atoms. In any sufficiently large information processing device, there are two or more sets of atoms (or whatever its made of) processing the same information, such they they could operate independently of each other if they weren't spatially intertwined. Why are they one person when spatially intertwined, but two people when they are apart? That they 'could have' gone on independently is a counterfactual in the situation that they are both receiving the same inputs. You 'could' be peeled apart into two people, but both halves of your parts are currently still making up 1 person. Personhood is in the pattern. Not the atoms or memory or whatever. There's only another person when there is another sufficiently different pattern. merge is equivalent to 'spatially or logically reintegrate, then shed atoms or memory allocation as desired'

It's helpful in these sorts of problems to ask the question "What would evolution do?". It always turns out to be coherent, reality-based actions. Even though evolution, to the extent that it "values" things, values different things than I do, I'd like my actions to be comparably coherent and reality-based.

Regarding the first horn: Regardless of whether simple algorithms move "subjective experience" around like a fluid, if the simple algorithms take some resources, evolution would not perform them.

Regarding the second horn: If there was an organism that routinely split, merged, played lottery, and priced side-bets on whether it had won the lottery, then, given zero information about whether it had won the lottery, it would price the side-bet at the standard lottery odds. Splitting and merging, so long as the procedure did not provide any new information, would not affect its price.

Regarding the third horn: Evolution would certainly not create an organism that hurls itself off cliffs without fear. However, this is not because of it "cares" about any thread of subjective experience. Rather, this is because of the physical continuity. Compare this with evolution's choice in an environment where there are "transporters" that accurately convey entities by molecular disassembly/reassembly. Creatures which had evolved in that environment would certainly step through those transporters without fear.

I can't answer the fourth or fifth horns; I'm not sure I understand them.

Whenever I read about "weight of experience", "quantum goo", "existentness" etc. I can't keep myself of also thinking of "vital spark", "phlogiston", "ether" and other similar stuff... And it somehow spoils the whole fun...

In the history of mankind, hard looking (meta-)physical dilemmas were much more often resolved by means of elimination rather than by introduction of new "essences". The moral of the history of physics so far is that relativity typically trumps absoluteness in the long run.

For example, I would not be surprised at all, if it turned out that experienced Born probabilities would not be absolute, but would depend on some reference frame (in a very high dimensional space) just like the experience of time, speed, mass, etc. depends on the relativistic frame of reference.

Marcello and I use the term "reality-fluid" to remind ourselves that we're confused.

3Psy-Kosh14y
Dunno about others but I agree that terms like that seem to indicate a serious confusion, or at least something that I am very confused about, and seems others here do too. We're more using them as a way of talking about our confusion. Just noting "there's something here we're failing to comprehend" isn't enough to help us comprehend it. It's more a case of "we're not sure what concepts to replace those terms with", at least so it is to me.

Following Nominull and Furcas, I bite the third bullet without qualms for the perfectly ordinary obvious reasons. Once we know how much of what kinds of experiences will occur at different times, there's nothing left to be confused about. Subjective selfishness is still coherent because you're not just an arbitrary observer with no distinguishing characteristics at all; you're a very specific bundle of personality traits, memories, tendencies of thought, and so forth. Subjective selfishness corresponds to only caring about this one highly specific bundle: only caring about whether someone falls off a cliff if this person identifies as such-and-such and has such-and-these specific memories and such-and-those personality traits: however close a correspondence you need to match whatever you define as personal identity.

The popular concepts of altruism and selfishness weren't designed for people who understand materialism. Once you realize this, you can just recast whatever it was you were already trying to do in terms of preferences over histories of the universe. It all adds up to, &c., &c.

3Wei Dai14y
I agree that giving up anticipation does not mean giving up selfishness. But as Dan Armak pointed out there is another reason why you may not want to give up anticipation: you may prefer to keep the qualia of anticipation itself, or more generally do not want to depart too much from the subjective experience of being human. Eliezer, if you are reading this, why do you not want to give up anticipation? Do you still think it means giving up selfishness? Is it for Dan Armak's reason? Or something else?
1komponisto14y
The (only) trouble with this is that it doesn't answer the question about what probabilities you_0 should assign to various experiences 5 seconds later. Personal identity may not be ontologically fundamental, it may not even be the appropriate sort of thing to be programmed into a utility function -- but at the level of our everyday existence (that is, at whatever level we actually do exist), we still have to be able to make plans for "our own" future.
5Z_M_Davis14y
I would say that the ordinarily very useful abstraction of subjective probability breaks down in situations that involve copying and remerging people, and that our intuitive morality breaks down when it has to deal with measure of experience. In the current technological regime, this isn't a problem at all, because the only branching we do is quantum branching, and there we have this neat correspondence between quantum measure and subjective probability, so you can plan for "your own" future in the ordinary obvious way. How you plan for "your own" future in situations where you expect to be copied and merged depends on the details of your preferences about measure of experience. For myself, I don't know how I would go about forming such preferences, because I don't understand consciousness.
2Douglas_Knight14y
Quantum suicide is already a problem in the current regime, if you allow preference over measure. Splitting and merging adds another problem, but I think it is a factual problem, not an ethical problem. At least, I think that there is a factual problem before you come to the ethical problem, which may be the same as for Born measure.

Just an aside - this is obviously something that Eliezer - someone highly intelligent and thoughful - has thought deeply about, and has had difficulty answering.

Yet most of the answers - including my own - seem to be of the "this is the obvious solution to the dilemma" sort.

8DanArmak14y
...Only each obvious solution proposed is different.
0Stuart_Armstrong14y
Bien entendue...
0[anonymous]14y
It's philosophy, what'd u expect?
0casebash8y
People often miss a solution that it obvious in retrospect.

Bonus: if you're uncomfortable with merging/deleting copies, you can skip that part! Just use the lottery money to buy some computing equipment and keep your extra copies running in lockstep forever. Is this now an uncontroversial algorithm for winning the lottery, or what?

More near-equivalent reformulations of the problem (in support of the second horn):

  • A trillion copies will be created, believing they have won the lottery. All but one will be killed (1/trillion that your current state leads directly to your future state). If you add some uniportant differentiation between the copies - give each one a speratate number - then the situation is clearer: you have one chance in a trillion that the future self will remember your number (so your unique contribution will have 1/trillion chance of happening), while he will be certain to believe he has won the lottery (he gets that belief from everyone.

  • A trillion copies are created, each altruistically happy that one among the group has won the lottery. One of them at random is designated the lottery winner. Then everyone else is killed.

  • Follow the money: you (and your copies) are not deriving utility from winning the lottery, but from spending the money. If each copy is selfish, there is no dilema: the lottery winnings divided amongst a trillion cancels out the trillion copies. If each copy is altruistic, then the example is the same as above; in which case there is a mass of utility generated from the copies, which vanish when the copies vanish. But this extra mass of utility is akin to the utility generated by: "It's wonderful to be alive. Quick, I copy myself, so now many copies feel it's wonderful to be alive. Then I delete the copies, so the utility goes away".

3casebash8y
"You (and your copies) are not deriving utility from winning the lottery, but from spending the money" I would say that you derive utility from knowing that you've won money you can spend. But, if you only get $1, you haven't won very much. I think that a better problem would be if you split if your favourite team won the super bowl. Then you'd have a high probability of experiencing this happiness, and the number of copies wouldn't reduce it.
0Stuart_Armstrong8y
Neat!

Time and (objective or subjective) continuity are emergent notions. The more basic notion they emerge from is memory. (Eliezer, you read this idea in Barbour's book, and you seemed to like it when you wrote about that book.)

Considering this: yes, caring about the well-being of "agents that have memories of formerly being me" is incoherent. It is just as incoherent as caring about the well-being of "agents mostly consisting of atoms that currently reside in my body". But in typical cases, both of these lead to the same well known and evolutionarily useful heuristics.

I don't think any of the above implies that "thread of subjective experience" is a ridiculous thing, or that you can turn into being Britney Spears. Continuity being an emergent phenomenon does not mean that it is a nonexistent one.

As for what happens ten seconds after that, you have no way of knowing how many processors you run on, so you shouldn't feel a thing

Here's the problem, as far as I can see. You shouldn't feel a thing, but that would also be true if none of you ever woke up again. "I won't notice being dead" is not an argument that you won't be dead, so lottery winners should anticipate never waking up again, though they won't experience it (we don't anticipate living forever in the factual world, even though no one ever notices being dead).

I'm sure there's some reason this is considered invalid, since quantum suicide is looked on so favorably around here. :)

The reason is simply that, in the multiple worlds interpretation, we do survive-- we just also die. If we ask "Which of the two will I experience?" then it seems totally valid to argue "I won't experience being dead."

2randallsquared14y
So it basically comes back to pattern-as-identity instead of process-as-identity. Those of me who survive won't experience being dead. I think you can reach my conclusions by summing utility across measure, though, which would make an abrupt decrease in measure equivalent to any other mass death.

I still have trouble biting that bullet for some reason. Maybe I'm naive, I know, but there's a sense in which I just can't seem to let go of the question, "What will I see happen next?" I strive for altruism, but I'm not sure I can believe that subjective selfishness - caring about your own future experiences - is an incoherent utility function; that we are forced to be Buddhists who dare not cheat a neighbor, not because we are kind, but because we anticipate experiencing their consequences just as much as we anticipate experiencing our own. I don't think that, if I were really selfish, I could jump off a cliff knowing smugly that a different person would experience the consequence of hitting the ground.

I don't really understand your reasoning here. It's not a different person that will experience the consequences of hitting the ground, it's Eliezer+5. Sure, Eliezer+5 is not identical to Eliezer, but he's really, really, really similar. If Eliezer is selfish, it makes perfect sense to care about Eliezer+5 too, and no sense at all to care equally about Furcas+5, who is really different from Eliezer.

9orthonormal14y
Suppose I'm duplicated, and both copies are told that one of us will be thrown off a cliff. While it makes some kind of sense for Copy 1 to be indifferent (or nearly indifferent) to whether he or Copy 2 gets tossed, that's not what would actually occur. Copy 1 would probably prefer that Copy 2 gets tossed (as a first-order thing; Copy 1's morals might well tell him that if he can affect the choice, he ought to prefer getting tossed to seeing Copy 2 getting tossed; but in any case we're far from mere indifference). There's something to "concern for my future experience" that is distinct from concern for experiences of beings very like me.
4Furcas14y
I have the same instincts, and I would have a very hard time overriding them, were my copy and I put in the situation you described, but those instincts are wrong.
2orthonormal14y
Emend "wrong" to "maladapted for the situation" and I'll agree.
5DanielVarga14y
These instincts are only maladapted for situations found in very contrived thought experiments. For example, you have to assume that Copy 1 can inspect Copy 2's source code. Otherwise she could be tricked into believing that she has an identical copy. (What a stupid way to die.) I think our intuitions are already failing us when we try to imagine such source code inspections. (To put it another way: we have very few in common with agents that can do such things.)
3orthonormal14y
It would suffice, instead, to have strong evidence that the copying process is trustworthy; in the limit as the evidence approaches certainty, the more adaptive instinct would approach indifference between the cases.
0[anonymous]9y
good thought experiment, but I actually would be indifferent, as long as I actually believed that my copy was genuine, and wouldn't be thrown off a cliff. Unfortunately I can't actually imagine any evidence that would convince me of this. I wonder if that's the source of your reservations too--if the reason you imagine Copy 1 caring is because you can't imagine Copy 1 being convinced of the scenario.

And the third horn of the trilemma is to reject the idea of the personal future - that there's any meaningful sense in which I can anticipate waking up as myself tomorrow, rather than Britney Spears. Or, for that matter, that there's any meaningful sense in which I can anticipate being myself in five seconds, rather than Britney Spears. In five seconds there will be an Eliezer Yudkowsky, and there will be a Britney Spears, but it is meaningless to speak of the current Eliezer "continuing on" as Eliezer+5 rather than Britney+5; these are simply

... (read more)

Oddly, I feel myself willing to bite all three bullets. Maybe I am too willing to bite bullets? There is a meaningful sense in which I can anticipate myself being one of the future people who will remember being me, though perhaps there isn't a meaningful way to talk about which of those many people I will be; I will be all of them.

I suggested that, in some situations, questions like "What is your posterior probability?" might not have answers, unless they are part of decision problems like "What odds should you bet at?" or even "What should you rationally anticipate to get a brain that trusts rational anticipation?". You didn't comment on the suggestion, so I thought about problems you might have seen in it.

In the suggestion, the "correct" subjective probability depends on a utility function and a UDT/TDT agent's starting probabilities, which ... (read more)

I'm terribly sorry, but I had to delete that comment because it used the name of M-nt-f-x, which, if spoken in plain text, causes that one to appear. He Googles his name.

Here was the original comment by outlawpoet:

Marc Geddes was a participant in the SL4 list back when it was a bit more active, and kind of proverbial for degenerating into posting very incoherent abstract theory. He's basically carrying on the torch for M-nt-f-x and Achimedes Plutonium and other such celebrities of incommunicable genius.

Now please stop replying to this thread!

The problem is that copying and merging is not as harmless as it seems. You are basically doing invasive surgery on the mind, but because it's performed using intuitively "non-invasive" operations, it looks harmless. If, for example, you replaced the procedure with rewriting "subjective probability" by directly modifying the brain, the fact that you'd have different "subjective probability" as a result won't be surprising.

Thus, on one hand, there is an intuition that the described procedure doesn't damage the brain, and on the... (read more)

The problem with anthropic reasoning and evidence is that unlike ordinary reasoning and evidence, it can't be transferred between observers. Even if "anthropic psychic powers" actually do work, you still should expect all other observers to report that they don't.

5DanArmak14y
And not only do they report that they don't see it working for you. The bigger problem is that they report that it doesn't even work for them when they try it.

I bite the third bullet. I am not as gifted with words as you are to describe why biting it is just and good and even natural if you look at it from a certain point of view, but...

You are believing in something mystical. You are believing in personal identity as something meaningful in reality, without giving any reason why it ought to be, because that is how your algorithm feels from the inside. There is nothing special about your brain as compared to your brain+your spinal cord, or as compared to your brain+this pencil I am holding. How could there be... (read more)

1Johnicholas14y
EY seems to have equated the third bullet with throwing oneself off of cliffs. Do you throw yourself off of cliffs? Why or why not?
8Nick_Tarleton14y
It sounds to me like EY is equating the second bullet with "perfect altruism is coherent, as is caring only about one's self at the current moment, but nothing in between is." To that, though, as Furcas says, one can be selfish according to similarity of pattern rather than ontologically privileged continuity.
1torekp14y
Or one could be selfish according to a non-fundamental, ontologically reducible continuity. At least, I don't see why not. Has anyone offered an argument for pattern over process? randallsquared has it dead right, I think.
3Z_M_Davis14y
I don't throw myself off cliffs for very roughly the same reason I don't throw other people off cliffs.
9SilasBarta14y
And for the same reason you buy things for yourself more often than for other people? And for the same reason you (probably) prefer someone else falling off a cliff than yourself?
2Z_M_Davis14y
I was trying to be cute.
4SilasBarta14y
Considering that your cute comment was consistent with your other comments in this discussion, I think I can be forgiven for thinking you were serious. Actually, which of your other comments here are just being cute?
5Z_M_Davis14y
Right, so of course I'm rather selfish in the sense of valuing things-like-myself, and so of course I buy more things for myself than I do for random strangers, and so forth. But I also know that I'm not ontologically fundamental; I'm just a conjunction of traits that can be shared by other observers to various degrees. So "I don't throw myself off cliffs for very roughly the same reason I don't throw other people off cliffs" is this humorously terse and indirect way of saying that identity is a scalar, not a binary attribute. (Notice that I said "very roughly the same reason" and not "exactly the same reason"; that was intentional.)
-2SilasBarta14y
And ... you expected everyone else to get that out of your cute comment? You know, sometimes you just have to throw in the towel and say, "Oops. I goofed." ETA: I'm sure that downmod was because this comment was truly unhelpful to the discussion, rather than because it made someone look bad.
4Z_M_Davis14y
Oops. I goofed.
4Douglas_Knight14y
I am sad to see this comment. Perhaps you were mistaken in how clear the comment was to how broad an audience, but I think the original comment was valuable and that we lose a lot of our ability to communicate if we are too careful.
0SilasBarta14y
Except this wasn't an issue of being too careful, and it definitely doesn't count as good communication! Z_M_Davis made a remark that was both poorly-reasoned and supportive of every other comment he left in the discussion (in trivializing the privileged state of any choice of identity). If he had been arguing against the third horn, okay, maybe it could have been read as "oh, he's cleverly mocking a position he disagrees with". But then he comes back with, "I was trying to be cute." Okay, so he's ... doing self-parody. Great -- we all need to be able to laugh at ourselves. So what's his real position, then? Oh, you see, he was making a very subtle point about identity being scalar rather than binary (which has some as-of-yet unspecified implication for the merit of his position). And there was a hidden argument in there that allows him to see his life as no different from any others and yet still act in preference to himself. And it was obvious what distinction he was making by using the words "very roughly the same reason" instead of "exactly the same reason". I'm sorry, but that's just not "how it works". You can claim illusion of transparency issues if the assumed common knowledge is small, and you have a reasonable basis for assuming it, and your full explanation doesn't look blatantly ad hoc. In other words, anywhere but here. I'm sorry to belabor the point, but yes, sometimes you just have to admit you goofed. Mistakes are okay! We all make them! But we don't all try to say "I meant to do that".
1Z_M_Davis14y
See Furcas's comment. I never said it was no different. Elsewhere in the thread, I had argued that selfishness is entirely compatible with biting the third bullet. Egan's Law. I disagree; if it had been obvious, I wouldn't have had to point it out explicitly. Maybe the cognitive history would help? I had originally typed "the same reason," but added "very roughly" before posting because I anticipated your objection. I think the original was slightly funnier, but I thought it was worth trading off a little of the humor value in exchange for making the statement more defensible when taken literally. I'm curious. If what actually happened looks ad hoc to you, what's your alternative theory? If you don't trust what I say about what I was thinking, then what do you believe instead? You seem to think I've committed some error other than writing two admittedly somewhat opaque comments, but I'm not sure what it's supposed to be.
0Alicorn14y
Is there any chance that you will soon mature / calm down / whatever it is you need to do to stop being so hostile, so frequently? This is only the latest example of you coming down on people with the utmost contempt for fairly minor offenses, if they're offenses at all. It looks increasingly like you think anyone who conducts themselves in the comments differently than you prefer ought to be beheaded or something. It's really unpleasant to read, and I don't think it's likely to be effective in getting people to adopt your standards of behavior. (I can think of few posters I'm less motivated to emulate than you, as one data point.) Edit: I downvoted the parent of this comment because I would like to see fewer comments that resemble it.
-3SilasBarta14y
I hate to play tu quoque, but it's rather strange of you to make this criticism considering that just a few months ago you gave a long list of very confining restrictions you wanted on commenters, enforced by bannings. Despite their best efforts, no one could discern the pattern behind what made something beyond-the-pale offensive, so you were effectively asking for unchecked, arbitrary authority to remove comments you don't like. You even went so far as to ask to deputize your own hand-picked cadre of posters "with their heads on right" to assist in classifying speech as unacceptable! Yes, Alicorn, I've been very critical of those who claim objectivity in modding people they're flaming, but I don't think I've ever demanded the sort of arbitrary authority over the forum that you feel entitled to. I will gladly make my criticisms more polite in the future, but I'm not going to apologize for having lower karma than if I abused the voting system the way some of you seem to. And in the meantime, perhaps you could make it a habit of checking if the criticisms you make of others could apply to yourself. I'm not asking that you be perfect or unassailable in this respect. I'm not even asking that you try to adhere to your own standards. I just ask that you check for whether you're living up to it. Edit: I didn't downvote the parent of this comment because I'm not petty like that.
-6Alicorn14y
-7SilasBarta14y
1Nominull14y
A lot of people, especially religious people, equate lack of belief in a fundamental meaning of life with throwing oneself off cliffs. Eliezer is committing the same sort of mistake.
3loqi14y
No, I think he's just pointing out that the common intuitions behind anticipatory fear are grossly violated by the third horn. I'd like to see you chew this bullet a bit more, so try this version. You are to be split (copied once). One of you is randomly chosen to wake up in a red room and be tortured for 50 years, while the other wakes up in a green room and suffers a mere dust speck. Ten minutes will pass for both copies before the torture or specking commences. How much do you fear being tortured before the split? Does this level of fear go up/down when you wake up in a red/green room? To accept the third horn seems to imply that you should feel no relief upon waking in the green room.
4Nick_Tarleton14y
Only if you assume feelings of relief should bind to reality (or reality+preferences) in a particular way.
6loqi14y
Good point, I should have phrased that differently: "To accept the third horn seems to imply that any relief you feel upon waking in the green room is just 'legacy' human intuition, rather than any rational expectation of having avoided future suffering."
3gwern14y
You know, your example is actually making that horn look more attractive: replace the torture to the person with '50000 utilities subtracted from the cosmos', etc, and then it's obvious that the green-room is no grounds for relief since the -50000 is still a fact. More narrowly, if you valued other persons equal to yourself, then the green room is definitely no cause for relief. We could figure out how much you value other people by varying how bad the torture, is and maybe adding a deal where if the green-room person will flip a fair coin (heads, the punishment is swapped; tails, no change), the torture is lessened by n. If you value the copy equal to yourself, you'll be willing to swap for any difference right down to 1, since if it's tails, there's no loss or gain, but if it's heads, there's a n profit. Now, of course even if the copy is identical to yourself, and even if we postulate that somehow the 2 minds haven't diverged (we could do this by making the coinflip deal contingent on being the tortured one - 2 identical rooms, neither of which knows whether they are the tortured one; by making it contingent, there's no risk in not taking the bet), I think essentially no human would take the coinflip for just +1 - they would only take it if there was a major amelioration of the torture. Why? Because pain is so much realer and overriding to us, which is a fact about us and not about agents we can imagine. (If you're not convinced, replace the punishments with rewards and modify the bet to increase the reward but possibly switch it to the other fellow; and imagine a parallel series of experiments being run with rational agents who don't have pain/greed. After a lot of experiments, who will have more money?)
1loqi14y
Yes, and this hypothesis can even be weakened a bit, since the other persons involved are nearly identical to you. All it takes is a sufficiently "fuzzy" sense of self. To clarify what you mean by "haven't diverged"... does that include the offer of the flip? E.g., both receive the offer, but only one of the responses "counts"? Because I can't imagine not taking the flip if I knew I was in such a situation... my anticipation would be cleanly split between both outcomes due to indexical uncertainty. It's a more complicated question once I know which room I'm in.
0gwern14y
Well, maybe I wasn't clear. I'm imagining that there are 2 green rooms, say, however, one room has been secretly picked out for the torture and the other gets the dustspeck. Each person now is made the offer: if you flip this coin, and you are not the torture room, the torture will be reduced by n and the room tortured may be swapped if the coin came up heads; however, if you are the torture room, the coin flip does nothing. Since the minds are the same, in the same circumstances, with the same offer, we don't need to worry about what happens if the coins fall differently or if one accepts and the other rejects. The logic they should follow is: if I am not the other, then by taking the coin flip I am doing myself a disservice by risking torture, and I gain under no circumstance and so should never take the bet; but if I am the other as well, then I lose under no circumstance so I should always take the bet. (I wonder if I am just very obtusely reinventing the prisoner's dilemma or newcomb's paradox here, or if by making the 2 copies identical I've destroyed an important asymmetry. As you say, if you don't know whether "you" have been spared torture, then maybe the bet does nothing interesting.)
0loqi14y
I'm not sure what "not being the other" means here, really. There may be two underlying physical processes, but they're only giving rise to one stream of experience. From that stream's perspective, its future is split evenly between two possibilities, so accepting the bet strictly dominates. Isn't this just straightforward utility maximization? The reason the question becomes more complicated if the minds diverge is that the concept of "self" must be examined to see how the agent weights the experiences of an extremely similar process in its utility function. It's sort of a question of which is more defining: past or future. A purely forward-looking agent says "ain't my future" and evaluates the copy's experiences as those of a stranger. A purely backward-looking agent says "shares virtually my entire past" and evaluates the copy's experiences as though they were his own. This all assumes some coherent concept of "selfishness" - clearly a purely altruistic agent would take the flip. The identical copies scenario is a prisoner's dilemma where you make one decision for both sides, and then get randomly assigned to a side. It's just plain crazy to defect in a degenerate prisoner's dilemma against yourself. I think this does destroy an important asymmetry - in the divergent scenario, the green-room agent knows that only his decision counts. Speaking for my own values, I'm still thoroughly confused by the divergent scenario. I'd probably be selfish enough not to take the flip for a stranger, but I'd be genuinely unsure of what to do if it was basically "me" in the red room.
0orthonormal14y
The SIA predicts that you will say "no".

Let's explore this scenario in computational rather than quantum language.

Suppose a computer with infinite working memory, running a virtual world with a billion inhabitants, each of whom has a private computational workspace consisting of an infinite subset of total memory.

The computer is going to run an unusual sort of 'lottery' in which a billion copies of the virtual world are created, and in each one, a different inhabitant gets to be the lottery winner. So already the total population after the lottery is not a billion, it's a billion billion, spre... (read more)

0[anonymous]14y
If everyone performs the copying strategy, then after the lottery and the copying but before the merge, the majority of all agents (summed over all the worlds) have won the lottery. So if copy-merge works at all, Yu'el is still favored, and so is everyone else.

....you have to throw away the idea that your joint subjective probabilities are the product of your conditional subjective probabilities....If you win the lottery, the subjective probability of having still won the lottery, ten seconds later, is ~1.

If copying increases your measure, merging decreases it. When you notice yourself winning the lottery, you are almost certainly going to cease to exist after ten seconds.

1ESRogs10y
Which is to pick Nick's suggestion, and something like horn #2, except perhaps without the last part where you "anticipate that once you experience winning the lottery you will experience having still won it ten seconds later." Scanning through the comments for "quantum suicide," it sounds like a few others agree with you.

I don't think you have the third horn quite right. It's not that you're equally likely to wake up as Brittany Spears. It's that the only meaningful idea of "you" exists only right now. Your subjective anticipation of winning the lottery in five minutes should be zero. You clearly aren't winning the lottery or in five minutes.

Also, isn't that more of a quintlemma?

Maybe I am too late to comment here and it is already covered in collapsed comments, but it looks like that it is possible to make this experiment in real life.

Imagine that instead of copying, I will use waking up. If I win, I will be waked up 3 times and informed that I won and will be given a drug which will make me forget the act of awakening. If I lose, I will be wakened only one time and informed that I lost. Now I have 3 to 1 observer moments where I informed about winning.

In such setup in is exactly the Sleeping beauty problem, with all its pro and contra, which I will not try to explore here.

Tafter thinking about the Anthropic Trilemma for awhile, I've come up with an alternative resolution which I think is better than, or at least simpler than, any of the other resolutions. Rather than try to construct a path for consciousness to follow inductively forwards through time, start at the end and go backwards: from the set of times at which an entity I consider to be a copy of me dies, choose one at random weighted by quantum measure, then choose uniformly at random from all paths ending there.

The trick is that this means that while copying your m... (read more)

I deny that increasing the number of physical copies increases the weight of an experience. If I create N copies of myself, there is still just one of me, plus N other agents running my decision-making algorithms. If I then merge all N copies back into myself, the resulting composite contains the utility of each copy weighted by 1/(N+1).

My feeling about the Boltzmann Brain is: I cheerfully admit that there is some chance that my experience has been produced by a random experience generator. However, in those cases, nothing I do matters anyway. Thus I d... (read more)

Banned for attempted top-posting of Geddes-standard gibberish.

3wedrifid14y
Geddes-standard? I don't understand the reference. An unflattering comparison to the biologist? Edit: Thanks for the explaination EY. (And my expressed contempt for whoever thinks a query trying to understand the reference must be punished.)

I don't know the answer either. My best guess is that the question turns out to involve comparing incommensurable things, but I haven't pinned down which things. (As I remarked previously, I think the best answer for policy purposes is to just optimize for total utility, but that doesn't answer the question about subjective experience.)

But one line of attack that occurs to me is the mysterious nature of the Born probabilities.

Suppose they are not fundamental, suppose the ultimate layer of physics -- maybe superstrings, maybe something else -- generates the... (read more)

1Eliezer Yudkowsky14y
That's indeed what we should all be hoping for. But what possible set of "axioms" for subjective experience - never mind what possible underlying physics - could correspond to the Born probabilities, while solving the computer-processor trilemma as well?
3rwallace14y
Well... following this line of thought, we should expect that the underlying physics is not special, because any physics that satisfies certain generic properties will lead to subjective experience of the Born probabilities. Suppose we can therefore without loss of generality take the underlying physics to be equivalent to a digital computer programmed in a straightforward way, so that the quantum and computer trilemmas are equivalent. Is there any set of axioms that will lead (setting aside other intuitions for the moment) to subjective experience of the Born probabilities in the case where we are running on a computer and therefore do know the underlying physics? If there is, that would constitute evidence for the truth of those axioms even if they are otherwise counterintuitive; if we can somehow show that there is not, that would constitute evidence that this line of thought is barking up the wrong tree.
1Psy-Kosh14y
Elaborate on that bit please? Thanks.
0rwallace14y
Well basically, we start off with the claim (which I can't confirm of my own knowledge, but have no reason to doubt) that the Born rule has certain special properties, as explained in the original post. We observe that the Born rule seems to be empirically true in our universe. We would like an explanation as to why our universe exhibits a rule with special properties. Consider the form this explanation must take. It can't be because the Born rule is encoded into the ultimate laws of physics, because that would only push the mystery back a few steps. It should be a logical conclusion that we would observe the Born rule given any underlying physics (within reason). Of course there is far too much armchair handwaving here to constitute proof, but I think it at least constitutes an interesting conjecture.
0Psy-Kosh14y
Well, even if it turns out that there're special properties of our physics that are required to produce the Born rule, I'd say that mystery would be a different, well, kind of mystery. Right now it's a bit of "wtf? where is this bizzaro subjective nonlinearity etc coming from? and it seems like something 'extra' tacked onto the physics" If we could reduce that to "these specific physical laws give rise to it", then even though we'd still have "why these laws and not others", it would, in my view, be an improvement over the situation in which we seem to have an additional law that seems almost impossible to even meaningfully phrase without invoking subjective experience. I do agree though that given the special properties of the rule, any special properties in the underlying physics that are needed to give rise to the rule should be in some sense "non arbitrary"... that is, should look like, well, like a nonaribitrarily selected physical rule.
0Eliezer Yudkowsky14y
Sounds like a right question to me. Got an answer? A related problem: If we allow unbounded computations, then, when we try to add up copies, we can end up with different limiting proportions of copies depending on how we approach t -> infinity; and we can even have algorithms for creating copies such that their proportions fail to converge. (1 of A, 3 of B, 9 of A, 27 of B, etc.) So then either it is a metaphysical necessity that reality be finite, because otherwise our laws will fail to give correct answers; or the True Rules must be such as to give definitive answers in such a situation.
1rwallace14y
I'm afraid I'm not familiar enough with the Born probabilities to know how to approach an answer -- oh, I've been able to quote the definition about squared amplitudes since I was a wee lad, but I've never had occasion to actually work with them, so I don't have any intuitive feel about their implications. As for the problem of infinity, you're right of course, though there are other ways for that to arise too -- for example, if the underlying physics is analog rather than digital. Which suggests it can't be fiated away. I don't know what the solution is, but it reminds me of the way cardinality says all shapes contain the same number of points, so it was necessary to invent measure to justify the ability to do geometry.
0Psy-Kosh14y
Deeply fundamentally analog physics, ie, infinite detail, would just be another form of infinity, wouldn't it? So it's a variation of the same problem of "what happens to all this when there's an infinity involved?"
1bogus14y
To the best of our understanding, there's no such thing as "infinite detail" in physics. Physical information is limited by the Bekenstein bound.
0Psy-Kosh14y
Sorry, I may have been unclear. I didn't mean to make a claim that physics actually does have this property, but rather I was saying that if physics did have this property, it would just be another instance of an infinity, rather than an entirely novel source for the problem mentioned. (Also, I'm unclear on the BB, if it takes into account possible future tech that may be able to manipulate the geometry of spacetime to some extent. ie, if we can do GR hacking, would that affect the bound or are the limits of that effectively already precomputed into that?)
0rwallace14y
Yes, that is my position on it.

The third option seems awfully close to the second. In the second, you anticipate winning the lottery for a few seconds, and then going back to not. In the third, the universe anticipates winning the lottery for a few seconds, and then going back to Britney.

I'm just wondering what would Britney Spears say when she reads this.

Now you could just bite this bullet. You could say, "Sounds to me like it should work fine." You could say, "There's no reason why you shouldn't be able to exert anthropic psychic powers." You could say, "I have no problem with the idea that no one else could see you exerting your anthropic psychic powers, and I have no problem with the idea that different people can send different portions of their subjective futures into different realities."

I think there are other problems that may prevent the "anthropic psychic powers" example from working (maybe co

... (read more)

The odds of winning the lottery are ordinarily a billion to one. But now the branch in >which you win has your "measure", your "amount of experience", temporarily >multiplied by a trillion. So with the brief expenditure of a little extra computing power, >you can subjectively win the lottery - be reasonably sure that when next you open >your eyes, you will see a computer screen flashing "You won!"

As I see it, the odds of being any one of those trillion "me"s in 5 seconds is 10^21 to one(one trillion... (read more)

0jacob_cannell12y
I have nearly the same viewpoint and was surprised to find what seems to me to be the obvious solution so far down this thread. One little nitpick: From your analysis, I think you mean you expect there is a 1 in a billion chance there will be 1 trillion "people who are me who experience winning" at 5 sec.

I'm coming in a bit late, and not reading the rest of the posts, but I felt I had to comment on the third horn of the trilemma, as it's an option I've been giving a lot of thought.

I managed to independently invent it (with roughly the same reasoning) back in high school, though I haven't managed to convince myself of it or, for that matter, to explain it to anyone else. Your explanation is better, and I'll be borrowing it.

At any rate. One of your objections seems to be "...to assert that you can hurl yourself off a cliff without fear, because whoever ... (read more)

I figured it out! Roger Penrose is right about the nature of the brain!

just kidding.

In quantum copying and merging, every "branch" operation preserves the total measure of the original branch,

Maybe branch quantum operations don't make new copies, but represent already existing but identical copies "becoming" no longer identical?

In the computer program analogy: instead of having one program at time t and n slightly different versions at time t+1, start out with n copies already existing (but identical) at time t, and have each one change in the branching. If you expect a t+2, you need to start with at least n^2 copie... (read more)

2Psy-Kosh14y
But where does all the quantum interference stuff come from then?
0MichaelHoward14y
I'm not trying to resolve any quantum interference mysteries with the above, merely anthropic ones. I have absolutely no idea where the born probabilities come from.
2Psy-Kosh14y
Sorry, I was unclear. I meant "if what you say is the correct explanation, then near as I can tell, there shouldn't be anything resembling quantum interference. In your model, where is there room for things to 'cancel out' if copies just keep multiplying like that?" Or did I misunderstand what you were saying?
2MichaelHoward14y
Ah, sorry if I wasn't clear. The copies wouldn't multiply. In the computer program analogy, you'd have the same number of programs at every time step. So instead of doing this... Step1: "Program". Step2: "Program0", "Program1". Step3: "Program00", "Program01", "Program10", "Program11". You do this... Step1: "Program", "Program", "Program", "Program", ... Step2: "Program0", "Program1", "Program0", "Program1", ... Step3: "Program00", "Program01", "Program10", "Program11", ... For the sake of simplicity this is the same algorithm, but with part of the list not used when working out the next step. If our universe did this, surely at any point in time it would produce exactly the same experimental results as if it didn't. If we're not experiencing the last instant of experienceable existence, I think that may imply that the second model is closer to the truth, and also that someone who runs redundantly on three processors gets three times as much weight as someone who runs on one processor, for the reasons above.
1Psy-Kosh14y
Ah, okay. Sorry I misunderstood.
0[anonymous]14y
Ah, sorry if I wasn't clear. The copies wouldn't multiply. In the computer program analogy, you'd have the same number of programs at every time step. So instead of doing this... Step1: "Program". Step2: "Program0", "Program1". Step3: "Program00", "Program01", "Program10", "Program11". You do this... Step1: "Program", "Program", "Program", "Program", ... Step2: "Program0", "Program1", "Program0", "Program1", ... Step3: "Program00", "Program01", "Program10", "Program11", ... For the sake of simplicity this is the same algorithm, but with part of the list not used when working out the next step. If our universe did this, surely at any point in time it would produce exactly the same experimental results as if it didn't.

The "second horn" seems to be phrased incorrectly. It says:

"you can coherently anticipate winning the lottery after five seconds, anticipate the experience of having lost the lottery after fifteen seconds, and anticipate that once you experience winning the lottery you will experience having still won it ten seconds later."

That's not really right - the fate of most of those agents that experience a win of the lottery is to be snuffed out of existence. They don't actually win the lottery - and they don't experience having won it eleven... (read more)

3Johnicholas14y
If I understand the proposed merging procedure correctly, the procedure treats the trillion observers who experience a win of the lottery symmetrically. None of them are "snuffed" any more than any other. For each of the observers, there is a continuous space-time-causality "worm" connecting to the future self who spends the money. This space-time-causality worm is supposed to be as analogous as possible to the one that connects any ordinary moment in your life to your future self. The difference is that this one merges (symmetrically) with almost a trillion others, all identical.
0timtyler14y
I see, I think. I can't help wondering what the merge procedure does with any flipped bits in the diff, though. Anyway, horn 2 now seems OK - I think it describes the situation.
-1timtyler14y
Rereading the comments on this thread, the problem is more subtle than I had thought - and I had better retract the above comment. I am inclined towards the idea that copying doesn't really alter the pattern - but that kind of anthropic reasoning seems challenging to properly formalise under the given circumstances.
0Eliezer Yudkowsky14y
Yup! If you can't do the merge without killing people, then the trilemma is dissolved.

I strive for altruism, but I'm not sure I can believe that subjective selfishness - caring about your own future experiences - is an incoherent utility function; that we are forced to be Buddhists who dare not cheat a neighbor, not because we are kind, but because we anticipate experiencing their consequences just as much as we anticipate experiencing our own. I don't think that, if I were really selfish, I could jump off a cliff knowing smugly that a different person would experience the consequence of hitting the ground.

It strikes me that this is a p... (read more)

I will bite the first horn of the trilemma. I'm will argue that the increase in subjective probability results from losing information and that it is no different from other situations where you lose information in such a way to make subjective probabilities seem higher. For example, if you watch the lotto draw, but then forget every number except those that match your ticket, your subjective probability that you won will be much higher than originally.

Let's imagine that if you win the lottery that a billion copies of you will be created.

t=0: The lottery i... (read more)

1Chris_Leong6y
So, I (as casebash) originally wanted to bite the first horn of the dilemma, then I wanted to bite different horns depending on assumptions. Suppose we embrace physical continuity. Then all we've done by create copies is manage the news. Your original is just as likely to be in any state, but you now no longer know if you are the original or a clone. But even putting this aside, the ability to merge copies seems to contradict the idea of additional copies being given extra weight. If we embrace this idea, it seems more accurate to say that we can only add or delete copies. On the other hand suppose we reject it. What would psychological continuity mean? Imagine a physical agent A and a clone C which undergo identical experiences over 10 seconds. Now we use A and C as our psychologically continuous agents, but there's no reason why we couldn't construct an psychologically continuous agent that is A during the even numbered seconds and C during the odd numbered seconds. In fact, there are a ridiculous number of overlapping, psychologically continuous agents, all of which ought to be equally valid. This seems absurd, so I'd go so far as to suggest that if we reject physical continuity, we ought to reject the notion of continuity altogether. This would led us to take the third horn.

1) the probability of ending up in the set of winners is

1/billion

2) the probability of being (a specific) one of the trillion is

1/(b * t)

the probability of being a 2) given you are awake is

p(2 | awake) = P(awake | 2) * p(2)

                    p(awake)


            = 1 * 1E-21
              ---------
                   1



            = very small

Buy a ticket. Suspend your computer program just before the lottery drawing - which should of course be a quantum lottery, so that every ticket wins somewhere. Program your computational environment to, if you win, make a trillion copies of yourself, and wake them up for ten seconds, long enough to experience winning the lottery. Then suspend the programs, merge them again, and start the result. If you don't win the lottery, then just wake up automatically.

How would you do that???

0wedrifid12y
Are you asking for a solution to the engineering problem of how to convert yourself into an Em? I can't help there. Once you have that, this part with the lottery seems simple. The 'merge them again' would be tricky both on the philosophy and the engineering side (perhaps harder than converting to an Em in the first place.)

The odds of winning the lottery are ordinarily a billion to one. But now the branch in which you win has your "measure", your "amount of experience", temporarily multiplied by a trillion.

This seems perhaps too obvious, but how can branches multiply probability by anything greater than 1? Conditional branches follow the rules of conjunctive probability . . .

Probability in regards to the future is simply a matter of counting branches. The subset of branches in which you win is always only one in a billion of all branches - and any f... (read more)

The flaw is that anticipation should not be treated as a brute thing. Anticipation should be a tool used in the service of your decision theory. Once you bring in some particular decision theory and utility function, the question is dissolved (if you use TDT and your utility function is just the total quality of simulated observer moments, then you can reverse engineer exactly Nick Bostrom's notion of "anticipate." So if I had to go with an answer, that would be mine.)

Two people disagreeing about what they should anticipate is like two people arguing about whether a tree falling in an empty forest makes a sound. They disagree about what they anticipate, yes, but they behave identically.

2Vladimir_Nesov13y
Do that. Isn't as straightforward as it perhaps looks, I still have no idea how to approach the problem of anticipation. (Also, "total quality of simulated observer moments"?)
-2paulfchristiano13y
Do you mean try to reverse engineer a notion of anticipation, or try to dissolve the question? For the first, I mean to define anticipation in terms of what wagers you would make. In this case, how you treat a wager depends on whether having a simulation win the wager causes something good to happen to your utility function in one simulated copy, or in a million of them. Is that fair enough? I don't see why we care about anticipation at all, except as it bears on our decision making. I don't really understand how the second question is difficult. Whatever strategy you choose, you can predict exactly what will happen. So as long as you can compare the outcomes, you know what you should do. If you care about the number of simulated paperclips that are ever created, then you should take an even paperclip bet on whether you won the lottery if the paperclips would be created before the extra simulations are destroyed. Otherwise, you shouldn't. How do you describe a utility function that cares twice as much what happens to a consciousness which is being simulated twice?

I'm curious why no one mentioned Solomonoff prior here. Anticipation of subjective experience can be expressed as: what is a probability of experiencing X, given my prior experiences. Thus we "swap" ontological status of objective reality and subjective experience, then we can use Solomonoff prior to infer probabilities.

When one wakes up as a copy, one experiences instantaneous arbitrary space-time travel, thus Solomonoff prior for this experience should be lower, than that of wake-up-as-original one (if original one can wake up at all).

Given that approach, it seems that our subjective experience will tend to be as much "normal" as it allowed by simplest computable laws of physics.

0red7514y
It seems I've given too little information to make it worth thinking on it. Here's detailed explanation. I'll abbreviate thread of subjective experience as TSE. 1. If I make 10^6 copies of myself, then all 10^6+1 continuations of TSE are indistinguishable to external observer. Thus all these continuations are invariant under change of TSE, and it seems that we can assign equal probability to them. Yes, we can, but: 2. If TSE is not ontologically fundamental, then it is not bound by spacetime, laws of physics, universe, Everett multiverse, etc. There will be no logical contradiction, if you will find youself next instant as Boltzmann brain, or in one of infinitely many universes of level 4 multiverse, or outside your own lightcone. Thus: 3. Every finite set of continuations of TSE has zero probability. And finally: 4. We have no options, but Solomonoff prior to infer what we will experience next.

a truly remarkable observation: quantum measure seems to behave in a way that would avoid this trilemma completely

Which is why Roger Penrose is so keen to show that consciousness is a quantum phenomenon.

We have a strong subjective sense of personal experience which is optimized for passing on genes, and which thus coincides with the Born probabilities. In addition, it seems biased toward "only one of me" thinking (evidence: most people's intuitive rejection of MWI as absurd even before hearing any of the physics, and most people's intuitive sense that if duplicated, 'they' will be the original and 'someone else' will be the copy). The plausible ev-psych explanation for this, ISTM, is that you won't ever encounter another version of your actual... (read more)

0Johnicholas14y
Suppose that the Ebborians gamble. What odds would it give for event X? Suppose ve gives odds of 2:1 (probability of 2/3). A bookie takes the bet, and in half of the branches, collects 2 (from the two "descendants"), and in half of the branches, pays out 1, for an average profit of 0.5. I think your argument leads to the Ebborians being vulnerable to Dutch books.
1orthonormal14y
Er, your math is the wrong way around, but your point at first seems right: the Ebborian sees 2/3 odds, so ve is willing to pay the bookie 2 if X doesn't happen, and get paid 1 (split between copies, as in correlated decision theory) if X does happen. However, if instead the Ebborian insists on paying 2 for X not happening, but on each copy receiving 1 if X happens, the Dutch book goes away. Are there any inconsistencies that could arise from this sort of policy? Perhaps the (thus developed) correlated decision theory only works for the human form of subjective probability? Or more probably, I'm missing something.
0Johnicholas14y
From the bookie's perspective, the "each copy" deal corresponds to 1:1 odds, right?

The third horn is a fake, being capable of being defined in or out of existence at will. It I am indifferent to my state ten seconds from now, it is true; if my current utility function includes a term for my state ten seconds from now, it is false.

The 'thread of subjective experience' is not the issue; whether I throw myself of the cliff, will depend on whether I am currently indifferent as to whether my future self will die.

0Tyrrell_McAllister14y
I don't follow you. You write What do you mean by calling the horn "fake"? What is the "it" that is true or false?

Why is bullet three a fake reduction? I've bit it before and even see bullet three as something to aspire to.

[-][anonymous]14y00

"I still have trouble biting that bullet for some reason.... I don't think that, if I were really selfish, I could jump off a cliff knowing smugly that a different person would experience the consequence of hitting the ground."

Replace "jump off a cliff" with "heroin overdose." Sure you could, and many do. Not caring about the future is actually very common in humans, and it's harder for smart people to understand this attitude because many of us have very good impulse control. But I still find it strange that you seem to want ... (read more)

Is there a contradiction in supposing that the total subjective weight increases as unconnected threads of subjective experience come into existence, but copies branching off of an existing thread of subjective experience divide up the weight of the parent thread?

For what it's worth, here is the latest attempt by philosophers of physics to prove the Born rule from decision theory.

0rwallace14y
Interesting paper, but from skimming it without grokking all the mathematics, it looks to me like it doesn't quite prove the Born rule from decision theory, only proves that given the empirical existence of the Born rule in our universe, a rational agent should abide by it. Am I understanding the paper correctly?
0steven046114y
My understanding is the proof doesn't use empirical frequencies -- though if we observed different frequencies, we'd have to start doubting either QM or the proof. The question is just whether the proof's assumptions are true rationality constraints or "wouldn't it be convenient if" constraints. Everett and Evidence is another highly relevant paper.
0Johnicholas14y
I think the paper starts from the empirical existence of Born rule "weights" and attempts to explain in what sense they should be treated, decision-theoretically, as classical probabilities (since in the MWI sense, everything that might happen does happen) - but I admit I didn't grok the mathematics either.
[-][anonymous]14y00

Sentences in this comment asserted with only 95% probability of making sense, read on at own peril.

There's a mainstream program to derive the Born probabilities from physics and decision theory which David Wallace, especially, has done a lot of work on. If I remember correctly, he distinguishes two viewpoints:

  • "Subjective Uncertainty", which says you're a stage of some 4D space-time worm and you're indexically uncertain which worm, because many of them have stages that are exactly alike
  • "Objective Determinism", which says you'll contin
... (read more)
[-][anonymous]14y00

In quantum copying and merging, every "branch" operation preserves the total measure of the original branch,

Maybe branch quantum operations don't make new copies, but represent already existing but identical copies "becoming" no longer identical?

In the computer program analogy: instead of having one program at time t and n slightly different versions at time t+1, start out with n copies already existing (but identical) at time t, and have each one change in the branching. If you expect a t+2, you need to start with at least n^2 copie... (read more)

[-][anonymous]14y00

In quantum copying and merging, every "branch" operation preserves the total measure of the original branch,

Maybe branch quantum operations don't make new copies, but represent already existing but identical copies "becoming" no longer identical?

In the computer program analogy: instead of having one program at time t and n slightly different versions at time t+1, start out with n copies already existing (but identical) at time t, and have each one change in the branching. If you expect a t+2, you need to start with at least n^2 copie... (read more)

[-][anonymous]14y00

Just an aside - this is obviously something that Eliezer - someone highly intelligent and thoughful - has thought deeply about, and has had difficulty answering.

Yet most of the answers - including my own - seem to be of the "this is the obvious solution to the dilemma" sort.

Since I have a theory of Correlated decision making, let's use it! :-)

Let's look longer at the Nick Bostrom solution. How much contribution is there towards "feeling I will have won the lottery ten seconds from now" from "feeling I have currently won the lottery"? By the rules of this set-up, each of the happy copies contributes one trillionth towards that result.

(quick and dirty argument to convince you of that: replace the current rules by one saying "we will take the average feeling of victory across the trillion copies"; s... (read more)

0Eliezer Yudkowsky14y
So... what does it feel like to be merged into a trillion exact copies of yourself? Answer: it feels like nothing, because you couldn't detect the event happening. So in terms of what I expect to see happen next... if I've seen myself win the lottery, then in 10 seconds, I expect to still see evidence that I won the lottery. Even if, for some reason, I care about it less, that is still what I see... no?
1Stuart_Armstrong14y
See my other reformulations; here there is no "feeling of victory", but instead, you have scenarios where all but one of the trillions is spared, the others are killed. Then your expecations - if you didn't know what the other trillion had been told or shown - is that there is 1/trillion chance that the you in 10 seconds will still remember evidence that he has won the lottery. You can only say the you in 10 seconds will remember winning the lottery with certainty because you know that all the other copies also remember winning the lottery. Their contributions bumps it up to unity.

I've thought about this before and I think I'd have to take the second horn. Argument: assuming we can ignore quantum effects for the moment, imagine setting up a computer running one instance of some mind. There're no other instances anywhere. Shut the machine down. Assuming no dust theory style immortality (which, if there was such a thing, would seem to violate born stats, and given that we actually observe the validity of born stats...), total weight/measure/reality-fluid assigned to that mind goes from 1 to 0, so it looks reasonable that second horn t... (read more)

I have an Othello/Reversi playing program.

I tried making it better by applying probabilistic statistics to the game tree, quite like antropic reasoning. It then became quite bad at playing.

Ordinary minimax with A-B did very well.

Game algorithms that ignore density of states in the game tree, and only focus on minimaxing, do much better. This is a close analogy to the experience trees of Eliezer, and therefore a hint that antropic reasoning here has some kind of error.

Kim0

2rwallace14y
That's because those games are nonrandom, and your opponent can be expected to single out the best move. Algorithms for games like backgammon and poker that have a random element, do pay attention to density of states. (Oddly enough, so nowadays do the best known algorithms for Go, which surprised almost everyone in the field when this discovery was made. Intuitively, this can be seen as being because the game tree of Go is too large and complex for exhaustive search to work.)

Re: "But if I make two copies of the same computer program, is there twice as much experience, or only the same experience? Does someone who runs redundantly on three processors, get three times as much weight as someone who runs on one processor?"

Do they get three times as much "weight" - in some moral system?

Er, that depends on the moral system in question.

0Nominull14y
We're not talking about morality here, we're talking anthropics. http://j.photos.cx/AnthropicPrinciple-135.jpg
0timtyler14y
Oh. I see. Thanks! Put it this way: I would pay 100$ for a 101$ return if I didn't win the lottery.