Irrationality Game III

by CellBioGuy3 min read12th Mar 2014209 comments

15

Personal Blog

The 'Irrationality Game' posts in discussion came before my time here, but I had a very good time reading the bits written in the comments section.  I also had a number of thoughts I would've liked to post and get feedback on, but I knew that being buried in such old threads not much would come of it.  So I asked around and feedback from people has suggested that they would be open to a reboot!

I hereby again quote the original rules:

Please read the post before voting on the comments, as this is a game where voting works differently.

Warning: the comments section of this post will look odd. The most reasonable comments will have lots of negative karma. Do not be alarmed, it's all part of the plan. In order to participate in this game you should disable any viewing threshold for negatively voted comments.

Here's an irrationalist game meant to quickly collect a pool of controversial ideas for people to debate and assess. It kinda relies on people being honest and not being nitpickers, but it might be fun.

Write a comment reply to this post describing a belief you think has a reasonable chance of being true relative to the the beliefs of other Less Wrong folk. Jot down a proposition and a rough probability estimate or qualitative description, like 'fairly confident'.

Example (not my true belief): "The U.S. government was directly responsible for financing the September 11th terrorist attacks. Very confident. (~95%)."

If you post a belief, you have to vote on the beliefs of all other comments. Voting works like this: if you basically agree with the comment, vote the comment down. If you basically disagree with the comment, vote the comment up. What 'basically' means here is intuitive; instead of using a precise mathy scoring system, just make a guess. In my view, if their stated probability is 99.9% and your degree of belief is 90%, that merits an upvote: it's a pretty big difference of opinion. If they're at 99.9% and you're at 99.5%, it could go either way. If you're genuinely unsure whether or not you basically agree with them, you can pass on voting (but try not to). Vote up if you think they are either overconfident or underconfident in their belief: any disagreement is valid disagreement.

That's the spirit of the game, but some more qualifications and rules follow.

If the proposition in a comment isn't incredibly precise, use your best interpretation. If you really have to pick nits for whatever reason, say so in a comment reply.

The more upvotes you get, the more irrational Less Wrong perceives your belief to be. Which means that if you have a large amount of Less Wrong karma and can still get lots of upvotes on your crazy beliefs then you will get lots of smart people to take your weird ideas a little more seriously.

Some poor soul is going to come along and post "I believe in God". Don't pick nits and say "Well in a a Tegmark multiverse there is definitely a universe exactly like ours where some sort of god rules over us..." and downvote it. That's cheating. You better upvote the guy. For just this post, get over your desire to upvote rationality. For this game, we reward perceived irrationality.

Try to be precise in your propositions. Saying "I believe in God. 99% sure." isn't informative because we don't quite know which God you're talking about. A deist god? The Christian God? Jewish?

Y'all know this already, but just a reminder: preferences ain't beliefs. Downvote preferences disguised as beliefs. Beliefs that include the word "should" are are almost always imprecise: avoid them.

That means our local theists are probably gonna get a lot of upvotes. Can you beat them with your confident but perceived-by-LW-as-irrational beliefs? It's a challenge!

Additional rules:

  • Generally, no repeating an altered version of a proposition already in the comments unless it's different in an interesting and important way. Use your judgement.
  • If you have comments about the game, please reply to my comment below about meta discussion, not to the post itself. Only propositions to be judged for the game should be direct comments to this post. 
  • Don't post propositions as comment replies to other comments. That'll make it disorganized.
  • You have to actually think your degree of belief is rational.  You should already have taken the fact that most people would disagree with you into account and updated on that information. That means that  any proposition you make is a proposition that you think you are personally more rational about than the Less Wrong average.  This could be good or bad. Lots of upvotes means lots of people disagree with you. That's generally bad. Lots of downvotes means you're probably right. That's good, but this is a game where perceived irrationality wins you karma. The game is only fun if you're trying to be completely honest in your stated beliefs. Don't post something crazy and expect to get karma. Don't exaggerate your beliefs. Play fair.
  • Debate and discussion is great, but keep it civil.  Linking to the Sequences is barely civil -- summarize arguments from specific LW posts and maybe link, but don't tell someone to go read something. If someone says they believe in God with 100% probability and you don't want to take the time to give a brief but substantive counterargument, don't comment at all. We're inviting people to share beliefs we think are irrational; don't be mean about their responses.
  • No propositions that people are unlikely to have an opinion about, like "Yesterday I wore black socks. ~80%" or "Antipope Christopher would have been a good leader in his latter days had he not been dethroned by Pope Sergius III. ~30%." The goal is to be controversial and interesting.
  • Multiple propositions are fine, so long as they're moderately interesting.
  • You are encouraged to reply to comments with your own probability estimates, but  comment voting works normally for comment replies to other comments.  That is, upvote for good discussion, not agreement or disagreement.
  • In general, just keep within the spirit of the game: we're celebrating LW-contrarian beliefs for a change!

I would suggest placing *related* propositions in the same comment, but wildly different ones might deserve separate comments for keeping threads separate.

Make sure you put "Irrationality Game" as the first two words of a post containing a proposition to be voted upon in the game's format.

Here we go!

EDIT:  It was pointed out in the meta-thread below that this could be done with polls rather than karma so as to discourage playing-to-win and getting around the hiding of downvoted comments.  If anyone resurrects this game in the future, please do so under that system  If you wish to test a poll format in this thread feel free to do so, but continue voting as normal for those that are not in poll format. 

15

209 comments, sorted by Highlighting new comments since Today at 11:36 AM
New Comment
Some comments are truncated due to high volume. (⌘F to expand all)Change truncation settings

Irrationality Game: Less Wrong is simply my Tyler Durden—a disassociated digital personality concocted by my unconcious mind to be everything I need it to be to cope with Camusian absurdist reality. 95%.

4Dallas7yI am very curious as to what your evidence for backing up this proposition is or would be.
2[anonymous]7yIf you're right, your unconscious mind is awesome.
1ThisSpaceAvailable7ySo, I'm supposed to upvote this unless I believe that I am a figment of your imagination? This seems a bit cheaty. Maybe I should post "No one other than ThisSpaceAvailable believes that this statement is true. 99.99%"
1kokotajlod7yAre we supposed to judge the rationality of this statement based on how rational we think it is for you to believe it, or based on how rational we think it is for us to believe it?
1gothgirl4206667yWow. I think this one might win.

Irrationality Game: (meta, I like this idea)

Flush toilets are a horrible mistake. 7b/99%

4Slackson7y"If it's yellow let it mellow, if it's brown flush it down." This is one of the first things I remember learning, growing up with tank water.
4ChristianKl7yBased on what reasoning?
2faul_sname7yUpvoted for "horrible". I don't see how their impact is all that bad -- at 3.5 GPF (which is standard), that means that, for example, all of the flush toilets in California together use about 750,000 acre-feet of water per year. Compared to the 34 million acre feet used in the same state for agriculture, it's clear that flush toilets use a significant but still pretty small fraction of the water in the state, but "horrible" is an overstatement. (I choose California because it is a populous state that regularly has water shortages).
2trist7yI admit to hyperbole, now, with a little more thought, I would have worded it differently. Both to clarify that it's pretty far down on our list of societal problems, and that it's more an individual level mistake rather than a systematic one (though there are systematic benefits to fewer flush toilets).
2rthomas67yHow is this not just a preference?
5trist7yI suppose my actual belief is that flush toilets are a mistake outside of urban areas, I don't have much experience with urban living or what other poop strategies could work with it. Advantages, flush toilets: * Provide easy long distance transport of human waste in urban environments. * Exchanges weekly-to-yearly chores for purchased services. Disadvantages, flush toilets: * Create additional dependency on water (and by extension outside water districts, electricity). * Turn (vast amounts of) drinking water into black water. * Create a waste product from human manure, which is a valuable resource (fertile soil) when dealt with properly. * Adds significantly to the cost of housing (especially outside sewer districts).
5Error7yI still think you're overconfident, so upvoting, but the justification is convincing enough to make me update from near-zero to something noticeably above zero. I never thought of it quite that way.
4trist7yTranslate it to "In x% of new non-urban houses, there are options better than flush toilets." My confidence in my confidence assignment isn't very high yet though, so I am quite open to being overconfident. And obviously both lists are non-exhaustive. * Flush toilets handle large numbers of people for a long time fairly easily. * Flush toilets get clogged.
0Lumifer7yYou forgot the "not having the whole village die from a cholera outbreak" part :-/
3CellBioGuy7yThe running water revolution came around the same time as the sanitation revolution. I'm not 100% sure you necessarily HAVE to have one to have the benefits of the other, though it helps. Modern composting toilets and hot-composting of human manure is quite safe if done properly. Flush toilets definitely get the sanitation thing done, but perhaps rather than 'mistake' we could call them an 'inefficient first draft that still works well'.

is quite safe if done properly

That's the thing -- it's basically an issue of idiot-proofing. Many things are "safe if done properly" and still are not a good idea because people in general are pretty bad at doing things properly.

Flush toilets are idiot-proof to a remarkable degree. Composting human manure, I have my doubts.

0[anonymous]7yAlmost all of them are, but I've seen a couple of them which are very easy to accidentally flush improperly in such a way that water will keep on running until someone notices and fixes it.
0rthomas67yThat clears things up a lot, and I changed my downvote to an upvote. EDIT: To be clear, I disagree with you. My thoughts on your disadvantages list: * Flush toilets do create additional dependency on water, however if one already has running water and depends on it for drinking and washing, how significant is the additional water dependency for flush toilets? * The reason flush toilets use potable water is an economic one. It is simply cheaper to use one unified water system instead of two, when someplace already has running water. The cost of the wasted drinking water is negligible compared to the cost of building a second plumbing system. * This point is the most interesting to me. I have no information on the usefulness of human manure, and would be interested to know if human manure would have a comparable market value to cattle manure or synthetic fertilizer. I am skeptical because of the tendency for human waste to carry human diseases. * I have no disagreements with this disadvantage, but simply feel that the vast, vast majority of people would be willing to pay for the extra cost in housing if they already had indoor plumbing.
2NancyLebovitz7yI expect this is too expensive to be worth it. but instead of a whole second water system, it's theoretically possible to use gray water from bathing and showering for flushing. On second thought, this might actually make sense for apartment buildings and hotels, since some gray water could be stored and sent downhill for flushing-- you wouldn't need a pump in the bathroom.
1roystgnr7yAustin's "Dillo Dirt" [https://en.wikipedia.org/wiki/Dillo_Dirt] is made from yard waste and treated human sewage. Less-treated sewage gets used to fertilize ranchland. As you suspected, there's more than a little controversy over whether the result is well-composted enough for health and aesthetics [http://www.thedailyswarm.com/headlines/austin-shitty-limits/], but it's mixed up with concern over the standards for various non-fecal pollutants. Presumably whatever closed loop fertilization trist is advocating wouldn't have to worry so much about the various kinds of industrial and medical waste people dump down their drains.
0gothgirl4206667yI think you're playing it wrong? You upvote if you disagree.
0rthomas67yI do disagree. Did you read the rest of my comment? I originally downvoted because the rules also say to downvote if someone expresses a preference disguised as a belief.

Irrationality game: Every thing which exists has subjective experience (80%). This includes things such as animals, plants, rocks, ideas, mathematics, the universe and any sub component of an aforementioned system.

5khafra7yBy "any subcomponent," do you mean that the powerset of the universe is composed of conscious entities, even when light speed and expansion preclude causal interaction within the conscious entity? Because, if the universe is indeed spatially infinite, that means that the set of conscious entities is the infinity of the continuum; and I'm really confused by what that does to anthropic reasoning.
2NoSuchPlace7yIf you replace consciousness with subjective experience I believe your statement is correct. Also once you have one infinity you can take power sets again and again. As far as I understand it breaks anthropic reasoning because now your event space is to big to be able to define a probability measure [http://en.wikipedia.org/wiki/Event_%28probability_theory%29#Events_in_probability_spaces] . For the time being I have concluded that anthropic reasoning doesn't work because of a very similar argument though I will revise my argument once I have learned the relevant math.
3shminux7yHow would one define subjective experience for rocks and atoms?
6NoSuchPlace7yDefining subjective experience is hard for the same reason that defining red is hard, since they are direct experiences. However in this case I can't get around this by pointing at examples. So the only thing I can do is offer an alternative phrasing which suffers from the same problem: If you accept that our experiences are what an algorithm feels like from on the inside [http://wiki.lesswrong.com/wiki/How_an_algorithm_feels] then I am saying that everything feels like something from the inside.
0ThisSpaceAvailable7yBesides the issue of "subjective experience" that has already been brought up, there's also the question of what "thing" and "exists" mean. Are abstract concepts "things"? Do virtual particles "exist"? By including ideas, you seem to be saying "yes" to the first question. So do subjective experiences have subjective experiences themselves? Also, it's "an aforementioned". That's especially important when speaking.
0NoSuchPlace7yI believe some form of MUH [http://en.wikipedia.org/wiki/Mathematical_universe_hypothesis] is correct so when I say exist I mean the same thing as in mathematics (in the sense of quantifying over various things). So by a thing I mean anything for which it is (at least in principle) possible to write down a mathematically precise definition. Presumably abstract ideas and virtual particles fall under this category though in neither case am I sure because I don't know what you mean by abstract idea/I don't know enough physics. I not sure whether it possible to give a definition for subjective experience so I don't know whether subjective experiences have subjective experiences. Substituted an a for an an.

Irrationality game:

Most posthuman societies will have violent death rate much higher than humans ever had. Most poshumans who will ever live will die at wars. 95%

Interesting. So, you have Robin Hanson's belief that we won't get a strong singleton; but you lack his belief that emulated minds will be able to evaluate each other's abilities with enough confidence that trade (taking into account the expected value of fighting) will be superior to fighting? That's quite the idiosyncratic position, especially for 95% confidence.

Not a bad hypothesis, but your confidence level is too high... hence my upvote.

0[anonymous]7ySame here. I'd give it around 20%.
5NancyLebovitz7yAre you imagining that outcome because you expect resource shortages? Peaceful lives are just too boring? Posthumans are generally too alien to each other for stable cooperation? Now that I think about it, I find the last alternative pretty plausible.
4Nornagest7ySeems fairly reasonable on its face, actually; once you've gotten rid of disease and age, what's left is accidents, violence, and suicide if you're counting that separately from violence. Upvoted, though, because I think you're undercounting accidents (more Americans already die in automotive accidents than die violently, by a large margin; I'd expect the same is true for the rest of the First World but haven't seen statistics) and making too strong a statement about the structure of posthuman society.
0aausch5yi think the concept of death is extremely poorly defined under most variations of posthuman societies; death as we interpret it today depends on a number of concepts that are very likely to break down or be irrelevant in a post-human-verse -------------------------------------------------------------------------------- take, for example, the interpretation of death as the permanent end to a continuous distinct identity: if i create several thousand partially conscious partial clones of myself to complete a task (say, build a rocketship), and then reabsorb and compress their experiences, have those partial clones died? if i lose 99.5% of my physical incarnations and 50% of my processing power to an accident, did any of the individual incarnations die? have i died? what if some other consciousness absorbs them (with or without my, or the clones', permission or awareness)? what if i become infected with a meme which permanently alters my behavior? my identity?

You (the reader) do not exist.

EDIT: That was too punchy and not precise. The reasoning behind the statement:

Most things which think they are me are horribly confused gasps of consciousness. Rational agents should believe the chances are small that their experiences are remotely genuine.

EDIT 2: After thinking about shminux's comment, I have to retract my original statement about you readers not existing. Even if I'm a hopelessly confused Boltzmann brain, the referent "you" might still well exist. At minimum I have to think about existence more. Sorry!

4[anonymous]7yCogito, ergo upvoto. :-)
3[anonymous]7y.
2solipsist7yTo quote the adage, I'm a solipsist, and am surprised everyone else isn't too. I think any intelligent agent should conclude that it is probably something akin to a Boltzmann brain [http://en.wikipedia.org/wiki/Boltzmann_brain]. You could plausibly argue that I am cheating with pronoun references (other people might agree with the solipsistic logic, but centered around them). Is that what you are asking? EDIT Not really. I think some of the problems with AIXI may be AIXI acting rationally where the desired behavior is irrational, but that's the only time I can think of it coming up outside of a philosophy discussion.
2[anonymous]7yI often use the concept of Boltzmann brain to relax or fall asleep. Thinking that this is the only moment you will ever get to feel alive and you will die a few moments from now is a good way to put your mind "in the now". That said, if it actually were true I would expect the reality I perceive to be radically different. Almost everything I know about the outside world is really consistent and ordered and everything I've ever experienced supports the mainstream physical model of the universe. I don't think there would be an entire history of the universe and Earth and such, which I'm able to confirm relatively well by going into the museum and and considering the evidence, if this were just a random fluke. I would expect many things to be far more incoherent. I still think there's a low chance it's true. Not really low chance, the chance is probably higher than the chance that I will win in a lottery or that biblical God exists. And this belief doesn't have much decision theoretic importance so I would probably ignore it even if I knew for sure that it's true. Btw, how do you resolve the paradox that you can't trust your own senses and reasoning? This game assumes that users actually are real people because otherwise asking about their opinions would be pointless. But now that you explained it I decided to change my downvote to upvote because I think the probability of this being true is low.
2maia7yCould you be more specific about what you mean by that?
2solipsist7yOf all possible minds thinking the thought that I am thinking right now, most aren't living on earth, posting to Less Wrong. Most are random fluctuations in high-entropy soup [http://en.wikipedia.org/wiki/Boltzmann_brain] after the heat-death of the universe, or bizarre minds capable of belief but not thought, or other deluded or confused agents. In all but a negligible fraction of these, you, maia, do not exist.
0rthomas67yWhat degree of certainty do you place on that belief?
0solipsist7yI could put numbers to it, but it would really be pulling them out of my butt -- how certain are you that anthropic reasoning is valid? If it is valid (which seems more likely than not), then you quickly run into the problem of Boltzmann brains. Some people try to exorcise Boltzmann brains from their anthropic viewpoint, but I have no problem with biting the bullet that most brains are Boltzmann brains. The practical implications of that belief, assuming the world is as it appears to be, are (I believe) minimal.
-5gothgirl4206667y

Irrationality Game: I am something ontologically distinct from my body; I am much simpler and I am not located in the same spacetime. 50%

EDIT: Upon further reflection, my probability assignment would be better represented as the range between 30% and 50%, after factoring in general uncertainty due to confusion. I doubt this will make a difference to the voting though. ;)

2NancyLebovitz7yWhy would you be much simpler if you were ontologically distinct?
0kokotajlod7yThe ";" was meant to be simply a "and also" rather than a "therefore." I think that I'm much simpler than my body, and that is one of the reasons why I think I'm ontologically distinct. With 50% probability. EDIT: Another answer I endorse: If I'm ontologically distinct from my body, then who knows how complicated I am--but apply Occam's Razor, such that I'm only as complicated as I need to be, and the result will be that I'm much simpler than my body. Anyone who believes that uploading would preserve consciousness should agree with this, since uploading (can) change the medium of computation to a simpler one.
1Eugine_Nier7yWhy is this a 19? I thought this was a restatement of the "official LW position". Or would people argue that an uploaded kokotajlod wouldn't be the real kokotajlod?
3Randy_M7yI guess if you read it loosely. I think the official LW position would be (correct me if I am wrong) an em of kokotajlod that has high enough fidelity to replicate his decision making process is him; what he is is a particular set of hueristics, instincts, etc, that accompany his body but could theoretically exist outside it. That does match his statement if one reads it as refering to something more like a platonic concept than a spiritual essence.
1kokotajlod7yI think the "official LW position" is more reductionist than my Irrationality Game statement. Even if people think that I am a computation, they probably don't think I'm some sort of Platonic Form, but rather that I'm just a certain type of physical object that "implements" a computation. That's my understanding of typical LW thought, at any rate. Which is why I chose the statement that I did. :)
1khafra7yWhat does "spacetime" mean? Is the real "you" neither a causal descendent, nor a causal ancestor, of any of your body's actions? I'd have to put that down somewhere around argmin probability. Or do you just mean that you consider the real you to be something like a platonic computation, which your material body instantiates? That's not too far off from some realms of LW semi-orthodoxy.
1kokotajlod7yGood questions. I'll explain my reasoning: Basically, after thinking about consciousness for a while, and personal identity, I've come to assign high probability to some sort of dualism/idealism being true. It might still be a sort of reductionist dualism, i.e. platonic computations. So yes, the "platonic computation" theory would count. Do you think my original post ought to be revised given this information? I hope I haven't been misleading. As for spacetime and causation: If I'm a platonic form, I'm not in spacetime, nor am I causally related to my body in any normal sense. It all depends on how we define causation, and I tend to be reductionist/eliminativist about causation.
0khafra7yI don't think you've been any more misleading than a dualist is pretty much required to be. The basic ambiguities of dualism do, of course, remain: 1. How does the non-spacetime stuff produce subjective experience, when spacetime stuff can't? 2. How does your subjective experience correlate with the environment and actions of your material body, just as if there were two-way causation going on? (even when you reduce causation to a Pearl-style net, or to the large-scale behavior of many individually time-reversible components, this question remains).
2kokotajlod7y(1) It's not about producing subjective experience, it is about being subjective experience. The idea here is that massive, vague mereological fusions of subatomic stuff just aren't the sort of thing that can be subjective experiences. Just as no fundamental particle can be a chariot, since chariots have parts. (2) I have no idea yet. I'm considering some sort of interactionist dualism, or some sort of idealism, or some sort of mathematical multiverse theory with self-contained mathematical structures that play the role of Platonic computations, with measure assigned by a simplicity weighting that generates the appearance of a physical world. And of course I'm considering reductionist physicalism, reductionist mathematical multiverse theory, etc. as well. That's where the other 50% comes in.
0[anonymous]7yTaboo “I”. For all the ways of interpreting that claim that I can come up with, I'd give a probability either much, much higher or much, much lower than 50%.
0kokotajlod7yCould you list the ways? I'm interested to hear which ways you think would give a probability much higher than 50%. Also, telling me to taboo "I" is telling me to give a successful analysis of consciousness; if I could already do that, I wouldn't assign 50% probability to it being one thing, and 50% probability to it being another.
0[anonymous]7yI mean, sometimes by the word “I” I mean my System 2 (as in “I'm not sure what the effect of stricter gun control on the homicide rate would be”), sometimes I mean my System 1 (as in “I'm not scared of spiders”), sometimes I mean my body (as in “I've got a backache”), sometimes I mean my car (as in “I'm parked over there”), etc. Which one did you mean there? ...Oh. Now I can see a reasonable non-tautological, non-tautologically false interpretation of your entry.

Irrationality game - there is a provident, superior entity that is in no way infinite (I wonder if people here would call that God. As a "superman theist" I had to put "odds of God (as defined in question)" at 5% but identify as strongly theist in the last census)

Edit: forgot odds. 80%

3NoSuchPlace7yCould you explain your reasoning? I'm very curious about this.
5Ixiel7yI was brought up Catholic, and quickly decided religion (later updated to human scribes millennia ago and blind faith therein) didn't really understand the difference between "bigger than I can understand" and "infinite." I also have a life so cartoonishly awesome (let me know if you have a solution to this, but I honestly believe if I laid down the facts people would think I'm lying), I figured what I called God not only exists but likes me more than everybody else. As I grew up, I "tested" the theory a few times, but never with any scientific rigor, and I think I'd have to call the results positive but not statistically significant. I have no problem assuming no god at the beginning of a discussion, and if I had strong enough evidence I'd like to think I'd admit I'm wrong. I also don't correlate anything about God with misunderstanding what "death" means - or as many Catholics call it life after death. I know it's a minority view here and would never trot it out in normal discourse, but it seemed appropriate for the venue.
3RowanE7yI'm intruiged - a life so awesome that it's implausible for a member of an internet forum with thousands of members (especially this one, which is dedicated to the "science of winning at life" and has an average IQ so high it should be impossible) to actually be living it strikes me as a high bar, and additionally the idea of it being so awesome that it's (at least subjectively) convincing evidence for the hypothesis "god is real and he loves me". The only thing that comes to my mind that looks like it meets that criteria and isn't blatantly supernatural would be winning the lottery, but since lottery winners are often less happy in the end I don't expect that's it. I can't promise I'll believe you, but I'd upvote just for sating my curiosity.
5Ixiel7yNah, not implausible I exist, but I rarely post, so have no track record. It's amazing how many people are above average online... Mentally, I'm materially above average intelligence, but understand that that only goes so far. And I cultivate rationality (I'm here aren't I?) Socially, I'm reasonably well liked by everyone I know, people tell me I have a decent sense of humor. I'm engaged to a beautiful blonde doctor, who is eerily similar to the woman I prayed to meet as a teenager, and has been able to put up with my strangeness for four years. Bodily, I have no known history of any genetic diseases and have never been dependant on prescription drugs. Though I admit the surgeon general would like me to lose a pound or two. Not "Mommy why's he like that?" fat though. Financially, I own my own house, and if I (and they) decide to have kids, my kids and grandkids will never have to work, assuming I don't earn/inherit/win/cetera a penny, and my stocks gain 0% (Admittedly, they could crash). I tell people I have a Forrest Gump approach. "Lt. Dan said I didn't have to worry about money any more. And I said, well, that's good. One less thing." Attitudinally, I'm hugely optimistic. Not every day, but more often than not, I wake up and am struck by the wonder of how unlikely my good fortune is. I know it sounds out there, and it is, but it's also true. Hand to God. Or Bacon, or whomever you'd like, if you dig propriomanual verification.
3RowanE7yThe only parts of that list that seem out-there are the fiancee-eerily-similar-to-prayers (alternate explanations: The kind of people you actively seek out and the kind of people you pray to meet are going to resemble each other, human memories are fallible to the point that they can be completely fabricated, so a vague similarity might adjust to become an eerily close similarity) and the financial status (I'm not sure what my prior for this is supposed to be, since "rich enough never to have to work for three generations" varies a lot on expenditure, probably more than just the top 1% could get there if they were frugal).
1Ixiel7yOh agreed. My awesome life is not a good proof, but while I came with a high IQ out of the box, I hadn't learned the tools of thinking yet to the necessary degree. It's loosely confirmatory, but not a silver bullet. I was just saying that prompted me to have the idea a decade or two before (inadequitely still, but I knew it) testing the idea. My confidence may be too high, but really, it hasn't been a priority to test mostly because I can't think of a good one that doesn't come at too high a cost for too little benefit. I've never really tried to prove my Fiancé's aunt who I've never met exists either. Open to ideas. And the grandkids thing is something I came up with to give my fiancé perspective when she was just my girlfriend. Anything with "illion" in it becomes "a bunch" to non math folk. Think about 4 mil. Now think about 8. I'd say (based on nothing but anecdote) if you said each to fifty men-on-the-street, you'd have at least 85 people thinking of pretty much the same pile of gold doubloons.
0Will_Newsome7yCommit to not talk to anyone about the results of your test. No hinting, nothing. If you do this the experiment is worthless. Don't even mention it; if asked about it, say 'I promised beforehand that I would not give out information'. Take a coin, flip it twenty times or so, record heads and tails. This is not a good test but it's simple and easy and can at least theoretically provide some information.
-1falenas1087yI agree with RomanE in that this doesn't seem all that unusual. I'm an undergrad in college right now so I don't have the monetary security, fiancee, or house, but everything else applies to me as well. There are a couple of things that could help to explain this, in my cause and probably in yours. Are you fairly neurotypical in a way that doesn't interfere with your social life or physical well-being? Did you grow up with a middle, upper middle, or upper class family? Do you grow up in a developed nation? Are you of a racial class that is generally privileged in your area? (E.g., white in America.) If these are true, or even just the first three, I would say it's not all that unusual. Not lower than 5%, anyway. I don't think that, given the above, it is unusual enough to warrant an explanation outside the ordinary.
0Ixiel7yYes, upper, yes, and yes. As stated in the other reply, I do not currently count it as strong evidence.
2Will_Newsome7yThat's impossible, He likes me more than everybody else. (.000001% confident) (Seriously though, I do believe there are god-like phenomena, and they seem to be suspiciously favorable toward me. (Excuse me while I knock on wood and stare plaintively at the fourth wall.))
0TsviBT7yI too am curious.
1gothgirl4206667yWhat are your odds on this?
0Ixiel7yOh thanks. Fixed.

The universe is finite, and not much bigger than the region we observe. There is no multiverse (in particular Many Worlds Interpretation is incorrect and SIA is incorrect). There have been a few (< million) intelligent civilisations before human beings but none of them managed to expand into space, which explains Fermi's paradox. This also implies a mild form of the "Doomsday" argument (we are fairly unlikely to expand ourselves) but not a strong future filter (if instead millions or billions of civilisations had existed, but none of them expanded, there would be a massive future filter). Probability: 90%.

1polymathwannabe7yI don't know how to vote on this. I have very strong suspicions that MWI is incorrect (its Copernican allure is its only favorable point), but I disagree that the universe is finite. I feel inclined toward SIA, but I generally reject anthropic reasoning (that's perhaps a statement about myself rather than about your arguments). (Also, I require more detailed arguments to dissolve Fermi's paradox because I don't believe paradoxes exist in reality.)
0shminux7yHow would you evaluate correctness of something untestable?
0polymathwannabe7yI don't know whether this counts as a correctness assessment, but my expectations do not vary with the trueness of MWI, so it's a needless hypothesis.
0drnickbone7yI'd suggest that since you agree with some parts but disagree with others, you assign probability a lot less than 90% to the whole hypothesis. So you should think I'm irrationally overconfident in the whole lot, and upvote please! If you want some detail, I start from the "Great Filter" argument (see http://hanson.gmu.edu/greatfilter.html [http://hanson.gmu.edu/greatfilter.html] ). I find it very hard to believe that there is a super-strong future filter ahead of us (such that we have < 1 in a million or < 1 in a billion chance of passing it and then expanding into space). But a relatively weak filter implies that rather few civilizations can have got to our stage of development - there can't have been millions or billions of them, or some would have got past the filter and expanded, and we would not expect to see the world as we do in fact see it. The argument to the universe being finite (and not too big) then follows from there being a limited number of civilizations. SIA and MWI must also be wrong, because they each imply a very large or infinite number of civilizations.
1Nisan7yYour conclusion doesn't follow from your premises. The lack of a strong filter implies that a not insignificant proportion of civilizations colonize space. This is consistent with there being a large universe containing many intergalactic civilizations we will never observe because of the expansion of the universe.
0drnickbone7yNo, in that large universe model we'd expect to be part of one of the expanded, intergalactic civilisations, and not part of a small, still-at-home civilisation. So, as I stated "we would not expect to see the world as we do in fact see it". Clearly we could still be part of a small civilisation (nothing logically impossible about being in a tiny minority), or we could be in some sort of zoo or ancestor simulation within a big civilisation. But that's not what we'd expect to see. You might want to see Ken Olum's paper for more on this: http://arxiv.org/abs/gr-qc/0303070 [http://arxiv.org/abs/gr-qc/0303070] Incidentally, Olum considers several different ways out of the conflict between expectation and observation: the finite universe is option F (page 5) and that option seems to me to be a lot more plausible than any of the alternatives he sketches. But if you disagree, please tell me which option you think more likely.
0ThisSpaceAvailable7yI find that sort of anthropic argument to Prove Too Much. For instance, our universe is about 14 billion years old, but many models have the universe existing trillions of years into the future. If the universe were to survive 280 billion years, then that would put us within the first 5% of the universe's lifespan. So, if we take an alpha of 5%, we can reject the hypothesis that the universe will last more than 280 billion years. We can also reject the hypothesis that more than 4 trillion humans lives will take place, that any given 1-year-old will reach the age of 20, that humans will have machines capable of flight for more than 2000 years, etc. Olum appears to be making a post hoc argument. The probability that the right sperm would fertilize the right egg and I would be conceived is much less than 1 in a billion, but that doesn't mean I think I need a new model. The probability of being born prior to a galactic-wide expansion may be very low, but someone has to be born before the expansion. What's so special about me, that I should reject the possibility that I such a person?
0drnickbone7yThat sounds like "Copernican" reasoning (assume you are at a random point in time) rather than "anthropic" reasoning (assume you are a random observer from a class of observers). I'm not surprised the Copernican approach gives daft results, because the spatial version (assume you are at a random point in space) also gives daft results: see here in this thread [http://lesswrong.com/lw/jvg/irrationality_game_iii/ap01] point 2. Incidentally, there is a valid anthropic version of your argument: the prediction is that the universe will be uninhabitable 280 billion years from now, or at least contain many fewer observers than it does now. However, in that case, it looks like a successful prediction. The recent discovery that the stars are beginning to go out [http://blogs.scientificamerican.com/life-unbounded/2012/11/19/the-stars-are-beginning-to-go-out/] and that 95% of stars that will ever form have formed already [http://www.wired.com/wiredscience/2012/11/universe-making-stars/] is just the sort of thing that would be expected under anthropic reasoning. But it is totally surprising otherwise. The correct application of anthropic reasoning only rejects this as a hypothesis about the average number of observers in a civilisation, not about human beings specifically. If we knew somehow (on other grounds) that most civilisations make it to 10 trillion observers, we wouldn't predict any less for human beings. That's an instance of the same error: anthropic reasoning does NOT reject the particular hypothesis. We already know that an average human lifespan is greater than 20, so we have no reason to predict less than 20 for a particular child. (The reason is that observing one particular child at age 1 as a random observation from the set of all human observations is no less probable if she lives to 100 than if she lives to 2). Anthropic reasoning is like any Bayesian reasoning: observations only count as evidence between hypotheses if they are more likely on one hypo
0drnickbone7yI got a bit distracted by the "anthropic reasoning is wrong" discussion below, and missed adding something important. The conclusion that "we would not expect to see the world as we in fact see it" holds in a big universe regardless of the approach taken to anthropic reasoning. It's worth spelling that out in some detail. 1. Suppose I don't want to engage in any form of anthropic reasoning or observation sampling hypothesis. Then the large universe model leaves me unable to predict anything much at all about my observations. I might perhaps be in a small civilisation, but then I might be in a simulation, or a Boltzmann Brain, or mad, or a galactic emperor, or a worm, or a rock, or a hydrogen molecule. I have no basis for assigning significant probability to any of these - my predictions are all over the place. So I certainly can't expect to observe that I'm an intelligent observer in a small civilisation confined to its home planet. 2. Suppose I adopt a "Copernican" hypothesis - I'm just at a random point in space. Well now, the usual big and small universe hypotheses predict that I'm most likely going to be somewhere in intergalactic or interstellar space, so that's not a great predictive success. The universe model which most predicts my observations looks frankly weird... instead of a lot of empty space, it is a dense mass of "computronium" running lots of simulations of different observers, and I'm one of them. Even then I can't expect to be in a simulation of a small civilisation, since the sim could be of just about anything. Again, not a great predictive success. 3. Suppose I adopt SIA reasoning. Then I should just ignore the finite universes, since they contribute zero prior probability. Or if I've decided for some reason to keep all my universe hypotheses finite, then I should ignore all but the largest ones (ones with 3^^^3 or more galaxies). Among t
0Squark7yThis sort of anthropic reasoning is wrong. Consider the following experiment. A fair coin is tossed. If the result is H, you are cloned into 10^10 copies, and all of those copies except one are placed in the Andromeda galaxy. Another copy remains in the Milky Way. If the result is T, no cloning occurs and you remain in the Milky Way. Either way, the "you" in Milky Way has no immediate direct way to know about the result of the coin toss. Someone, call her "anthropic mugger", comes to you an offers a bet. She can perform an experiment which will reveal the result of the coin toss (but she hasn't done it yet). If you accept the bet and the coin toss turns out to be H, she pays you 1$. If you accept the bet and the coin toss turns out to be T, you pay her 1000$. Do you accept the bet? Reasoning along the same lines as you did to conclude there are no large civilizations, you should accept the bet. But this means your expected gain before the coin toss is -499.5$. So, before the coin toss it is profitable for you to change your way of reasoning so you won't be tempted to accept the bet. There's no reason to accept the bet unless in the cloning scenario you care much less about the copy of you in Milky Way than in the no-cloning scenario. So, there's no reason to assume there are no large civilizations if the existence of large civilizations wouldn't make us care much less about our own.
0drnickbone7yThere are a number of problems with that: 1) You don't specify whether the bet is offered to all my copies or just to one of them, or if to just one of them, whether it is guaranteed to be the one in the Milky Way. Or if the one in the Milky Way knows he is in the Milky Way when taking the bet, and so on. Suppose I am offered the bet before knowing whether I am in Andromeda or Milky Way. What odds should I accept on the coin toss: 50/50? Suppose I am then told I am in the Milky Way... what odds should I now accept on the coin-toss: still 50/50? If you say 50/50 in both cases then you are a "double-halfer" (in the terminology of Sleeping Beauty problems) and you can be Dutch-booked. If you answer other than 50/50 in one case or the other, then you are saying there are circumstances where you'd bet at odds different (probably very different) from the physical odds of a fair coin toss and without any context, that sounds rather crazy. So whatever you say, there is a bullet to bite. 2) I am, by the way, quite aware of the literature on Anthropic Decision Theory (especially Stuart Armstrong's paper) and since my utility function is roughly the average utility for my future copies (rather than total utility) I feel inclined to bet with the SSA odds. Yes, this will lead to the "me" in the Milky Way making a loss in the case of "H" but at that stage he counts for only a tiny slither of my utility function, so I think I'll take the risk and eat the loss. If I modify my reasoning now then there are other bets which will lead to a bigger expected loss (or even a guaranteed loss if I can be Dutch-booked). Remember though that I only assigned 90% probability to the original hypothesis. Part of the remaining 10% uncertainty is that I am not fully confident that SSA odds are the right ones to use. So the anthropic mugger might not be able to make $500 off me (I'm likely to refuse the 1000:1 bet), but he probably could make $5 off me. 3) As in many such problems, you overs
0Squark7yI meant that the bet is offered to the copy in the Milky Way and that he knows he is in the Milky Way. This is the right analogy with the "large civilizations" problem since we know we're in a small civilization. In your version of the problem the clones get to bet too, so the answer depends on how your utility is accumulated over clones. If you have a well-defined utility function and you're using UDT, everything makes sense IMO. It doesn't change anything in principle. You just added another coin toss before the original coin toss which affects the odd of the latter.
0drnickbone7yWell we currently observe that we are in a small civilisation (though we could be in a zoo or simulation or whatever). But to assess the hypotheses in question we have to (in essence) forget that observation, create a prior for small universe versus big universe hypotheses, see what the hypotheses predict we should expect to observe, and then update when we "notice" the observation. Alternatively, if you adopt the UDT approach, you have to consider what utility function you'd have before knowing whether you are in a big civilization or not. What would the "you" then like to commit the "you" now to deciding? If you think you'd care about average utility in that original situation then naturally the small civilisations will get less weight in outcomes where there are big civilisations as well. Whereas if there are only small civilisations, they get all the weight. No difficulties there. If you think you'd care about total utility (so the small civs get equal weight regardless) then be carefully that it's bounded somehow. Otherwise you are going to have a known problem with expected utilities diverging (see http://lesswrong.com/lw/fg7/sia_fears_expected_infinity/ [http://lesswrong.com/lw/fg7/sia_fears_expected_infinity/]). A metaphorical coin with unknown (or subjectively-assigned) odds is quite a different beast from a physical coin with known odds (based on physical facts). You can't create crazy-sounding conclusions with metaphorical coins (i.e. situations where you bet at million to 1 odds, despite knowing that the coin toss was a fair one.)
0Squark7yI think that I care about a time-discounted utility integral within a future light-cone. Large civilizations entering this cone don't reduce the utility of small civilizations. I don't believe in different kinds of coins. They're all the same Bayesian probabilities. It's a meta-Occam razor: I don't see any need for introducing these distinct categories.
0drnickbone7yI'm not sure how you apply that in a big universe model... most of it is lies outside any given light-cone, so which one do you pick? Imagine you don't yet know where you are: do you sum utility across all light-cones (a sum which could still diverge in a big universe) or take the utility of an average light cone. Also, how do you do the time-discounting if you don't yet know when you are? My initial guess is that this utility function won't encourage betting on really big universes (as there is no increase in utility of the average lightcone from winning the bet), but it will encourage betting on really dense universes (packed full of people or simulations of people). So you should maybe bet that you are in a simulation, running on a form of dense "computronium" in the underlying universe.
0Squark7yThe possible universes I am considering already come packed into a future light cone (I don't consider large universes directly). The probability of a universe is proportional to 2^{-its Kolmogorov complexity} so expected utility converges. Time-discounting is relative to the vertex of the light-cone. Not really. Additive terms in the utility don't "encourage" anything, multiplicative factors do.
1drnickbone7yI was a bit surprised by this... if your possible models only include one light-cone (essentially just the observable universe) then they don't look too different from those of my stated hypothesis (at the start of the thread). What is your opinion then on other civilisations in the light-cone? How likely are these alternatives? * No other civilisations exist or have existed in the light-cone apart from us. * A few have existed apart from us, but none have expanded (yet) * A few have existed, and a few have expanded, but we can't see them (yet) * Lots have existed, but none have expanded (very strong future filter) * Lots have existed, and a few have expanded (still a strong future filter), but we can't see the expanded ones (yet) * Lots have existed, and lots have expanded, so the light-cone is full of expanded civilisations; we don't see that, but that's because we are in a zoo or simulation of some sort. Here's how it works. Imagine the "mugger" offers all observers a bet (e.g. at your 1000:1 on odds) on whether they believe they are in a simulation, within a dense "computronium" universe packed full of computers simulating observers. Suppose only a tiny fraction (less than 1 in a trillion) universe models are like that, and the observers all know this (so this is equivalent to a very heavily weighted coin landing against its weight). But still, by your proposed utility function, UDT observers should accept the bet, since in the freak universes where they win, huge numbers of observers win $1 each, adding a colossal amount of total utility to the light-cone. Whereas in the more regular universes where they lose the bet, relatively fewer observers will lose $1000 each. Hence accepting the bet creates more expected utility than rejecting it. Another issue you might have concerns the time-discounting. Suppose 1 million observers live early on in the light-cone, and 1 trillion live late in the light-cone (and again all observers know this). The mugg
0Squark7yMy disagreement is that the anthropic reasoning you use is not a good argument for non-existence of large civilizations. I am using a future light cone whereas your alternatives seem to be formulated in terms of a past light cone. Let me say that I think the probability to ever encounter another civilization is related to the ratio {asymptotic value of Hubble time} / {time since appearance of civilizations became possible}. I can't find the numbers this second, but my feeling is such an occurrence is far from certain. Very good point! I think that if the "computronium universe" is not suppressed by some huge factor due to some sort of physical limit / great filter, then there is a significant probability such a universe arises from post-human civilization (e.g. due to FAI). All decisions with possible (even small) impact on the likelihood of and/or the properties of this future get a huge utility boost. Therefore I think decisions with long term impact should be made as if we are not in a simulation whereas decisions which involve purely short term optimizations should be made as if we are in a simulation (although I find it hard to imagine such a decision in which it is important whether we are in a simulation). The effective time discount function is of rather slow decay because the sum over universes includes time translated versions of the same universe. As a result, the effective discount falls off as 2^{-Kolmogorov complexity of t} which is only slightly faster than 1/t. Nevertheless, for huge time differences your argument is correct. This is actually a good thing, since otherwise your decisions would be dominated by the Boltzmann brains appearing far after heat death.
0drnickbone7yIt is about 1/t x 1/log t x 1/log log t etc. for most values of t (taking base 2 logarithms). There are exceptions for very regular values of t. Incidentally, I've been thinking about a similar weighting approach towards anthropic reasoning, and it seems to avoid a strong form of the Doomsday Argument (one where we bet heavily against our civilisation expanding). Imagine listing all the observers (or observer moments) in order of appearance since the Big Bang (use cosmological proper time). Then assign a prior probability 2^-K(n) to being the nth observer (or moment) in that sequence. Now let's test this distribution against my listed hypotheses above: 1. No other civilisations exist or have existed in the universe apart from us. Fit to observations: Not too bad. After including the various log terms in 2^-K(n), the probability of me having an observer rank n between 60 billion and 120 billion (we don't know it more precisely than that) seems to be about 1/log (60 billion) x 1/log (36) or roughly 1/200. Still, the hypothesis seems a bit dodgy. How could there be exactly one civilisation over such a large amount of space and time? Perhaps the evolution of intelligence is just extraordinarily unlikely, a rare fluke that only happened once. But then the fact that the "fluke" actually happened at all makes this hypothesis a poor fit. A better hypothesis is that the chance of intelligence evolving is high enough to ensure that it will appear many times in the universe: Earth-now is just the first time it has happened. If observer moments were weighted uniformly, we would rule that out (we'd be very unlikely to be first), but with the 2^-K(n) weighting, there is rather high probability of being a smaller n, and so being in the first civilisation. So this hypothesis does actually work. One drawback is that living 13.8 billion years after the Big Bang, and with only 5% of stars still to form, we may simply be too late to be the first among many. If there were going
0drnickbone7yI was assuming that the "vertex" of your light cone is situated at or shortly after the Big Bang (e.g. maybe during the first few minutes of nucleosynthesis). In that case, the radius of the light cone "now" (at t = 13.8 billion years since Big Bang) is the same as the particle horizon "now" of the observable universe (roughly 45 billion light-years). So the light-cone so far (starting at Big Bang and running up to 13.8 billion years) will be bigger than Earth's past light-cone (starting now and running back to the Big Bang) but not massively bigger. This means that there might be a few expanded simulations who are outside our past light-cone (so we don't see them now, but could run into them in the future). Still if there are lots of civilisations in your light cone, and only a few have expanded, that still implies a very strong future filter. So my main point remains: given that a super-strong future filter looks very unlikely, most of the probability will be concentrated on models where there are only a few civilisations to start with (so not many to get filtered out; a modest filter does the trick). Ahh... I was assuming you discounted faster than that, since you said the utilities converged. There is a problem with Kolmogorov discounting of t. Consider what happens at t = 3^^^3 years from now. This has Kolmogorov complexity K(t) much much less than log(3^^3) : in most models of computation K(t) will be a few thousand bits or less. But the width of the light-cone at t is around 3^^^3, so the utility at t is dominated by around 3^^^3 Boltzmann Brains, and the product U(t) 2^-K(t) is also going to be around 3^^^3. You'll get similar large contributions at t = 4^^^^4 and so on; in short I believe your summed discounted utility is diverging (or in any case dominated by the Boltzmann Brains). One way to fix this may be to discount each location in space and time (s,t) by 2^-K(s,t) and then let u(s,t) represent a utility density (say the average utility per Plan
0Squark7yNo, it can be located absolutely anywhere. However you're right that the light cones with vertex close to Big Bang will probably have large weight to low K-complexity. This looks correct, but it is different from your initial argument. In particular there's no reason to believe MWI is wrong or anything like that. It is guaranteed to converge and seems to be pretty harsh on BBs either. Here is how it works. Every "universe" is an infinite sequence of bits encoding a future light cone. The weight of the sequence is 2^{-K-complexity}. More precisely I sum over all programs producing such sequences and give weight 2^{-length} to each. Since sum of 2^-{length} over all programs is 1 I get a well-defined probability measures. Each sequence gets assigned a utility by a computable function that looks like integral over space-time with temporal discount. The temporal discount here can be fast e.g. exponential. So the utility function is bounded and its expectation value converges. However the effective temporal discount is slow since for every universe, its sub-light-cones are also within the sum. Nevertheless its not so slow that BBs come ahead. If you put the vertex of the light cone at any given point (e.g. time 4^^^^4) there will be few BBs within the fast cutoff time and most far points are suppressed due to high K-complexity.
0drnickbone7yAh, I see what you're getting at. If the vertex is at the Big Bang, then the shortest programs basically simulate a history of the observable universe. Just start from a description of the laws of physics and some (low entropy) initial conditions, then read in random bits whenever there is an increase in entropy. (For technical reasons the programs will also need to simulate a slightly larger region just outside the light cone, to predict what will cross into it). If the vertex lies elsewhere, the shortest programs will likely still simulate starting from the Big Bang, then "truncate" i.e. shift the vertex to a new point (s, t) and throw away anything outside the reduced light cone. So I suspect that this approach gives a weighting rather like 2^-K(s,t) for light-cones which are offset from the Big Bang. Probably most of the weight comes from programs which shift in t but not much in s. That's what I thought you meant originally: this would ensures that the utility in any given light-cone is bounded, and hence that the expected utility converges. I disagree. If models like MWI and/or eternal inflation are taken seriously, then they imply the existence of a huge number of civilisations (spread across multiple branches or multiple inflating regions), and a huge number of expanded civilisations (unless the chance of expansion is exactly zero). Observers should then predict that they will be in one of the expanded civilisations. (Or in UDT terms, they should take bets that they are in such a civilisation). Since our observations are not like that, this forces us into simulation conclusions (most people making our observations are in sims, so that's how we should bet). The problem is still that there is a poor fit to observations: yes we could be in a sim, and it could look like this, but on the other hand it could look like more or less anything. Incidentally, there are versions of inflation and many worlds which don't run into that problem. You can always take
0Squark7yIn some sense it does, but we must be wary of technicalities. In initial singularity models I'm not sure it makes sense to speak of "light cone with vertex in singularity" and it certainly doesn't make sense to speak of a privileged point in space. In eternal inflation models there is no singularity so it might make space to speak of the "Big Bang" point in space-time, however it is slightly "fuzzy". I don't think it does. If we are not in a sim, our actions have potentially huge impact since they can affect the probability and the properties of a hypothetical expanded post-human civilization. In UDT it doesn't make sense to speak of what "actually exists". Everything exists, you just assign different weights to different parts of "everything" when computing utility. The "U" in UDT is for "updateless" which means that you don't update on being in a certain branch of the wavefunction to conclude other branches "don't exist", otherwise you lose in counterfactual mugging.
2drnickbone7ySo: if a bet is offered that you are a sim (in some form of computronium) and it becomes possible to test that (and so decide the bet one way or another), you would bet heavily on being a sim? But on the off-chance that you are not a sim, you're going to make decisions as if you were in the real world, because those decisions (when suitably generalized across all possible light-cones) have a huge utility impact. Is that right? The problem I have is this only works if your utility function is very impartial (it is dominated by "pro bono universo" terms, rather than "what's in it for me" or "what's in it for us" terms). Imagine for instance that you work really hard to ensure a positive singularity, and succeed. You create a friendly AI, it starts spreading, and gathering huge amounts of computational resources... and then our simulation runs out of memory, crashes, and gets switched off. This doesn't sound like it is a good idea "for us" does it? This all seems to be part of a general problem with asking UDT to model selfish (or self-interested) preferences. Perhaps it can't [http://lesswrong.com/lw/jpr/sudt_a_toy_decision_theory_for_updateless/am31]. In which case UDT might be a great decision theory for saints, but not for regular human beings. And so we might not want to program UDT into our AI in case that AI thinks it's a good idea to risk crashing our simulation (and killing us all in the process). I've remarked elsewhere [http://lesswrong.com/lw/jpr/sudt_a_toy_decision_theory_for_updateless/am2t] that UDT works best against a background of modal realism [http://en.wikipedia.org/wiki/Modal_realism], and that's essentially what you've said here. But here's something for you to ponder. What if modal realism is wrong? What if there is, in fact, evidence that it is wrong, because the world as we see it is not what we should expect to see if it was right? Isn't it maybe a good idea to then - er - update on that evidence? Or does a UDT agent have to stay dog
2Squark7yIt depends on the stakes of the best. It's not an "off-chance". It is meaningless to speak of the "chance I am a sim": some copies of me are sims, some copies of me are not sims. It surely can: just give more weight to humans of a very particular type ("you"). Subjective expectations are meaningless in UDT. So there is no "what we should expect to see". Does it have to stay dogmatically committed to Occam's razor in the face of whatever it sees? If not, how would it arrive at a replacement without using Occam's razor? There must be some axioms at the basis of any reasoning system.
2drnickbone7yI thought we discussed an example earlier in the thread? The gambler pays $1000 if not in a simulation; the bookmaker pays $1 if the gambler is in a simulation. In terms of expected utility, it is better for "you" (that is, all linked instances of you) to take the gamble, even if the vast majority of light-cones don't contain simulations. No it isn't meaningless: chances simply become operationalised in terms of bets, or other decisions with variable payoff. The "chance you are a sim" becomes equal to the fraction of a util you are prepared to pay for a betting slip which pays out one util if you are a sim, and pays nothing otherwise. (Lots of linked copies of "you" take the gamble; some win, some lose.) Incidentally, in terms of original modal realism (due to David Lewis), "you" are a concrete unique individual who inhabits exactly one world, but it is unknown which one. Other versions of "you" are your "counterparts". It is usually not possible to group all your counterparts together and treat them as a single (distributed) being, YOU, because the counterpart relation is not an equivalence relation (it doesn't partition possible people into neat equivalence classes). As one example, imagine a long chain of possible people whose experiences and memories are indistinguishable from immediate neighbours in the chain (and they are counterparts of their neighbours). But there is a cumulative "drift" along the chain, so that the ends are very different from each other (and not counterparts). A subjective expectation is rather like a bet: it is a commitment of mental resource to modelling certain lines of future observations (and preparing decisions for such a case). If you spend most of your modelling resource on a scenario which doesn't materialise, this is like losing the bet. So it is reasonable to talk about subjective expectations in UDT; just model them as bets. Occam's razor here is just a method for weighting hypotheses in the prior. It is only "dogmatic" if
2Squark7yIt is not the case if the money can be utilized in a manner with long term impact. This doesn't give an unambiguous recipe to compute probabilities since it depends on how the results of the bets are accumulated to influence utility. An unambiguous recipe cannot exist since it would have to give precise answers to ambiguous questions such as: if there are two identical simulations of you running on two computers, should they be counted as two copies or one? UDT doesn't seem to work this way. In UDT, "you" are not a physical entity but an abstract decision algorithm. This abstract decision algorithm is correlated to different extent with different physical entities in different worlds. This leads to the question of whether some algorithms are more "conscious" than others. I don't think UDT currently has an answer for this, but neither do other frameworks. If we think of knowledge as a layered pie, with lower layers corresponding to knowledge which is more "fundamental", then somewhere near the bottom we have paradigms of reasoning such as Occam's razor / Solomonoff induction and UDT. Below them lie "human reasoning axioms" which are something we cannot formalize due to our limited introspection ability. In fact the paradigms of reasoning are our current best efforts at formalizing this intuition. However, when we build an AI we need to use something formal, we cannot just transfer our reasoning axioms to it (at least I don't know how to do it; meseems every way to do it would be "ingenuine" since it would be based on a formalism). So, for the AI, UDT (or whatever formalism we use) is the lowest layer. Maybe it's a philosophical limitation of any AGI, but I doubt it can be overcome and I doubt it's a good reason not to build an (F)AI.
2drnickbone7yOK, I was using $ here as a proxy for utils, but technically you're right: the bet should be expressed in utils (as for the general definition of a chance that I gave in my comment). Or if you don't know how to bet in utils, use another proxy which is a consumptive good and can't be invested (e.g. chocolate bars or vouchers for a cinema trip this week). A final loop-hole is the time discounting: the real versions of you mostly live earlier than the sim versions of you, so perhaps a chocolate bar for the real "you" is worth many chocolate bars for sim "you"s? However we covered that earlier in the thread as well: my understanding is that your effective discount rate is not high enough to outweigh the huge numbers of sims. Well this is your utility function, so you tell me! Imagine a hacker is able to get into the simulations and replace pleasant experiences by horrible torture. Does your utility function care twice as much if he hacks both simulations versus hacking just one of them? (My guess is that it does). And this style of reasoning may cover limit cases like a simulation running on a wafer which is then cut in two (think about whether the sims are independently hackable, and how much you care.)
0Squark7yIt wouldn't be exactly twice but you're more or less right. However, it has no direct relation to probability. To see this, imagine you're a paperclip maximizer. In this case you don't care about torture or anything of the sort: you only care about paperclips. So your utility function specifies a way of counting paperclips but no way of counting copies of you. From another angle, imagine your two simulations are offered a bet. How should they count themselves? Obviously it depends on the rules of the bet: whether the payoff is handed out once or twice. Therefore, the counting is ambiguous. What you're trying to do is writing the utility function as a convex linear combination of utility functions associated with different copies of you. Once you accomplish that, the coefficients of the combination can be interpreted as probabilities. However, there is no such canonical decomposition.
0drnickbone7yI think it works quite well with "you" as a concrete entity. Simply use the notion that "your" decisions are linked to those of your counterparts (and indeed, to other agents), such that if you decide in a certain way in given circumstances, your counterparts will decide that way as well. The linkage will be very tight for neighbours in the chain, but diminishing gradually with distance, and such that the ends of the chain are not linked at all. This - I think - addresses the problem of trying to identify what algorithm you are implementing, or partitioning possible people into those who are running "the same" algorithm.
0Squark7yActually I was speaking of a different problem, namely the philosophical problem of which abstract algorithms should be regarded as conscious (assuming the concept makes sense at all). The identification of oneself's algorithm is an introspective operation whose definition is not obvious for humans. For AIs the situation is clearer if we assume the AI has access to its own source code.
0Nisan7yOh I see, that makes sense.
0polymathwannabe7yI agree with this counterargument, but this thread being what it is, in which direction should I vote sub-comments?
0Squark7ySubcomments are voted by the ordinary rules

Irrationality Game: One can reliable and predictably make $1M / year, and it's not that difficult. (Confidence: 75%)

5[anonymous]7yWhat do you mean by “one”? Literally anyone at all? Anyone at least as smart as the average LWer? Something else?
2Alexei7yLet's say someone with an engineering bachelor's degree.
4polymathwannabe7yCare to describe how?
-1Alexei7yFiguring that out right now.
1John_Maxwell7yPlease elaborate!

Irrationality game:

There are other 'technological civilizations' (in the sense of intelligent living things that have learned to manipulate matter in a complicated way) in the observable universe: 99%

There are other 'technological civilizations' in our own galaxy: 75% with most of the probability mass in regimes where there are somewhere between dozens and thousands.

Conditional on these existing: Despite some being very old, they are limited by the hostile nature of the universe and the realities of practical manipulation of matter and energy to never co... (read more)

4Eugine_Nier7yOne statement per comment please.
2CellBioGuy7yProbably overdid it with that one. Will split things up more in subsequent comments.
0[anonymous]7yWhat do you mean by “complicated”? Were humans a technological civilization in 1900? In 1700? In 10,000 BC? In 50,000 BC?
0shokwave7yEven given other technological civilisations existing, putting "matter and energy manipulation tops out a little above our current cutting edge" at 5% is way off.
0ThisSpaceAvailable7yWay off in which direction?
0CellBioGuy7yThere's a lot you can do on the surface of a clement planet, and a lot you can do in a solar system without replicators that eat everything. Also depends on what you mean by 'above'.

Irrationality game:

Nice idea. This way I can safely test whether the Baseline of my opinion on LW topics is as contrarian as I think.

My proposition:

On The Simulation Argument I go for "(1) the human species is very likely to go extinct before reaching a “posthuman” stage" (80%)

Correspondingly on The Great Filter I go for failure to reach "9. Colonization explosion" (80%).

This is not because I think that humanity is going to self-annihilate soon (though this is a possibility).

0Squark7yWhat is the extinction scenario you have in mind?
1Gunnar_Zarncke7yHopefully no extinction during the next many thousends of years. Which extinction afterwards is difficult to predict. As I shortly argued in my baseline post I think that posthuman state is unlikely due to thermodynamics/complexity constraints.

Irrationality game: people are happier when living in traditional social structures, and value being part of their traditions[1]. The public existence of "weird" relationships (homosexuality, polyamory, BDSM, ...) is actively harmful to most people; the open practice of them is a net negative for world utility. Morally good actions include condemnation and censorship of such things.

[1] Or rather what they believe are their traditions; these beliefs may not be particularly well-correlated with reality.

3Lumifer7yThat looks to be a pretty standard neo-reactionary position.
0lmm7ySure, but I expect the proportion of LW that actually agrees with neoreaction to be small (and I disagree with most neoreactionary positions).
2[anonymous]7yProbability?
0lmm7yAh, sorry. 0.8
0John_Maxwell7yHow did you come to believe this?

Irrationality Game: Currently, understanding history or politics is a better avenue than studying AI or decision theory for dealing with existential risk. This is not because of the risk of total nuclear annihilation, but because of the possibility of political changes that result in setbacks to or an accelerated use and understanding of AI. 70%

6[anonymous]7yI think competition is far tougher in politics than in studying decision theory.

I'm 99% confident that dust specks in 3^^^3 eyes result in less lost utility than 50 years of torturing one person.

Utility seems underspecified here.

3[anonymous]7y
1rthomas67yI agree with you (maybe not 99% certainty though), and I'm surprised more people do not.That is, assuming the original stipulation of the dust specks causing only a "mild inconvenience" for everyone, and not some sort of bell curve of inconvenience with a mean of "mild". People around here seem to grok the idea of the hedonic treadmill, so why don't they apply that idea to this situation? Assuming all of those 3^^^3 people all truly only have a "mild inconvenience", I would argue that from a subjective point of view from each individual, the utility of their day as a whole has not been diminished at all. Actually, the more I think about it, the idea itself is poorly formed. It depends a lot on what sort of inconvenience the dust causes. If it causes 0.00000001% of the people to decide to shoot up a store or something, then I guess the one person being tortured would be better. But if the dust does not cause any sort of cascading effect, if it's truly isolated to the lost utility of the dust itself, then I'd say the dust is better.
3polymathwannabe7yElsewhere I argued that the pain from the dust specks doesn't add up (and is therefore not really comparable to one single person's torture) unless the victims are forming a hive mind. What the thought experiment is actually comparing is one instance of horrible pain versus many, many individual and not groupable instances of minor discomfort.
[-][anonymous]7y 2

.

[This comment is no longer endorsed by its author]Reply
4Will_Newsome7yIf this train of thought continues along its natural course, you have to wonder why you are being "shown" the experience you have this moment as you read this, rather than some other more interesting or influential moment. Also, it is not clear that you would want to use this kind of anthropic reasoning to determine a policy; people that are not conscious but think they are would incorrectly draw the same conclusions and thus muck up the social commons with their undue senses of specialness. ETA: There are a few other counterarguments similar to those in the previous paragraph. This has perturbed me for many years now, because the line of reasoning in the parent comment really does seem like the most intuitive approach to subjective anthropics. I'd be very satisfied to find a solution, but it seems equally likely that there's just something pretty wrong with our intuitions about (relative) existence, which has implications for which kinds of decision theories we should be willing to put our weight on. ETA2: And the UDT pragmatist in me wonders whether it even means anything for a hypothesis to be true, if it rationally shouldn't affect your decisions. If anything I would lean toward decision theoretic epiphenomenalism implying falsehood.
1RowanE7yThe part of this I disagree with most is putting the probability as low as 10%. I up voted, since that seemed like just putting a number to the word "significant", and the other claims seem about right.

Why are we reviving this at all?

[-][anonymous]7y 10

Just as a curiosity, this was the most downvoted comment in the original thread:

For a large majority of people who read this, learning a lot about how to interact with other human beings genuinely and in a way that inspires comfort and pleasure on both sides is of higher utility than learning a lot about either AI or IA. ~90%

(-44 points)

[This comment is no longer endorsed by its author]Reply
8falenas1087yThis is a time that the system of hiding votes less than -3 is a bad thing. In this thread, downvotes indicate that a belief that people may have thought was rare is actually pretty common on LW, which is something I'm interested in seeing.
7Douglas_Knight7yYou could do this with polls instead of karma. The advantage of karma is that it provides an incentive for people to play to win. The disadvantage is hiding comments.
4John_Maxwell7y"provides an incentive for people to play to win" You mean an incentive to hold irrational beliefs? Is that something we want to incentivize?
0ThisSpaceAvailable7yNo, not to hold irrational beliefs, but to admit to holding irrational beliefs.
1Fhyve7yI agree. I want to comment on some of the downvoted posts, but I don't want to pay the karma
0CellBioGuy7yGreat idea. I'll put a note in the post so that if anyone ever resurrects this in the future they'll do it that way.
4Ixiel7yShould we down vote posts with many propositions if we agree with a majority? One? All? There are already two split clusters for me.
0CellBioGuy7yHmm. I'd recommend if the split has one that's much stronger go with that vote, otherwise leave it at zero and explain in a comment.
3Eugine_Nier7yUnfortunately, since the first irrationality game, the hiding code was changed so that this is no longer possible.
3bramflakes7yAm I allowed to post about whether a counterfactual world would be "better" in some sense, if I specify something like "If Y had happened instead of X, the number of excess deaths from then till now would be lower / economic growth would have been better" ? I don't know whether that falls under preferences disguised as beliefs.
0CellBioGuy7yPerhaps you can try to turn it in a more generalized form?
0bramflakes7yHow do you mean?

Irrationality game:

Most progress in medicine in the next 50 years won't be due to advances in molecular biology and the production of drugs that are designed to target specific biochemical pathways but through other paradigms.

Probability: 75%

1polymathwannabe7yWhich other paradigms do you predict will become more relevant?
0ChristianKl7yI think that you can learn a lot via empirical measurement without needing to understand underlying biochemistry. Apart from direct measurement it also about developing better metrics for things like lung function, That partly why I invested significant effort into Quantified Self community building. Exploring the phenomenological aspect of illnesses provides an area with a lot of untapped potential for knowledge gathering. I think there are large returns found in studying human movement in detail. As it stands a method like Feldenkrais isn't has some studies to support it but no strong scientifically investigated base. It should be possible to use cameras and computers to get accurate models of human movement and investigate an approach like Feldenkrais in deeper scientific way. Relaxation is generally poorly understood. I think most patients could benefit from spending hours in floating tanks after having a major operation, yet few hospitals have floating tanks or maximize the relaxation in other ways. In general hospital beds and hospital food doesn't seem to be optimized for health outcomes. Having a well developed theory that can predict placebo effects accurately would be good. On the one hand it will make it easier to gather knowledge, on the other hand it will help doctors to frame their interactions in a way that helps patients. Having good empathy training for doctors has potential to improve healthcare. Psychological interventions like hypnosis. Biofeedback. That's just a list of possibilities I can think of. There are probably things that are unknown unknowns for me.
0polymathwannabe7yThose approaches risk turning therapy into stabs in the dark by neglecting the details of what is actually going on inside the black box.
0ChristianKl7yMost people who pretend that they know what goes on inside the black box are wrong anyway. Drug companies would like to predict which components work based on understanding biochemistry but they still have to run expensive trials in which over 90% of the components fail. Furthermore it becomes exponentially more expensive [http://pipeline.corante.com/archives/2012/03/08/erooms_law.php] to discover additional drugs in that way. That said, the idea of preaching blindness that currently haunts medical research is exactly about the virtue of stabing into the dark. If you would see what you are doing you wouldn't be objective anymore.

Irrationality game: The straightforward view of the nature of the universe is fundamentally flawed. 90%

By "fundamentally flawed", I mean things like:

  • I am currently dreaming.
  • The singularity has already happened, and this world maximizes my CEV.
  • I am a Boltzmann brain.
  • I am in a simulation.
  • Or some similar thing is true, but I haven't thought of it.
[-][anonymous]7y 0

Irrationality game: The Great Stagnation is actually occurring, and it is mostly due to fossil fuel depletion rather than (say) leftist politics or dysgenics. (60%)

[This comment is no longer endorsed by its author]Reply

Irrationality game: most opposition to wireheading comes from seeing it as weird and/or counterintuitive in the same way that most non-LWers see cryonics/immortalism as weird. Claiming to have multiple terminal values is an attempt to justify this aversion. 75%

The Hellenistic astronomers (300BC-0) were generally heliocentric. 90%

[-][anonymous]7y -2

Irrationality Game: We need a way to give feedback on irrationality game entries that the troll toll won't mess with. (98%)

[pollid:643]

Irrationality Game:

Everyone alive in developed nations today will die a fairly standard biological death by age:

150: 75%

250: 95%

(This latter figure accounts for the possibility that the stories of the odd Chinese monk living to age 200+ after only eating wild herbs from age 10 on up is actually true and not an exaggeration, or someone sticking to unreasonably-effective calorie restriction regimes religiously combined with some interesting metabolic rejiggering in the coming decade or two).

The majority (90+%) of people born in developed nations today will die a fairly standard biological death by age:

120: 85%

150: 99%

2Eugine_Nier7yI think the probability of a nuclear war or bio-engineered plague is higher than 5%.
1[anonymous]7yI think there's a good chance that a nuclear war would kill less than 90% of the population, though.
0[anonymous]7y(But now that I think about it, even the ones it doesn't kill straight away will be much less likely to live to 120 than they otherwise would.)

Irrationality Game:

Politics (in particular, large governments such as the US, China, and Russia) are a major threat to the development of friendly AI. Conditional on FAI progress having stopped, I give a 60% chance that it was because of government interference, rather than existential risk or some other problem.

[+][anonymous]7y -7