So now we're in a situation mildly close to an interesting epistemic situation, which is, winning the lottery. Winning the lottery or some better-optimized event provides a lot of incommunicable evidence that you're in a simulation. The typical anthropic problem in group epistemology -- your winning tells me nothing. I have a question for you: How serious a problem do you think this is in practice? If it's a common problem and been one throughout history, what social institutions would have evolved to help solve the problem? Or is solving the problem impossible? Only try to answer these if you're interested in the questions themselves of course.

You're looking at it all wrong, "you" are not "in" any simulation or universe. There exists instantiations of the algorithm, including the fact that it remembers winning the lottery, which is you in various universes and simulations and boltzman brains and other things, with certainty (for our purposes), and what you need to do depends on what you want ALL instances to do. It doesn't matter how many simulations of you are run, or what measure they have, or anything else like that, if your decisions within them don't matter for the multi... (read more)

This post is for sacrificing my credibility!

by Will_Newsome 1 min read2nd Jun 2012347 comments


Thank you for your cooperation and understanding. Don't worry, there won't be future posts like this, so you don't have to delete my LessWrong account, and anyway I could make another, and another.

But since you've dared to read this far:

Credibility. Should you maximize it, or minimize it? Have I made an error?


Don't be shallow, don't just consider the obvious points. Consider that I've thought about this for many, many hours, and that you don't have any privileged information. Whence our disagreement, if one exists?