Anthropic decision theory for selfish agents — LessWrong