Anthropic answers to logical uncertainties? — LessWrong