Far future, existential risk, and AI alignment — LessWrong