Scenario planning for AI x-risk — LessWrong