Exploring Last-Resort Measures for AI Alignment: Humanity's Extinction Switch — LessWrong