On “first critical tries” in AI alignment — LessWrong