Benefits/Risks of Scott Aaronson's Orthodox/Reform Framing for AI Alignment — LessWrong