Could an AI Alignment Sandbox be useful? — LessWrong