Reflections on Larks’ 2020 AI alignment literature review — LessWrong