[AN #84] Reviewing AI alignment work in 2018-19 — LessWrong