How can I reduce existential risk from AI? — LessWrong