[AN #122]: Arguing for AGI-driven existential risk from first principles — LessWrong