Are we too confident about unaligned AGI killing off humanity? — LessWrong