Can AI Agents with Divergent Interests Learn To Prevent Civilizational Failures? — LessWrong