Looking for AGI philosopher co-author on Draft Article: 'Optimising Peace to Constrain Risk of War from an Artificial Superintelligence' — LessWrong