Multi-agent predictive minds and AI alignment — LessWrong