Moral gauge theory: A speculative suggestion for AI alignment — LessWrong