Selfishness, preference falsification, and AI alignment — LessWrong