Acknowledging Human Preference Types to Support Value Learning — LessWrong