x
Towards an Axiological Approach to AI Alignment — LessWrong