Humans aren't agents - what then for value learning? — LessWrong