Why are we sure that AI will "want" something? — LessWrong