Could utility functions be for narrow AI only, and downright antithetical to AGI? — LessWrong