Scalar reward is not enough for aligned AGI
This post was authored by Peter Vamplew and Cameron Foale (Federation University), and Richard Dazeley (Deakin University) Introduction Recently some of the most well-known researchers in reinforcement learning Silver, Singh, Precup and Sutton published a paper entitled Reward is Enough, which proposes the reward-is-enough hypothesis: “Intelligence, and its associated abilities,...
Jan 17, 202223