Reward splintering for AI design — LessWrong