Trapping AIs via utility indifference — LessWrong