Uninfluenceable learning agents — LessWrong