On motivations for MIRI's highly reliable agent design research — LessWrong