Rejecting Violence as an AI Safety Strategy — LessWrong