Naive Hypotheses on AI Alignment — LessWrong