Clarifying how misalignment can arise from scaling LLMs — LessWrong