LLMs as amplifiers, not assistants — LessWrong