"LLMs Don't Have a Coherent Model of the World" - What it Means, Why it Matters — LessWrong