x
"I’ve observed a recurring pattern across frontier LLMs where, as multi-step reasoning depth increases, models sometimes maintain internal/persona coherence while drifting from semantic truth-states. I’m sharing this to ask whether this behavior is a known scaling byproduct or an evaluation blind spot. Example traces available if useful." — LessWrong