x
Investigating encoded reasoning in LLMs — LessWrong