x
Analysing CoT alignment in thinking LLMs with low-dimensional steering — LessWrong