The biggest event in 2025 for me is entering the field of AI safety research. In particular, I work with collaborators at Geodesic Research on designing a suite of health metrics to evaluate the "pathologies" in the chain-of-thought reasoning of Large Language Models, such as post-hoc, internalized, and encoded reasoning. I've also started reading the Philosophical Investigations by Ludwig Wittgenstein in my leisure time. To my pleasant surprise, Wittgenstein investigated these same phenomena in human language in his famous work. Given the speed of LLM development today and the urgent need for upholding monitorability of chain-of-thought reasoning (see this recent post from OpenAI), I find his work enlightening for drawing parallels between... (read 985 more words →)