This project was conducted as part of the SPAR Fall 2025 cohort. TL;DR * Chain-of-thought (CoT) monitoring may serve as a core pillar for AI safety if further advancements in AI capabilities do not significantly degrade the monitorability of LLM serial reasoning. * As such, we studied the effects of...
TL;DR: Most AI forecasts generally assume that if a conflict over Taiwan occurs, it will largely be about AI. I think there's a decent chance for a conflict before either side becomes substantially AGI-pilled. Thanks to Aaron Scher for comments on a draft of this post. I'm no China expert,...
More people should probably be thinking about research automation. If automating research is feasible prior to creating ASI it could totally change the playing field, vastly accelerating the pace of progress and likely differentially accelerating certain areas of research over others. There's a big risk, though, that AI capabilities research...
The title of this book is clickbait. "Encounters with Einstein" is a short collection of lectures given or written by Werner Heisenberg in the 1970's, only one of which discusses Einstein.[1] The remaining lectures discuss various aspects of the history and development of quantum mechanics and science in general: tradition...
Based on research performed as a PIBBSS Fellow with Tomáš Gavenčiak as well as work supported by EA Funds and Open Philanthropy. tl;dr: I'm investigating whether LLMs track and update beliefs during chain-of-thought reasoning. Preliminary experiments with older models (without reasoning training) have not been able to measure this; I...