Note: Appendix re-analysing with the addition of Gemini 3.1 data added 20/02/26 - this substantially changed the results One of the most attended to benchmarks for any new model these days is the METR estimated time horizon for “a diverse set of multi-step software and reasoning tasks”[1]. This comes in...
If it can be done by AI, it probably will be Cross posted from my Substack. Introduction I found myself in a discussion over the holidays with Seb Krier from Google DeepMind about what factors will be most important in determining the effect of powerful AI on the labour market....
cross posted from my Substack Introduction Following on from my last post on Comparative Advantage & AGI, more discussion on whether you might go hungry in a post AGI world. This time, we are looking at a hypothetical world where AGI has been developed and is in some sense misaligned:...
Cross-posted from my substack. Epistemic status: pretty confident, but willing to believe I’ve missed something somehow, hoping somebody smart will point it out if so. Level: pretty introductory, though assumes a bit of economics knowledge - if you’ve thought carefully or extensively about this particular topic I don’t expect you...
AI infrastructure numbers are hard to find with any precision. There are many reported numbers of “[company] spending Xbn on infrastructure this quarter” and “[company] has bought 100k H100s or “has a cluster of 100k H100s” but when I went looking for an estimate of how much compute a given...
Cross-posted from the EA Forum Introduction Many people, myself included, worry about how much we can trust aggregate forecasts (e.g., the Metaculus community median) that are based on the predictions of only a small number of individuals. This consideration also came up in my recent post analysing predictions of future...
If an AI system or systems goes wrong in the near term and causes harm to humans in a way which is consistent with or supportive of alignment being a big deal, what might it look like? I'm asking because I'm curious about potential fire-alarm scenarios (including things which just...