I expect it’ll take another week or two for everyone to fully digest the significance of Claude Mythos Preview. In the meantime, here are my initial thoughts. Gradually, then suddenly Mythos is radically better at cyber than any previous model: It isn’t the first model that can find vulnerabilities, of...
I see a lot of AI safety strategies that don’t fully engage with the complexity of the real world—and therefore are unlikely to succeed in the real world. To take a simple example: many strategies rely heavily on government playing a leading role through regulation and perhaps even nationalization. That’s...
My AI editor is essential to my writing flow and has made me a stronger and more consistent writer. I get a lot of questions about my setup, so I’m going to talk about how I think about the role of AI, how I set up my editing workflow, and...
We are in a strange situation: the big labs take AI risk more seriously than the government, and are doing a better job of preparing for it. There has been surprising progress on alignment over the last few years, and some strong work figuring out how to shape model behavior....
Nobody said the path would be clear. We know we need to prepare for AGI, but how do we do that if we don’t know whether it’s coming in 3 years or in 100? What about recursive self improvement: will that escalate to superintelligence, or fizzle out? And as the...
There’s broad (though not universal) agreement that present day AI is probably not conscious, but very little agreement about whether consciousness is likely to emerge as we move toward AGI. This isn’t an abstract question: AI consciousness has major implications for alignment. Further, a conscious AI might have moral rights...
I’m pleased to report that I have no new AI-related crises for you this week. Instead we get to focus on the fun parts, starting with physical constraints on AI development. Dylan Patel explains how power, GPUs, and memory will each be crucial bottlenecks on AI development over the next...