Many people who are paying attention to the trajectory of AI worry about its potential to concentrate power. I think this is a reasonable thing to worry about, with some important caveats. If someone builds a superintelligence, I think they are far more likely to die ignominiously with the rest...
This April 1st, I’m pleased to report that everything is fine. We did it! We saved the world. Congratulations, humanity. There are no more looming apocalypses, no desperate screaming crises, no unendorsedly miserable people on Earth, no creeping degeneration of death and aging existing as a perpetual affront against my...
In my occasional advising calls with aspiring AI Safety folks, one of the most common questions I get is “What courses should I take next?” I often find myself replying: “None; go do stuff instead.” Fabricando fit faber. By making, one becomes a maker. There are a lot of courses...
For several years now, I’ve been fielding calls from people who want to help make AI go well for humanity. Sometimes these folks are based outside the U.S., and often they ask me: With most of the labs concentrated in a few places (the U.S., China, and DeepMind in the...
This post is part of the sequence Against Muddling Through. Nate and Eliezer have written about why they think that even a mostly aligned system won’t safely generalize. This seems true whether “mostly aligned” means “we’ve hard-coded in most of the terminal goals we care about” or “it’s mostly done...
This post is part of the sequence Against Muddling Through. In an earlier post, I quoted Will McAskill: > Most directions I can fire a gun don’t hit the target; that doesn’t tell you much about how likely I am to hit the target if I’m aiming at it. What...
This post is part of the sequence Against Muddling Through. In previous posts, I discussed why I don’t think there’s a “sweet spot” for capabilities in which AIs are smart enough to solve alignment but too dumb to rebel. I also discussed why I don’t think we can get AIs...