Part 3 of a series. Here are part 1 and part 2. One of the things that always surprised me is how few people in AI were interested in AI safety and alignment purely out of intellectual curiosity. These topics raise the kind of novel, foundational problems that scientists typically...
Other people have written about reasons why we should trust AIs; the main one in my mind is that it’s possible to look at the computations they perform when producing an output (even if we struggle to understand them). I’m going to write about reasons why we shouldn’t trust AIs,...
Awareness and concern about the extinction risk posed by AI has been increasing the whole time I’ve been in the field. It feels like it’s finally going mainstream. But it’s also felt this way before… …picking up where we left off in my previous post about how I got into...
Suppose we succeed and bring AI to a screeching halt. Then what? What direction do we want to go? Can we actually stop AI from advancing at all? For how long? What are we going to do with whatever extra time we have to make the future a safer place...
I don’t know where the idea of “marginal risk” came from in AI policy. It sounds like BS. Yet another excuse to keep building dangerous AI systems… The basic idea is that instead of looking at how likely your AI system is to lead to millions of deaths, you ask...
This is a weekly round-up of things I’ve posted in the last week. InkHaven requires that I post a blog post every day, which is a lot. Especially for people subscribed to my blog. Someone requested I spare their inbox, so I haven’t been sending out every post. So now...
I think we need to Stop AI. Specifically we need to Stop AI Now. We can’t wait around. The standard metaphor is a runaway train heading towards a cliff. Let’s work with that. We don’t know when to stop. We don’t know where the cliff is. World’s most-cited-scientist (and my...