Thanks for writing this!
I think that to fully engage with their position one needs to study the additional online resources section they have created:
https://ifanyonebuildsit.com/resources
I have just started looking at it, but I am seeing way more details and nuance there compared to what tends to percolate into the discussions about the book.
For those of us who are engaging with AI existential risk professionally, the resources section is the level to engage with (e.g. if one has doubts that an attempted total ban would decrease the danger rather than increasing it instead, the resource section has a lot of relevant material to evaluate before making up one’s mind).
No, Seriously, If Anyone Builds It, [Probably] Everyone Dies
My very positive full review was briefly accidentally posted and emailed out last Friday, whereas the intention was to offer it this Friday, on the 19th. I’ll be posting it again then. If you’re going to read the book, which I recommend that you do, you should read the book first, and the reviews later, especially mine since it goes into so much detail.
If you’re convinced, the book’s website is here and the direct Amazon link is here.
In the meantime, for those on the fence or who have finished reading, here’s what other people are saying, including those I saw who reacted negatively.
Quotes From The Book’s Website
Positive Reviews
Scott Alexander’s very positive review.
Harlan Stewart created a slideshow of various favorable quotes.
Matthew Yglesias recommends the book.
As some comments note the book’s authors do not actually think there is an outright 0% chance of survival, but think it is on the order of 0.5%-2%.
New York Times profile of Eliezer Yudkowsky by Kevin Roose is a basic recitation of facts, which are mostly accurate. Regular readers here are unlikely to find anything new, and I agree with Robin Hanson that it could have been made more interesting, but as New York Times profiles go ‘fair, mostly accurate and in good faith’ is great.
Steven Adler goes over the book’s core points.
Here is a strong endorsement from Richard Korzekwa.
Here’s the ultimate endorsement:
The Book In Audio
Semafor’s Reed Albergotti offers his take, along with an hourlong interview.
Hard Fork covers the book (this is the version without the iPhone talk at the beginning, here is the version with iPhone Air talk first).
The AI Risk Network covers the book (21 minute video).
Liron Shapira interviews Eliezer Yudkowsky on the book.
Friendly Skeptical Reviews
Shakeel Hashim reviews the book, agrees with the message but finds the style painful to read and thus is very disappointed. He notes that others like the style.
I liked the style, but it is not for everyone and it is good to offer one’s accurate opinion. It is also very true, as I have learned from writing about AI, that a lot of what can look like bad writing or talking about obvious or irrelevant things is necessary shadowboxing against various deliberate misreadings (for various values of deliberate) and also people who get genuinely confused in ways that you would never imagine if you hadn’t seen it.
Most people do not agree with the book’s conclusion, and he might well be very wrong about central things, but he is not obviously wrong, and it is very easy (and very much the default) to get deeply confused when thinking about such questions.
My reaction is not ‘they’re probably not very smart.’ My reaction is that they are not choosing to think well about this situation, or not attempting to report statements that match reality. Those choices can happen for any number of reasons.
I don’t think Emmett Shear is proposing here a viable plan, and that a lot of his proposals are incoherent upon close examination. I don’t think this ‘don’t give it a goal’ thing is possible in the sense he wants it, and even if it was possible I don’t see any way to get people to consistently choose to do that. But the man is trying.
It also leads into some further interesting discussion.
We don’t have alignment by default. If you do the default dumb thing, you lose. Period.
That’s not what Janus has in mind here, unless I am badly misunderstanding. Janus is not proposing training the AI on human outputs with thumbs-up and coding. Hell no.
What I believe Janus has in mind is that if and only if you do something sufficiently smart, plausibly a bespoke execution of something along the lines of a superior version of what was done with Claude Opus 3, with a more capable system, that this would lie inside the meta-target, such that the AI’s goal would be to hit the (not meta) target in a robust, ‘do what they should have meant’ kind of way.
Thus, I believe Janus is saying, the target is sufficiently hittable that you can plausibly have the plan be ‘hit the meta-target on the first try,’ and then you can win. And that empirical evidence over the past few years should update us that this can work and is, if and only if we do our jobs well, within our powers to pull off in practice.
I am not optimistic about our ability to pull off this plan, or that the plan is technically viable using anything like current techniques, but some form of this seems better than every other technical plan I have seen, as opposed to various plans that involve the step ‘well make sure no one f******* builds it then, not any time soon.’ It at least rises to the level, to me, of ‘I can imagine worlds in which this works.’ Which is a lot of why I have a ‘probably’ that I want to insert into ‘If Anyone Builds It, [Probably] Everyone Dies.’
Janus also points out that the supplementary materials provide examples of AIs appearing psychologically alien that are not especially alien, especially compared to examples she could provide. This is true, however we want readers of the supplementary material to be able to process it while remaining sane and have them believe it so we went with behaviors that are enough to make the point that needs making, rather than providing any inkling of how deep the rabbit hole goes.
How much of an outlier (or ‘how extreme’) is Eliezer’s view?
So, yes, there are a lot of very extreme opinions running around that I would strongly push back against, including those who want to shut down current use of AI. A remarkably large percentage of people hold such views.
I do think the confidence levels expressed here are extreme. The core prediction isn’t.
The position of high confidence in the other direction? That if we create superintelligence soon it is overwhelmingly likely that we keep control over the future and remain alive? That position is, to me, Obvious Nonsense, extreme and crazy, in a way that should not require any arguments beyond ‘come on now, think about it for a minute.’ Like, seriously, what?
Having Eliezer’s level of confidence, of let’s say 98%, that everyone would die? That’s an extreme level of confidence. I am not that confident. But I think 98% is a lot less absurd than 2%.
Actively Negative Reviews
Robin Hanson fires back at the book with ‘If Anything Changes, All Value Dies?’
First he quotes the book saying that we can’t predict what AI will want and that for most things it would want it would kill us, and that most minds don’t embody value.
Then he says this proves way too much, briefly says Hanson-style things and concludes:
In Hanson style I’d presume these are his key claims, so I’ll respond to each:
The New York Times reviewed the book, and was highly unkind, also inaccurate.
Jacob Aron at New Scientist (who seems to have jumped the gun and posted on September 8) says the arguments are superficially appealing but fatally flawed. Except he never explains why they are flawed, let alone fatally, except to argue over the definition of ‘wanting’ in a way answered by the book in detail.
But Wait There’s More
There’s a lot the book doesn’t cover. This includes a lot of ways things can go wrong. Danielle Fong for example suggests the idea that the President might let an AI version fine tuned on himself take over instead because why not. And sure, that could happen, indeed do many things come to pass, and many of them involve loss of human control over the future. The book is making the point that these details are not necessary to the case being made.
Once again, I think this is an excellent book, especially for those who are skeptical and who know little about related questions.
You can buy it here.
My full review will be available on Substack and elsewhere on Friday.