This seems wise. The reception of the book in the community has been rather Why Our Kind Can't Cooperate, as someone whom I forget linked. The addiction to hashing-out-object-level-correctness-on-every-point-of-factual-disagreement and insistence on "everything must be simulacrum level 0 all the time"... well, it's not particularly conducive to getting things done in the real world.
I'm not suggesting we become propagandists, but I think pretty much every x-risk-worried Rat who disliked the book because e.g. the evolution analogy doesn't work, they would have preferred a different flavor of sci-fi story, or the book should have been longer, or it should have been shorter, or it should have proposed my favorite secret plan for averting doom, or it should have contained draft legislation at the back... if they would endorse such a statement, I think that (metaphorically) there should be an all-caps disclaimer that reads something like "TO BE CLEAR AI IS STILL ON TRACK TO KILL EVERYONE YOU LOVE; YOU SHOULD BE ALARMED ABOUT THIS AND TELLING PEOPLE IN NO UNCERTAIN TERMS THAT YOU HAVE FAR, FAR MORE IN COMMON WITH YUDKOWSKY AND SOARES THAN YOU DO WITH THE LOBBYISTS OF META, WHO ABSENT COORDINATION BY PEOPLE ON HUMANITY'S SIDE ARE LIABLE TO WIN THIS FIGHT, SO COORDINATE WE MUST" every couple of paragraphs.
I don't mean to say that the time for words and analysis is over. It isn't. But the time for action has begun, and words are a form of action. That's what's missing, is the words-of-action. It's a missing mood. Parable (which, yes, I have learned some people find really annoying):
A pale, frightened prisoner of war returns to the barracks, where he tells his friend: "Hey man, I heard the guards talking, and I think they're gonna take us out, make us a dig a ditch, and then shoot us in the back. This will happen at dawn on Thursday."
The friend snorts, "Why would they shoot us in the back? That's incredibly stupid. Obviously they'll shoot us in the head; it's more reliable. And do they really need for us to dig a ditch first? I think they'll just leave us to the jackals. Besides, the Thursday thing seems over-confident. Plans change around here, and it seems more logical for it to happen right before the new round of prisoners comes in, which is typically Saturday, so they could reasonably shoot us Friday. Are you sure you heard Thursday?"
The second prisoner is making some good points. He is also, obviously, off his rocker.
There are two steelmen I can think of here. One is "We must never abandon this relentless commitment to precise truth. All we say, whether to each other or to the outside world, must be thoroughly vetted for its precise truthfulness." To which my reply is: how's that been working out for us so far? Again, I don't suggest we turn to outright lying like David Sacks, Perry Metzger, Sam Altman, and all the other rogues. But would it kill us to be the least bit strategic or rhetorical? Politics is the mind-killer, sure. But ASI is the planet-killer, and politics is the ASI-[possibility-thereof-]killer, so I am willing to let my mind take a few stray bullets.
The second is "No, the problems I have with the book are things that will critically undermine its rhetorical effectiveness. I know the heart of the median American voter, and she's really gonna hate this evolution analogy." To which I say, "This may be so. The confidence and negativity with which you have expressed this disagreement are wholly unwarranted."
Let's win, y'all. We can win without sacrificing style and integrity. It might require everyone to sacrifice a bit of personal pride, a bit of delight-in-one's-own-cleverness. I'm not saying keep objections to yourself. I am saying, keep your eye on the fucking ball. The ball is not "being right," the ball is survival.
One is "We must never abandon this relentless commitment to precise truth. All we say, whether to each other or to the outside world, must be thoroughly vetted for its precise truthfulness." To which my reply is: how's that been working out for us so far?
[...]
We can win without sacrificing style and integrity.
But you just did propose sacrificing our integrity: specifically, the integrity of our relentless commitment to precise truth. It was two paragraphs ago. The text is right there. We can see it. Do you expect us not to notice?
To be clear, in this comment, I'm not even arguing that you're wrong. Given the situation, maybe sacrificing the integrity of our relentless commitment to precise truth is exactly what's needed!
But you can't seriously expect people not to notice, right? You are including the costs of people noticing as part of your consequentialist decision calculus, right?
I think that (metaphorically) there should be an all-caps disclaimer that reads something like "TO BE CLEAR AI IS STILL ON TRACK TO KILL EVERYONE YOU LOVE; YOU SHOULD BE ALARMED ABOUT THIS AND TELLING PEOPLE IN NO UNCERTAIN TERMS THAT YOU HAVE FAR, FAR MORE IN COMMON WITH YUDKOWSKY AND SOARES THAN YOU DO WITH THE LOBBYISTS OF META, WHO ABSENT COORDINATION BY PEOPLE ON HUMANITY'S SIDE ARE LIABLE TO WIN THIS FIGHT, SO COORDINATE WE MUST" every couple of paragraphs.
Yeah, I kind of regret not prefacing my pseudo-review with something like this. I was generally writing it from the mindset of "obviously the book is entirely correct and I'm only reviewing the presentation", and my assumption was that trying to "sell it" to LW users was preaching to the choir (I would've strongly endorsed it if I had a big mainstream audience, or even if I were making a top-level LW post). But that does feel like part of the our-kind-can't-cooperate pattern now.
To be fair, if you are reading reviews of IABIED on LessWrong, you are probably already pretty convinced of AI risk being a pretty big deal. But probably good to keep in mind the general vibe that we're all on the same team
guys I cut this but honestly do u consider riding a motorcycle to be within your risk budget? would you be excited or discouraging if a friend or loved one started riding a motorcycle? do you consider building superintellgience to be within your risk budget?
I understand that when a person feels a lot is on the line it is often hard for that person to not come across as sanctimonious. Maybe it’s unfair of me, but that is how this comes across to me. Eg “people who allegedly care”.
Death with Dignity:
>Q2: I have a clever scheme for saving the world! I should act as if I believe it will work and save everyone, right, even if there's arguments that it's almost certainly misguided and doomed? Because if those arguments are correct and my scheme can't work, we're all dead anyways, right?A: No! That's not dying with dignity! That's stepping sideways out of a mentally uncomfortable world and finding an escape route from unpleasant thoughts!”
This is a good insight about a possible reasoning mistake. Likewise, if more optimistic assumptions about AI are correct, you should not “step sideways” into an imaginary world where MIRI is right about everything “just to be safe”. Whatever problems come with AI need to be solved in the actual world, and in order to do that it is very very important to form good object-level beliefs about the problems
This is a review of the reviews, a meta review if you will, but first a tangent. and then a history lesson. This felt boring and obvious and somewhat annoying to write, which apparently writers say is a good sign to write about the things you think are obvious. I felt like pointing towards a thing I was noticing, like 36 hours ago, which in internet speed means this is somewhat cached. Alas.
Previously, I rode a motorcycle. I rode it for about a year while working on semiconductors until I got a concussion, which slowed me down but did not update me to stop, until it eventually got stolen. The risk in dying from riding a motorcycle for a year is about 1 in 800 depending on the source.
Previously, I sailed across an ocean. I wanted to calibrate towards how dangerous it was. The forums said the probability of dying from a transatlantic crossing is like 1 in 10,000.
Currently, the people I know working on AI are far more well calibrated to the risk of AI than the general public, and even me, and almost all of them I know think there is more than a 5% chance of an AI catastrophe. That is a 1 in 20 chance, which feels recklessly high.
The thing I wanted to point to was the mental gymnastics I observed in peoples book reviews (if im feeling more contentious I might come back and link to some examples) and the way it made me both disappointed and almost not want to say anything.
I think it's virtuous to note disagreements and it's cool to note agreements, but it's also even cooler and virtuous to avoid misleading people by not saying the trees are not there, when getting into the weeds, literally between the trees.
There are a bunch of people who are allegedly trying to change the world here. We all allegedly think lots of stuff is at stake. Building a coalition doesn't look like suppressing disagreements, but it does look like building around the areas of agreement.
If you think there's a 1 in 20 chance it could be so over, it feels to me the part where people are not doing the ‘yes the situation is insane’ even if that is immediately followed up with ‘im more hopeful than them tbc’ is weird.
On the first day of improv classes they teach you to say ‘yes, and’ instead of using ‘no’, which can kill the scene, when you don't know how to respond, or to move things along. So, yes, and -
Now is time for the history lesson. The Shanghai Communiqué..
The Shanghai Communiqué was a joint statement issued by the U.S. and China in 1972, breaking a plus 20 year freeze with no diplomatic relations. The communiqué ended a long period of isolation between the two countries and paved the way towards a formal normalization seven years later.
I think some people critique it for pushing the Taiwan issue down the line, but I like to think about things in the context in which they existed. There was a threat great enough that countries had the incentive to coordinate together.
These negotiations were happening in the shadow of the cold war, and relations with China were also about counter balancing the USSR. The Sino-Soviet split created space for triangular diplomacy, and improved relations with China, could pressure the Soviets to cooperate on a different arms race.
I bring up the communiqué because I think it is cool. It acknowledged disagreements, even clearly laid them out, and used some well worded phrases to get it across the line. But the success of the negotiation is, in my read, attributed to the focus on the areas of agreement. Focusing on disagreement would have sunk it before it even started.
It worries me that people who allegedly care about the future going well, and are also at least 5% concerned that AI is not going to go well, are also squandering opportunities to help wake the world up to the dangers that they themselves are seeing, even if they see them slightly differently from the authors.
That said, I am slightly more hopeful than I was yesterday, and hope to feel further more hopeful in tomorrow.