That's not much of a proxy. I'm relying on my subjective impressions from many reports. A more precise phrasing of my claim is that I've seen numerous reports of what I consider to be open contempt for the rule of law among elected officials, but judges in newsworthy cases have almost always looked like they're trying to take the law seriously.
Some of my impressions come from a private mailing list where conservative lawyers have been expressing dismay at the Trump administration's lack of interest in whether their actions could plausibly be defended in a court.
Did you know that Deng approved the 1989 crackdown on Tiananmen protesters
Yes, I'm aware that he did a few things that I consider evil. Wanting to keep his party in power is common enough among politicians that it's not much evidence of psychopathy. His overall attitude toward independent thought was a least no worse than average for a political leader.
A lot of what I have in mind is that Deng allowed more freedom than can readily explained by his self-interest, and Xi seems more Maoist than Deng.
But I wouldn't be surprised if you have better information about their personalities than do I.
A darker interpretation is that the (subconscious, but more real or substantial in some sense) goals of nearly all humans are to gain power and status, and utopian ideologies are merely a tool for achieving this.
The ideologies are partly a tool for that, but they have more effects on the wielder than a mere tool does. My biggest piece of evidence for that is the mostly peaceful collapse of the Soviet Union. I was quite surprised that the leaders didn't use more force to suppress dissent.
I am also somewhat dissatisfied with the basin of attraction metaphor, but for a slightly different reason.
I am concerned that an AI that functions as mostly corrigible in environments that resemble the training environment will be less corrigible when the environment changes significantly.
I'm guessing that a better metaphor would be based on evolutionary pressures. That would emphasize both the uncertainties about any given change, and the sensitivity to out-of-distribution environments.
Maybe a metaphor about how cats are sometimes selected for being friendly to humans? Or the forces that led to the peacock's tail?
Corrigibility would clearly be a nice property
Thinking of it as "a property" will mislead you about how Max's strategy works. It needs to become the AI's only top-level goal in order to work as Max imagines.
It sure looks like AI growers know how to instill some goals in AIs. I'm confused as to why you think they don't. Maybe you're missing the part where the shards that want corrigibility are working to overcome any conflicting shards?
I find it quite realistic that the AI growers would believe at the end of Red Heart that they probably had succeeded (I'll guess that they ended up 80% confident?). That doesn't tell us what probability we should put on it. I'm sure that in that situation Eliezer would still believe that the AI is likely not corrigible.
I don’t know what year the novel is actually set in,
It's an alternate timeline where AI capabilities have progressed faster than ours, likely by a couple of years.
Note this Manifold market on when the audiobook is released.
Novice investor participation is nowhere near what it was at the 2000 dot com peak. Current conditions look more like 1998. A bubble is probably coming, but there's lots of room still for increased novice enthusiasm.
you can't just train your ASI for corrigibility because it will sit and do nothing
I'm confused. That doesn't sound like what Max means by corrigibility. A corrigible ASI would respond to requests from its principal(s) as a subgoal of being corrigible, rather than just sit and do nothing.
Or did you mean that you need to do some next-token training in order to get it to be smart enough for corrigibility training to be feasible? And that next-token training conflicts with corrigibility?
Nothing importantly bearish happened in that month other than bullish deals
What happened that made a bunch of people more bearish is that AI stocks went up a good deal, especially some of the lesser known ones.
I'm unsure what exact time period you're talking about, but here are some of the more interesting changes between Aug 29 and Oct 15:
IREN +157%
CLSK +145%
APLD +136%
INOD +118%
NBIS +75%
MU +61%
AMD +47%
If I thought AI was mostly hype, those kinds of near-panic buying would have convinced me to change my mind from "I don't know" to "some of those are almost certainly in a bubble". (Given my actual beliefs, I'm still quite bullish on MU, and weakly bullish on half of the others).
Business models will change significantly. I speculated here about one likely change. Robotics-related business models will probably become important by 2030.