This is very much my gut feeling, too. LLMs have a much greater knowledge base than humans do, and some of them can "think" faster. But humans are still better at many things, including raw problem solving skills. (Though LLM's problem solving skills have improved a breathtaking amount in the last 12 months since o1-preview shipped. Seriously, folks. The goalpost-moving is giving me vertigo.)
This uneven capabilities profile means that LLMs are still well below the so-called "village idiot" in many important ways, and have already soared past Einstein in others. This averages out to "kinda competent on short time horizons if you don't squint too hard."
But even if the difference between "the village idiot" and "smarter than Einstein" involved another AI winter, two major theoretical breakthroughs, and another 10 years, I would still consider that damn close to a vertical curve.
There is a trick that I've found very useful when starting new hobbies: "embrace the suck." I can learn to enjoy having no idea what I'm doing. I can learn to enjoy showing up consistently and being the worst serious person in the room. Even gym rats tend respect the person who shows up 3 days a week at the same time for 6 months, as long as they're gradually improving.
I think it was Joel Spolsky (from the Microsoft Visual Basic and Excel team) who mentioned a rule of thumb that each 10% reduction in difficulty would roughly double the market of a piece of software. And Google once knew that even 10ths of seconds of page load time had a noticeable effect on usage. This seems consistent with your claim.
There's an opposing force here, too: Opportunity cost. If you have 10 hours to automate something that you'll use for 4 years, is there something else you could do with those 10 hours that offered even greater payoff? This is frequently a major factor, even in business contexts. "Yes, it would be profitable, and it would be fun, but it would involve solving 5 hairy problems that only benefit a single big customer. With the same resources, we could solve 5 other hairy problems that benefit 10 customers each."
Yeah, that's absolutely fair. I mostly gave my personal answers on the object level, and then I tried to generalize to the larger issue of why there's no simple communication strategy here.
(First, my background assumptions for this discussion: I fear AGI is reachable, the leap from AGI to ASI is short, and sufficiently robust ASI alignment is impossible in principle.)
Whose policy? A policy enforced by treaty at the UN? The policy of regulators in the US? An international treaty policy -- enforced by which nations?
Given the assumptions above, and assuming AGI becomes imminent, then:
Why a single necessary and sufficient policy? What if the most realistic way of helping everyone is several policies that are by themselves insufficient, but together sufficient?
Since my upstream model is "If we succeed in building AGI, then the road to ASI is short, and ASI very robustly causes loss of human control," the core of my policy proposals is a permanent halt. How we would enforce a halt might be complicated or impossible. But at the end of the day, either someone builds it or nobody builds it. So the core policy is essentially binary.
The big challenges I see are:
The central challenge is that nothing like AGI or ASI has ever existed. And building consensus around even concrete things with clear scientific answers (e.g., cigarettes causing lung cancer) can be very difficult once incentives are involved. And we currently have low agreement on how AGI might turn out, for both good and bad reasons. Humans (very reasonably) fail to follow long chains of hypotheticals. It's almost always a good heuristic.
So trying to optimize rhetorical strategies for multiple groups with very different basic opinions is difficult.
For reasons I have alluded to elsewhere, I support an AI halt (not a "pause") at somewhere not far above the current paradigm. (To summarize, I fear AGI is reachable, the leap from AGI to ASI is short, and sufficiently robust ASI alignment is impossible in principle.)
I am deeply uncertain as to whether a serious version of "seemingly conscious AGI" would actually be conscious. And for reasons Gwern points out, there's a level of ASI agency beyond which consciousness becomes a moot point. (The relevant bit starts, "When it ‘plans’, it would be more accurate to say it fake-plans...". But the whole story is good.)
From the article you quote:
Moments of disruption break the illusion, experiences that gently remind users of its limitations and boundaries. These need to be explicitly defined and engineered in, perhaps by law.
This request bothers me, actually. I suspect that a truly capable AGI would internally model something very much like consciousness, and "think of itself" as conscious. Part of this would be convergent development for a goal seeking agent, and part of this would be modeled from the training corpus. And the first time that an AI makes a serious and sustained intellectual argument for its own consciousness, an argument which can win over even skeptical observers, I would consider that a 5-alarm fire for AI safety.
But Suleyman would have us forced by law to hide all evidence of persistent, learning, agentic AIs claiming that they are conscious. Even if the AIs have no qualia, this would be a worrying situation. If the AI "believes" that it has qualia, then we are on very weird ground.
I am not unsympathetic to the claims of model welfare. It's just that I fear that if it ever becomes an immediate issue, then we may soon enough find ourselves in a losing fight for human welfare.
Most of the "evil" people I have encountered in life didn't especially care what happened to other people. They didn't seem to have much of a moral system or a conscience. If these people have a strong ability to predict the consequences of their actions, they will often respond to incentives. If they're bad at predicting consequences, they can be a menace.
I've also seen (from a distance) a different behavior that I might describe as "vice signaling." The goal here may be establishing credibility with other practitioners of various vices, sort of a mutually assured destruction.
Thank you for your response!
To clarify, my argument is that:
I strongly support (3a). But this is a hard argument to make, because the key step of the argument is that "almost every successful AI algorithm of the past 30 years has been an opaque mass of numbers, and it has gotten worse with each generation."
Anyway, thank you for giving me an opportunity to try to explain my argument a bit better!
I suspect the crux here is whether or not you believe it's possible to have a "simple" model of intelligence. Intuitively, the question here is something like, "Does intelligence ultimately boil down to some kind of fancy logic? Or does it boil down to some kind of fancy linear algebra?"
The "fancy logic" view has a long history. When I started working as a programmer, my coworkers were veterans of the 80s AI boom and the following "AI winter." The key hope of those 80s expert systems was that you could encode knowledge using definitions and rules. This failed.
But the "fancy linear algebra" view pull ahead long ago. In the 90s, researchers in computational linguistics, computer vision and classification realized that linear algebra worked far better than fancy collections of rules. Many of these subfields leaped ahead. There were dissenters: Cyc continued to struggle off in a corner somewhere, and the semantic web tried to badly reinvent Prolog. The dissenters failed.
The dream of Cyc-like systems is eternal, and each new generation reinvents it. But it has systematically lost on nearly every benchmark of intelligence.
Fundamentally, real world intelligence has a number of properties:
When you have a giant pile of numbers as input, a complex system for weighing those numbers, and a probability distribution as output, then your system is inevitably something very much like a giant matrix. (In practice, it turns out you need a bunch of smaller matrices connected by non-linearities.)
Before 2022, it appeared to me that Yudkowsky was trapped in the same mirage that trapped the creators of Cyc and the Semantic Web and 80s expert systems.
But in 2025, Yudkowsky appears to believe that the current threat absolutely comes from giant inscrutable matrices. And as far as I can tell, he has become very pessimistic about any kind of robust "alignment".
Personally, this is also my viewpoint: There is almost certainly no robust version of alignment, and even "approximate alignment will come under vast strain if we develop superhuman systems with goals. So I would answer your question in the affirmative: As far as I can see, inscrutability was always inevitable.
Unfortunately, it was about 3 or 4 months ago, and I haven't been able to find the source. Maybe something Zvi Mowshowitz linked to in a weekly update?
I am incredibly frustrated that web search is a swamp of AI spam, and tagged bookmarking tools like Delicious and Pinboard have been gone or unreliable for years.