I think context and inability to handle certain specific tasks (especially around computer use details) are holding things back right now more than intelligence.
What does this mean? Insofar as context is like working memory, that seems like a crucial component of "intelligence" in humans. It seems hard to define "intelligence" as distinct from working memory and the ability to handle a wide variety of tasks.
I would guess this is about "getting the right things into context", not "being able to usefully process what is in context". (AI already seems pretty good at the latter, for a broad though not universal set of tasks.)
I'm petty sure that when he talks about the damage of knowledge to intelligence he doesn't mean that the intelligence shouldn't have this knowledge in generation time. Rather, the issue in training - which you may ver well call a skill issue - is that by default local ad hoc explanations create superficial predictive success and get in the way of more general explanations. So the issue isn't having less knowledge, but rather having less early memorization.
best believe in intelligence explosions because you’re already living in one and have been for decades, that’s why GDP grows, this is all continuous with the existing hyper-exponential trend, previous techs also didn’t make GDP go up much, everything was slow diffusion.
This is such a weird and self-contradictory thought. The "existing hyper-exponential trend" is the one that has a vertical asymptote within the few handfuls of years or less. Or used to be, and will be again once "number of minds having ideas" stops being dependent on human reproduction. Yes, previous techs did make GDP growth go up, when those techs meaningfully increased number of minds, number of ideas per mind, or ability to diffuse good ideas faster across minds.
even they are now out in 2029 or so I think
Kokotajlo already claims to have begun working on AI-2032 branch where the timelines are pushed back, or that "we should have some credence on new breakthroughs e.g. neuralese, online learning, whatever. Maybe like 8%/yr? Of a breakthrough that would lead to superhuman coders within a year or two, after being appropriately scaled up and tinkered with."
I have two issues: one is the possibility that CoT-based AIs fail to reach the AGI, and another with the 8%/yr estimate of the chance of the next breakthrough.
Some podcasts are self-recommending on the ‘yep, I’m going to be breaking this one down’ level. This was very clearly one of those. So here we go.
As usual for podcast posts, the baseline bullet points describe key points made, and then the nested statements are my commentary.
If I am quoting directly I use quote marks, otherwise assume paraphrases.
Rather than worry about timestamps, I’ll use YouTube’s section titles, as it’s not that hard to find things via the transcript as needed.
This was a fun one in many places, interesting throughout, frustrating in similar places to where other recent Dwarkesh interviews have been frustrating. It gave me a lot of ideas, some of which might even be good.
AGI Is Still a Decade Away
LLM Cognitive Deficits
RL Is Terrible
How Do Humans Learn?
AGI Will Blend Into 2% GDP Growth
Wait what?
Note: The 2% number doesn’t actually come up until the next section on ASI.
ASI
I won’t go further into the same GDP growth or intelligence explosion arguments I seem to discuss in many Dwarkesh Patel podcast posts. I don’t think Andrej has a defensible position here, in the sense that he is doing some combination of denying the premise of AGI/ASI, not taking into account its implications in some places while acknowledging the same dynamics in others.
Most of all, this echoes the common state of the discourse on such questions, which seems to involve:
I’m fine with those who expect to at first encounter story #2 instead of story #1.
Except it totally, absolutely does not imply #3. Yes, these factors can slow things down, and 10 years are more than 2-5 years, but 10 years is still not that much time, and a continuous transition ends up in the same place, and tacking on some years for diffusion also ends up in the same place. It buys you some time, which we might be able to use well, or we might not, but that’s it.
What about story #4, which to be clear is not Karpathy’s or Patel’s? It’s possible that AI progress stalls out soon and we get a normal technology, but I find it rather unlikely and don’t see why we should expect that. I think that it is quite poor form to treat this as any sort of baseline scenario.
Evolution of Intelligence and Culture
Why Self Driving Took So Long
Future Of Education
Reactions
Peter Wildeford offers his one page summary, which I endorse as a summary.
Sriram Krishnan highlights part of the section on education, which I agree was excellent, and recommends the overall podcast highly.
Andrej Karpathy offered his post-podcast reactions here, including a bunch of distillations, highlights and helpful links.
Here’s his summary on the timelines question:
Those house parties must be crazy, as must his particular slice of Twitter. He has AGI 10 years away and he’s saying that’s 5-10X pessimistic. Do the math.
My slice currently overall has 4-10 year expectations. The AI 2027 crowd has some people modestly shorter, but even they are now out in 2029 or so I think.
That’s how it should work, evidence should move the numbers back and forth, and if you had a very aggressive timeline six months or a year ago recent events should slow your roll. You can say ‘those people were getting ahead of themselves and messed up’ and that’s a reasonable perspective, but I don’t think it was obviously a large mistake given what we knew at the time.
I agree with the second point (with error bars). The first point I would rate as ‘somewhat true.’ Much of the marketing is BS and much of the output is slop, no question, but much of it is not on either front and the models are already extremely helpful to those who use them.
Similarly, the first position here is obviously wrong, and the second position could be right on the substance but has one hell of a Missing Mood, 10-20 years before all jobs get automated is kind of the biggest thing that happened in the history of history even if the process doesn’t kill or diempower us.
It’s so crazy the amount to which vibes can supposedly shift when objectively nothing has happened and even the newly expressed opinions aren’t so different from what everyone was saying before, it’s that now we’re phrasing it as ‘this is long timelines’ as opposed to ‘this is short timelines.’
Yep, I read Altman as ~10 years there as well. Except that Altman was approaching that correctly as ‘quickly, there’s no time’ rather than ‘we have all the time in the world.’
It’s so crazy to think a big tech company would think ‘oops, it’s over, Dwarkesh interviews said so’ and regret or pull back on investment, also yeah it’s weird that Amazon was up 1.6% while AWS was down.
Why would you give Hotz credit for ‘GPT-12 won’t be AGI’ here, when the timeline for GPT-12 (assuming GPT-11 wasn’t AGI, so we’re not accelerating releases yet) is something like 2039? Seems deeply silly. And yet here we are. Similarly, people supposedly ‘look great’ when others echo previous talking points? In my book, you look good based on actual outcomes versus predictions, not when others also predict, unless you are trading the market.
I definitely share the frustration Liron had here:
In short, I don’t think a reasonable extrapolation from above plus AGI is ~2%.
But hey, that’s the way it goes. It’s been a fun one.