This is a valuable discussion to have, but I believe Tsvi has not raised or focused on the strongest arguments. For context, like Tsvi, I don't understand why people seem to be so confident of short timelines. However (though I did not read everything, and honestly I think this was justified since the conversation eventually seems to cycle and become unproductive) I generally found Abram's arguments more persuasive and I seem to consider short timelines much more plausible than Tsvi does.
I agree that "originality" / "creativity" in models is something we want to watch, but I think Tsvi fails to raise to the strongest argument that gets at this: LLMs are really, really bad at agency. Like, when it comes to the general category of "knowing stuff" and even "reasoning stuff out" there can be some argument around whether LLMs have passed through undergrad to grad student level, and whether this is really crystalized or fluid intelligence. But we're interested in ASI here. ASI has to win at the category we might call "doing stuff." Obviously this is a bit of a loose concept, but the situation here is MUCH more clear cut.
Claude cannot run a vending machine business without making wildly t...
Anyway, this is my crux. If we start to see competent agentic behavior I will buy into the short timelines view at 75% +
Seems good to flesh out what you mean by this if it's such a big crux. Ideally, you'd be able to flesh this out in such a way that bad vision (a key problem for games like pokemon) and poor motivation/adversarial-robustness (a key problem for vending claude because it would sort of knowingly make bad financial decisions) aren't highlighted.
Would this count as competent agentic behavior?
The AI often successfully completes messy software engineering tasks which require 1 week of work for a skilled human and which require checking back in with the person who specified the task to resolve ambiguities. The way the AI completes these tasks involves doing a bunch of debugging and iteration (though perhaps less than a human would do).
But people have been talking about LLM agents for years, and I’d be shocked if the frontier labs weren’t trying? Like, if that worked out of the box, we would know by now (?).
Agentic (tool-using) RLVR only started working in late 2024, with o3 the first proper tool-using reasoning LLM prototype. From how it all looks (rickety and failing in weird ways), it'll take another pretraining scale-up to get enough redundant reliability for some noise to fall away, and thus to get a better look at the implied capabilities. Also the development of environments for agentic RLVR only seems to be starting to ramp this year, and GB200 NVL72s that are significantly more efficient for RLVR on large models are only now starting to get online in large quantities.
So I expect that only 2026 LLMs trained with agentic RLVR will give a first reasonable glimpse of what this method gets us, the shape of its limitations, and only in 2027 we'll get a picture overdetermined by essential capabilities of the method rather than by contingent early-days issues. (In the worlds where it ends up below AGI in 2027, and also where nothing else works too well before that.)
Thanks, this is a really interesting conversation to read!
One thing I have not seen discussed much from either of these viewpoints (or maybe it is there and I just missed it) is how rare frontier-expanding intelligence is among humans, and what that means for AI. Among humans, if you want to raise someone, it's going to cost you something like 20-25 years and $2-500k. If you want to train a single scientist, on average you're going to have to do this about a few hundred to a thousand times. If you want to create a scientist in a specific field, much more than that. If you want to create the specific scientist in a specific field who is going to be able to noticeably advance that field's frontier, well, you might need to raise a billion humans before that happens, given the way we generally train humans.
If I went out in public and said, "Ok, based on this, in order to solve quantum gravity we'll need to spend at least a quadrillion dollars on education" the responses (other than rightly ignoring me) would be a mix of "That's an absurd claim" and "We're obviously never going to do that," when in fact that's just the default societal path viewed from another angle.
But, in this, ...
how rare frontier-expanding intelligence is among humans,
On my view, all human children (except in extreme cases, e.g. born without a brain) have this type of intelligence. Children create their conceptual worlds originarily. It's not literally frontier-expanding because the low-hanging fruit have been picked, but it's roughly the same mechanism.
Maybe this is a matter of shots-on-goal, as much as anything else, and better methods and insights are mostly reducing the number of shots on goal needed to superhuman rates rather than expanding the space of possibilities those shots can access.
Yeah but drawing from the human distribution is very different from drawing from the LP25 distribution. Humans all have the core mechanisms, and then you're selecting over variation in genetic and developmental brain health / inclination towards certain kinds of thinking / life circumstances enabling thinking / etc. For LP25, you're mostly sampling from a very narrow range of Architectures, probably none of which are generally intelligent.
So technically you could set up your laptop to generate a literally random python script and run it every 5 minutes. Eventually this would create an AGI, yo...
I have a couple things to add here to the conversation that I think will help:
I do not think that Noosphere’s comment did not contain an argument. The rest of the comment after the passage you cited tries to lay out a model for why continual learning and long-term memory might be the only remaining bottlenecks. Perhaps you think that this argument is very bad, but it is an argument, and I did not think that your reply to it was helpful for the discussion.
First, I don't buy Mateusz' conclusion from the whack-a-mole analogy. AI safety is hard because, once AIs are superintelligent, the first problem you don't catch can kill you. AI capability research is relatively easy because when you fail, you can try again. If AI safety is like a game of whack-a-mole where you lose the first time you miss, AI capabilities is like whack-a-mole with infinite retries.
I don't think the difference between "first problem you don't catch can kill you" and "when you fail, you can try again" is relevant here.
The thing I had in mi...
So I feel like Tsvi is actually right about a bunch of stuff but that his timelines are still way too long. I think of there as being stronger selection, on australopithecines and precursors, for bipedalism during interglacial periods, because it was hotter and bipedalism reduces the solar cross-section, and this is totally consistent with this not being enough/the right kind of selection over several interglacial periods to cause evolution to cough up a human. But if there had been different path dependencies, you could imagine a world where enough consec...
Tertiarily relevant annoyed rant on terminology:
I will persist in using "AGI" to describe the merely-quite-general AI of today, and use "ASI" for the really dangerous thing that can do almost anything better than humans can, unless you'd prefer to coordinate on some other terminology.
I don't really like referring to The Thing as "ASI" (although I do it too), because I foresee us needing to rename it from that to "AGSI" eventually, same way we had to move from AI to AGI.
Specifically: I expect that AGI labs might start training their models to be superhuman ...
Thanks for the efforts. Modeling good discourse around complicated subjects is hard and valuable.
Didn't know about enthymemes, cool concept.
1% chances of wiping all humans makes the current people working on ai worse than genocidal dictators and I dislike the appeasement of so many people trying to stay in good graces with them on the off chance of influencing them anyway. I think their behavior is clearly sociopathic if that term is to have any meaning at all and the only influence anyone has on them is simulated by them for strategic purposes. They are strongly following power gradients.
...During the live debate Tsvi linked to, TJ (an attendee of the event) referred to the modern LLM paradigm providing a way to take the deductive closure of human knowledge: LLMs can memorize all of existing human knowledge, and can leverage chain-of-thought reasoning to combine that knowledge iteratively, making new conclusions. RLVF might hit limits, here, but more innovative techniques might push past those limits to achieve something like the "deductive closure of human knowledge": all conclusions which can be inferred by some combination of existing know
2.1: This doesn't appear to follow from the previous two steps. EG, is a similar argument supposed to establish that, a priori, bridges are a long way off? This seems like a very loose and unreliable form of argument, generally speaking.
It seems fine to me; bridges were a long way off at most times at which bridges didn't exist! (What wouldn't be fine is continuing to make the a priori argument once there is evidence that we have many of the ideas.)
I'm surprised to see zero mentions of AlphaEvolve. AlphaEvolve generated novel solutions to math problems, "novel" in the "there are no records of any human ever proposing those specific solutions" sense. Of course, LLMs didn't generate them unprompted, humans had to do a lot of scaffolding. And it was for problems where it's easy to verify that the solution is correct; "low messiness" problems if you will. Still, this means that LLMs can generate novel solutions, which seems like a crux for "Can we get to AGI just by incrementally improving LLMs?".