An important point is that I frame this as creating insights at least semi-reliably, rather than being able to create any insights at all, because LLMs have already created insights, so it can't be due to a fundamental incapacity, but rather a practical incapability.
To be clear, practical incapability can be nearly as bad/good as fundamental incapability, so this nitpick doesn't matter too much.
Links below:
https://www.lesswrong.com/posts/GADJFwHzNZKg2Ndti/have-llms-generated-novel-insights#H8a4ub3vura8ttuPN (not too impressive, but definitely counts here)
https://www.lesswrong.com/posts/vvgND6aLjuDR6QzDF/my-model-of-what-is-going-on-with-llms?commentId=jLahLy4SRyA4Fuyc2 (an ancedote that an LLM managed to solve a step in synthesizing a chemical, and gave a plausible causal story for why it worked, and notably even when searching the internet/asking around, there still wasn't any discussion, implying that this wasn't in it's training data)
I consider these examples to be pretty weak. I was arguing that LLMs had not made novel insights (which inspired Abram to create the question), and I'm still surprised no has mentioned anything more convincing months later. It seems like just stumbling around blindly in idea-space would lead to more original ideas.
The math result is just filling in a step that the researchers could have done themselves and doesn't seem to go beyond standard techniques. Considering the IMO results one would expect better than this by now (I guess we'll see soon).
The chemistry one... even if it really didn't appear anywhere on the internet, maybe LLMs just suggest a lot of ideas and we only hear about the ones that worked. If I heard A LOT of examples like this, it would be convincing. But one just seems like getting lucky.
The key point is that I only need one example, so it doesn't matter whether or not the LLM simply got it lucky.
However, I do agree that if I assume that the LLM got lucky, this would be very bad in a practical sense, as this would make such insights take a long time to repeat by pure chance.
A key worldview here is that I think it's rarely productive to debate whether or not AI has a fundamentally incapable at a certain task, because if we allow arbitrary resources, and especially arbitrary designs subject only to the loosest requirements of existence, we can solve any problem with AI/create ASI trivially, and the actual question is whether a certain approach can actually do the task in question with limits on resources.
Another way to say it is that you should always focus on the quantitative questions over the qualitative questions.
In general, overfocusing on fundamental barriers that remain even with infinite compute and not focusing on the finite compute case is one of the biggest barriers to AI progress/good AI discourse over the years (though admittedly for theoretical analysis like computational complexity part of the issue is proving stuff about computation is quite hard, and a lot of the issue is we believe that cryptography exists combined with the fact that we are quite bad at opening up the black-box of a computer, which is necessary in order to solve a lot of problems in computational complexity)
I'll quote from a recent comment here:
My take on how recursion theory failed to be relevant for today's AI is that it turned out that what a machine could do if unconstrained basically didn't matter at all, and in particular it basically didn't matter what limits an ideal machine could do, because once we actually impose constraints that force computation to use very limited amounts of resources, we get a non-trivial theory and importantly all of the difficulty of explaining how humans do stuff lies here.
There was too much focus on "could a machine do something at all?" and not enough focus on "what could a machine with severe limitations could do?"
The reason is that in a sense, it is trivial to solve any problem with a machine if I'm allowed zero constraints except that the machine has to exist in a mathematical sense.
A good example of this is the paper on A Universal Hypercomputer, which shows how absurdly powerful computation can be if you are truly unconstrained:
Or davidad's comment on how every optimization problem is trivial under worst-case scenarios:
Another smaller point is I tend to have a somewhat more continuous model of how much creativity AIs currently have, and while there's a point to be made that sufficiently terrible/bad creativity is in practice not different from having 0 creativity, and it might turn out to be too inefficient to scale creativity because in-context learning scales poorly due to context windows not being a sufficient replacement for a long-term memory/continual thinking just being way more efficient than pure mesa-optimization, but I do claim that literally 0 creativity or insights is not very likely based on priors.
Really, this could be fixed by simply switching the question from "Are LLMs creative/have insights at all?" to "Is there any way to get an LLM to autonomously generate and recognize genuine scientific insights at least at the same rate as human scientists?".
And I'd easily agree that so far, LLMs have not actually done this, especially if we require it to be as good as the best human scientists.
To be clear, I'm not smuggling in any claim that would imply that since LLMs can do a non-zero amount of insights, that this reliably scales with the current architecture into enough insights to replace AI researchers and start AI takeoff by say 2030.
This post from Gwern tackles a question that I suspect could become very relevant for AI automating AI research (and jobs more generally), which is why don't current AIs produce frontier-expansion/insights semi-reliability beyond their training data, and what might be necessary for AI to create insights at least semi-reliably.
My takeaways are at the end.
An important point is that I frame this as creating insights at least semi-reliably, rather than being able to create any insights at all, because LLMs have already created insights, so it can't be due to a fundamental incapacity, but rather a practical incapability.
To be clear, practical incapability can be nearly as bad/good as fundamental incapability, so this nitpick doesn't matter too much.
Links below:
https://www.lesswrong.com/posts/GADJFwHzNZKg2Ndti/have-llms-generated-novel-insights#H8a4ub3vura8ttuPN (not too impressive, but definitely counts here)
https://www.lesswrong.com/posts/vvgND6aLjuDR6QzDF/my-model-of-what-is-going-on-with-llms?commentId=jLahLy4SRyA4Fuyc2 (an ancedote that an LLM managed to solve a step in synthesizing a chemical, and gave a plausible causal story for why it worked, and notably even when searching the internet/asking around, there still wasn't any discussion, implying that this wasn't in it's training data)
(Fun fact, this post/idea by Gwern itself was born out of an eruption of insight).
Quotes below:
Continual Learning:
Continual Thinking:
Hypothesis: Day-Dreaming Loop
Day-Dreaming Loop
LLM Analogy
Obstacles and Open Questions
Gwern's Implications
My Takeaways
If we grant the assumption that something like an insight-generator like the default mode network is necessary for AI research to be automated/jobs to be replaced by AI, I think there are a couple of very important implications that follow:
Number 1 is that the most capable AIs will be used internally, and by default we should expect pretty large divergences in capability between consumer AIs and in-house/researcher AIs, because only AI companies would want to spend the expense to get insights from AIs semi-reliably, and as Gwern said, it's an excellent moat for profit/something that differentiates them from competitors.
This means that takeoff will be far more local and internal to the company than people like Robin Hanson/Nora Belrose thought, and the one big model will likely win out over many smaller models that can't run a default mode network.
This also means that AI governance is going to be far more difficult, because even if AIs creating insights that led to automating AI research jobs that then lead to automating everything else, the consumer won't see this/average companies won't use this and instead use the optimized for cheap/fast with the good solutions already expensively made by AI companies.
This will make any discourse around AI very bad, because lots of people will keep arguing that AI can't make new insights/actually be AGI, even if that has in fact happened, and with poor enough transparency/explanation, this could easily keep the government in the dark, meaning AI governance never gets off the ground.
We also should expect a lot more internal tech development than external tech development by default, meaning that the more classic AI/human takeover stories become more plausible (especially so if we add in Drexlerian nanotech).
This also makes me more pessimistic on d/acc tech being developed in the event of extreme misuse/misalignment, because society won't change nearly as much as the technology.
Number 2 is that this should increase your probability that something like General Purpose Search is going to exist in AIs, because internal models will pay more of a compute tax to already generate insights for AI companies, so using more compute to run a General Purpose Search is already baked in.
Also means that mesa-optimizers/AIs in general will likely be more coherent/persistent than you might think.
Number 3 is that this makes the takeoff slower, because AIs will have to be more expensive in order to become cheap, and you can't get to the cheap/ultra-efficient ASI right at the start, you have to pay more compute to actually get to the cheap and fast ASIs of the future.
Number 4 is that the parallelizability of the insight generator means that it's much more feasible for AI to speed up the process of getting insights than you may initially expect, and in domains which are strongly bottlenecked by insight, this can mean unusually fast progress can happen in these domains.
How strongly you believe scientific fields are bottlenecked by insights relative to other variables like experimentation will determine a lot about how fast AI in general makes progress.
Number 5 is also related to the parallelizability point, which is that it's probably feasible for an AI to have more insights in a faster period than humans do, which makes progress even faster in areas where the strongest bottleneck is insight/simulation.
But that's just my thoughts on the matter, and I'd like to see more discussion of this point.