Gary Marcus now saying AI can't do things it can already do
January 2020, Gary Marcus wrote GPT-2 And The Nature Of Intelligence, demonstrating a bunch of easy problems that GPT-2 couldn’t get right. He concluded these were “a clear sign that it is time to consider investing in different approaches.” Two years later, GPT-3 could get most of these right. Marcus wrote a new list of 15 problems GPT-3 couldn’t solve, concluding “more data makes for a better, more fluent approximation to language; it does not make for trustworthy intelligence.” A year later, GPT-4 could get most of these right. Now he’s gone one step further, and criticised limitations that have already been overcome. Last week Marcus put a series of questions into chatGPT, found mistakes, and concluded AGI is an example of “the madness of crowds”. However, Marcus used the free version, which only includes GPT-4o. That was released in May 2024, an eternity behind the frontier in AI. More importantly, it’s not a reasoning model, which is where most of the recent progress has been. For the huge cost of $20 a month, I have access to GPT-o1 (not the most advanced model OpenAI offers, let alone the best that exists). I asked GPT-o1 the same questions Marcus did and it didn’t make any of the mistakes he spotted. First he asked it: > Make a table of every state in the US, including population, area and median household income, sorted in order of median household income. GPT-4o misses out a bunch of states. GPT-o1 lists all 50 (full transcript). Then he asked for a column added on population density. This also seemed to work fine. He then made a list of Canadian provinces and asked for a column listing how many vowels were in each name. I was running out of patience, so asked the same question about the US states. This also worked: To be clear, there are probably still some mistakes in the data (just as I’d expect from most human assistants). The point is that the errors Marcus identified aren’t showing up. He goes on to correctly point out that agent
This makes sense, just my concern is the 1.25 trend in transistor density is itself driven by exponentially increasing investment into semiconductor research. Going forward, most of that will be via spending on AI chips. So if AI chip spending stops growing, so does R&D spending on increasing transistor density.
I could see an argument though that mobile phones etc. would be able to fund enough investment to keep transistor research growing.