Was a philosophy PhD student, left to work at AI Impacts, then Center on Long-Term Risk, then OpenAI. Quit OpenAI due to losing confidence that it would behave responsibly around the time of AGI. Now executive director of the AI Futures Project. I subscribe to Crocker's Rules and am especially interested to hear unsolicited constructive criticism. http://sl4.org/crocker.html
Some of my favorite memes:
(by Rob Wiblin)
(xkcd)
My EA Journey, depicted on the whiteboard at CLR:
(h/t Scott Alexander)
Huh, I think continual learning would be a pretty big deal w.r.t. AI danger. Can you say more about why it wouldn't? Seems like it would dramatically increase horizon lengths for example.
Back when many expected takeoff in 2027 or so, it was pretty reasonable to assume that the probability of a conflict entirely unrelated to AI was low.[3] But the forecasters behind AI 2027 now expect takeoff in the 2030s.
I'm saddened that this is the takeaway from our new model! It seems misleading. Here is a graph of my+Eli's timelines over time:
https://x.com/DKokotajlo/status/1992316620254155028
ChatGPT goes on to say that I outperformed Rick (who also made predictions in 2020 in the spreadsheet.) However, looking over the data briefly, I'm not sure I agree with some of the scores, e.g. are there really robotaxis in 20+ cities now? And drone delivery?
Thank you! I continued your ChatGPT conversation and got the following tidbits of extra info:
For my 2020 predictions, I asked it to summarize my biggest sources of Brier loss:
Top 5 Brier losses from FALSE outcomes (overconfidence)
| Rank | Claim | p (2020) | Loss |
|---|---|---|---|
| 1 | 100+ drone swarms operationally fielded | 0.50 | 0.250 |
| 2 | 10%+ news articles AI-generated | 0.20 | 0.040 |
| 3 | $1 deepfake costs $100+ to detect | 0.35 | 0.123 |
| 4 | Orbital rocket reused 100 times | 0.30 | 0.090 |
| 5 | 90%+ Rio Tinto mines automated | 0.30 | 0.090 |
| Rank | Claim | p (2020) | Loss |
|---|---|---|---|
| 1 | AI wins International Math Olympiad | 0.15 | 0.723 |
| 2 | Neural net writes B-average college essays | 0.20 | 0.640 |
| 3 | Crypto restaurants near Berkeley | 0.20 | 0.640 |
| 4 | Combat drones <50 lbs fielded | 0.50 | 0.250 |
| 5 | AI beats top StarCraft players | 0.70 | 0.090 |
The IMO prediction alone is your single largest loss across all 2020 items.
Interestingly the two biggest hits to Brier score I took (on jan 1 2020) were from underestimating LLMs basically.
Also interestingly, it seems like maybe I took significantly more Brier loss from saying things wouldn't happen, that did happen, then from saying things would happen, that didn't? In other words I was too cautious overall rather than too credulous?
I basically agree with Eli, though I'll say that I don't think the gap between extrapolating METR specifically and AI revenue is huge. I think ideally I'd do some sort of weighted mix of both, which is sorta what I'm doing in my ATC.
Well done for making these predictions and scoring them!
Well done for making these predictions and scoring them!
A major plausible class of worlds in which we don't get superhuman coders by end of 2026 is worlds where the METR trend continues at roughly the same or only slightly greater slope than the slope it had in 2025. Right?
Insofar as zero is significantly smaller than epsilon, yes.
(adding to what bhalstead said: You are welcome to stop by our office sometime to chat about it, we'd love to discuss!)
Personally I agree that 125 years is complete overkill for automation of enough coding for the bottleneck to shift to experiments and research taste. That's a big part of why my parameter is set lower. However I want to think about this more. You can find more of my thoughts here.