(Apologies for the long comment).
I want to make a point about your arguments about the growth of time horizons being superexponential. I think they are generally correct, but I think they need to be downweighted somewhat in the timeline model.
This is how I understand your model:
Our starting point is to take the METR graph and extrapolate it exponentially, as they do, making a guess about what agentic coding time horizon would correspond to the AC milestone.
And then you include adjustments to this extrapolation, some of which are arguments about superexponential growth that don't have anything to do with AI R&D speedups feeding back into themselves. Because you are using a threshold on the METR graph to determine when ACs happen, these arguments about superexponential growth meaningfully affects your prediction of time to ACs.
I consider the casual network to look something like this:
(The METR time horizon level and the level of AI R&D speedup are both driven by the level of effective compute.)
Since we only truly care about AI R&D speedup, we must differentiate between arguments about how fast effective compute will advance or how these advances affect the R&D speedup (which both affect AI R&D speedup and the time to ACs), and arguments about how much effective compute will affect the METR time horizon (which is not what we ultimately care about).
The argument that superexponential growth is implied by infinite time horizons is purely an argument about the relationship between effective compute and the METR time horizon. Whether or not it is true does not change the level of effective compute you need to get ACs. This also applies to your second argument for superexponential growth (that doublings get easier to achieve naturally because less effective compute is needed to jump from 1 month to 4 months than from 1 week to 4 weeks, for example). Again this is only an argument about how increases in effective compute affect the METR time horizon graph, not how fast effective compute is increasing or how increases in effective compute increase AI R&D speedup.
Now this doesn’t mean you have to throw out this entire section of the model. Importantly, it seems like there should be at least some correlation between the relationship between effective compute and the METR time horizon and the relationship between effective compute and the AI R&D speedup. But unless this correlation is 1-1, arguments about superexponential growth that come from the relationship between effective compute and the METR time horizon should be downweighted.
Here’s a toy model to illustrate this better:
Imagine there are four effective compute levels: X1, X2, X3, and X4. X1 is where we are at right now. Let’s say that if METR is exponential in relation to effective compute, we hit Y horizon length at effective compute level X4. On the other hand, if METR is superexponential in relation to effective compute, we hit Y horizon length at capability level X2. Let’s imagine that we thought we would get ACs at effective compute level X4, around where METR was supposed to hit horizon length Y if it were exponential. Suppose we now know that the METR graph is superexponential and will hit Y at X2. How should that affect our expectation of when we will hit ACs? If the correlation between the relationship of effective compute to the METR time horizon and the relationship of effective compute to AI R&D speedup is 1-1, we should update to X2. If there is no correlation, we should keep our estimation at X4. If there is some correlation, maybe we say X3?
The consequences of this are, I think, slightly longer timelines from the model.
2. A working hypothesis: I propose that even though there are multiple possible outcomes, including ones where you, I, and everyone, will very much not be OK, people should live their day to day under the hypothesis they will be OK. Not just because I think that is the most likely outcome, but also because, as I said, it is best not to dwell on the parts of the probability space that are outside your control. This was true for most people during the cold war regarding the possibility of a total nuclear war, and is true now.
I think I disagree slightly with this idea. It feels like a local optimum to just ignore the parts of the probability space where you won't be ok. It feels like a local optimum in the sense that its easier to attain but is inferior to the global optimum. For me, the global optimum (in the sense that this point is harder to attain but better for you and the world), which I think the post you are responding to captures quite well, is to stare The Truth in the face: map the true probability of doom the best you can (whether its high or low), and accept it fully and act and feel appropriately.
If I, my friends, my family, my country, my species, and my planet are going to die, I want to know. I want to know not only so I can do my part to make that not happen, but I also want to know so that I can behave the way I want to on my deathbed. So I can prepare myself to comfort others if one day the doom starts to seem inevitable. So I can be maximally grateful for every second I still have on this planet. So I can live without regrets. So I can do good while I still can.
This is hard. I have spent a lot of time struggling with accepting all of this. However, I think I'm getting there. And I think it has brought me to a much better place, both for myself and my planet, than where I would have ended up if I had chosen to act as if I was going to be ok.
I don't think this global optimum is for everyone. At least not right now. I don't tell most of my friends and family about my perspective on doom. Especially not unprompted. Some people can't help, and some people will suffer significantly if they knew.
But for those of us who can, let's try.
Great post - enjoyable read and connected some concepts I hadn't considered together before.
The first thing that immediately comes to mind when I think about how to act in such an environment is reputation: trying to determine which actors are adversarial based on my (or other's) previous interactions with them. I think I would try this before resorting to the other three tactics.
For example, before online ratings became a thing, chain restaurants had one significant advantage over local restaurants: if you were driving through and needed a place to stop and eat, the chain may have an expected quality that is worse than the local restaurant, but at least you knew you weren't going to be sick because of their reputation.
And now that we have Yelp, I can just look there and see which restaurants consistently get 4-5 stars and no complaints of food poisoning, chain or not.
Of course if a restaurant was really trying to deceive you, it could bury negative reviews in thousands of bot ratings (I actually don't know if this is truly possible given moderation of many rating platforms). And when the environments get more adversarial and the benefits of deceiving you become higher (or the costs of taking the honest route become higher), I imagine a reputation-based filter like this could be easily thwarted. So this feels like an intuitive and lower-cost first line of defense, but not a reason to fully retire the big guns that you talked about in this post.
Thanks for the reply!
It is a very confusing point and I didn't explain it well, sorry. I also might just be fundamentally confused and wrong. Hopefully this comment can explain it well enough so you can either shoot it down as incorrect or accept it.
First of all, it might be easier to understand if we replace "effective compute" with "general capabilities" in my original comment. Effective compute causally affects capabilities which causally affects both the METR time horizon measurement and the AI R&D speedup variable. So we can screen off the effective compute node and replace it with a general capabilities node.
I mostly agree with this. However, I think a model having very high time horizons is direct evidence for capabilities being high, which is then direct evidence for AI R&D speedup being high (and thus more likely to be past the AC threshold).
This leads to an important distinction. If you want to argue that AI R&D speedup will be higher (or certain thresholds of AI R&D speedup like AC being reached faster) using the METR graph, your argument fundamentally has to be an argument about capabilities being higher (arriving sooner) or an argument about the relationship between capabilities and AI R&D speedup.
I don't think that your first argument about the superexponential nature of the METR graph (that infinite time horizons in finite time implies it must be superexponential) is either of these. It seems to be an argument purely about the relationship between capabilities and the METR graph.
I also think your second argument for superexponential growth (that subsequent doublings should get easier because they require less new capabilities than earlier doublings) is mostly an argument about the relationship between capabilities and the METR graph. Although I could maybe see it being an argument about the relationship between capabilities and AI R&D speedup (the last few capabilities to be unlocked provide huge boosts to AI R&D speedup?).
Basically, my core concern is if neither of these arguments are truly arguments about capabilities arriving faster, or about the relationship between capabilities and AI R&D speedup, then why are they being used to update the estimation of time to ACs?
In contrast, arguments about AI R&D feeding back into itself, or about compute investment slowing down, are arguments directly about how fast capabilities will advance. So these validly affect your estimation of time to ACs.
If you disagree with me and think that the arguments for superexponential growth are not just arguments about the relationship between capabilities and the METR graph, then we can focus our discussion there. To the extent that these are also arguments about the things we truly care about, you should adjust time to ACs based on them; this is what I was trying to capture with the correlation stuff in my first comment.
If you do end up agreeing, I also don't really know how to incorporate this into the model. It seems hard and messy. I think that it is true that using the METR graph is the best thing we can do right now. I just don't think that these arguments about the superexponential nature of the METR graph should affect the estimation of time to ACs. But I do think that it is superexponential...and I think that it is the best way to estimate time to ACs...so again, messy :(. Hope this makes more sense!