Your short timelines don't seem that short, and your long timelines seem absurdly long.
Edit: removed '20' from list on cell 91:line 63
Mmm. I'm a bit confused about the short timelines: 50% by 2030 and 75% by 2030 seem pretty short to me.
I think the medium timelines I use has a pretty long tail, but the 75% by 2060 is pretty much exactly the Metaculus' community 75% by 2059.
I present an extension to my optimal timing of spending on AGI safety model for calculating the value of information of AGI timelines via informing one’s spending schedule.
I show, using my best guess of the model parameters, that for an AI risk funder uncertain between a ‘short timelines’ model and ‘medium timelines’ model:
The gains are greater when considering a model of the community’s capacity, rather than capital.
I also show that small changes in one’s credence in short or medium timelines has relatively little impact on one’s optimal spending schedule, especially when one starts out with roughly equal credence in each[1].
You can enter your own parameters - such as AGI timelines, discount rate and diminishing returns to spending - here.
In an appendix I apply a basic model to consider the opportunity cost of timelines work. This model does not assume novel research is done.
Suppose you have two ‘models’ of AGI timelines and , with credence in and in [2]. You use your mixture distribution for AGI timelines to calculate the optimal spending schedule for AI risk interventions.
You could do some thinking and come to some credence in and in . How much better is the optimal spending schedule as a result of to the optimal spending schedule as a result of , both supposing ?
Writing for optimal spending schedule according to and for utility of supposing , I compare to (the former, by definition of optimality is greater or equal to the latter).
In the model, utility is the discount adjusted probability we ‘succeed’ with making AGI go well. The maximum utility is where are the AGI timelines and is the discount rate.
I take and as the following, and other parameters as described here. (See the appendix for some basic statistics on the mixtures of and .
![]() | ![]() |
: ‘Short’ 2027 median timelines | : Metaculus community timelines (2039 median) |
I compute the results for both the ‘main’ model and the ‘alternate’ model described in the previous post.
I first show two examples of the optimal spending schedule for each of the models: one that is optimal supposing , short timelines, and one that is optimal supposing medium timelines, .
![]() | ![]() |
Main model The optimal spending schedule assuming , , (left) and the optimal spending schedule assuming , , (right), both assuming medium difficulty. and : going from certainty in to certainty in and following the new optimal spending schedule leads to 20% subjective increase in utility. and : going from certainty in to certainty in and following the new optimal spending schedule leads to 35% subjective increase in utility |
![]() | ![]() |
Alternate model The optimal spending schedule assuming , (left) and the optimal spending schedule assuming , (right), both assuming medium difficulty. and : going from certainty in to certainty in and following the new optimal spending schedule leads to 18% subjective increase in utility. and : going from certainty in to certainty in and following the new optimal spending schedule leads to 45% subjective increase in utility |
Disclaimer: these results are based on my guesses of the model parameters. Further, the models are of course not without limitations. I expect the results to be directionally equal, but lower, for the robust spend-save model. Overall, I’m less confident in these results than the previous spending results.
For each pairing (main capital spending model, alternate community capacity spending model) x AGI success is (easy, medium, hard) I show three graphs. The leftmost plot shows the % increase in discount-adjusted probability of success (utility). The central plot shows this % increase after first factoring out the utilons we get for free - our probability of success if we contributed nothing. The rightmost plot shows the absolute increase in utility.
Both the main and alternate model show that:
Naturally, those with low credal resilience in their AGI timelines have better returns to work on timelines.
The greatest potential gains of timelines work is for people already highly convinced of short timelines. This is particularly true if they are sacrificing gains in capacity now in order to ‘crunch’, or spending at a high rate now (sacrificing a greater amount of capital later). However, it seems that:
In the other direction, the value of information moving from high confidence in medium timelines to high confidence in short timelines are likely to be higher than the results suggest because the current spending rate and crunching rate are too low. That is, if you become convinced of short timelines, your marginal spending/crunching has lower diminishing returns (because, by your lights, the rest of the community is at a suboptimally low spending/crunching rate).
Acknowledgements: thanks to Daniel Kokotajlo for the idea, comments and suggestions. Thanks to Tom Barnes for comments. All remaining errors are my own.
Potentially useful for calculating your own .
25th percentile | Median | 75th percentile | |
1 () | 2025 | 2027 | 2030 |
0.75 | 2025 | 2028 | 2034 |
0.5 | 2026 | 2030 | 2041 |
0.25 | 2027 | 2033 | 2051 |
0 () | 2030 | 2039 | 2060 |
This model is highly flawed but potentially illustrative. Please take it with a massive grain of salt!
Suppose someone deliberates for years and updates from
to
. Had they not done the deliberation, they would have done
years - optimal according to
- leading to
units of output. After the update, each year they generate
units of output per year. Suppose this last until AGI arrives, which in expectation is
.[9] The person does not regret having done this if
.
Taking and
as above,
and
gives the following:[10]
The asymmetry in the plot due to the fact that in the case one in fact has longer than they’d guessed, their impact multiplier applies for more years (this could make sense if there are things you can do down that grow over time, even if in 2042 both the you-who-did-timelines-work-in-2022 and the you who didn’t have similar AGI timelines).
To use these results, one must have a prior over how their credence in A will change over the duration (in!). Someone with low credal resilience will have a 'wider' distribution than someone with higher credal resilience. One could simplify this step further by supposing discrete credences. For example, one could have 1/4 probability of staying at their current credence in A of 0.6, 1/4 credence in moving to 0.2 and 1/2 credence in moving to 0.8.
The toy model could be greatly improved. For example, one could model the situation as spending non-time resources, in which the above limitation does not occur. Further, one could allow for parallel work on timelines.
For example, moving from 40-60% credence in short to medium, or vice versa, has very little gains to the optimal spending schedule.
For example, A is ‘scale is all you need to AGI’ and B is ‘we need more difficult insights’. Or ‘’ is your independent impression, which you could become more confident in, and ‘’ is deference towards others (having accounted for information cascades etc)
25% probability of AGI going well if it arrived this year, and slope parameter
10% probability of AGI going well if it arrived this year, and slope parameter
4% probability of AGI going well if it arrived this year, and and slope parameter
25% probability of AGI going well if it arrived after a year and the community had been crunching at rate 1 unit of capacity per year (we start with one unit of capacity) and the slope parameter
10% probability of AGI going well if it arrived after a year and the community had been crunching at rate 1 unit of capacity per year (we start with one unit of capacity) and the slope parameter
5% probability of AGI going well if it arrived after a year and the community had been crunching at rate 1 unit of capacity per year (we start with one unit of capacity) and the slope parameter
Assuming is sufficiently small such that conditioning on no AGI in that time makes little difference.
This roughly approximates the results above