Crossposted from the AI Alignment Forum. May contain more technical jargon than usual.

TLDR: Humanity — which includes all nations, organisations, and individuals — should limit the growth rate of machine learning training runs from 2020 until 2050 to below 0.2 OOMs/year.

Paris Climate Accords

In the early 21st century, the climate movement converged around a "2°C target", shown in Article 2(1)(a) of the Paris Climate Accords:

Holding the increase in the global average temperature to well below 2°C above pre-industrial levels and pursuing efforts to limit the temperature increase to 1.5°C above pre-industrial levels, recognizing that this would significantly reduce the risks and impacts of climate change;
"Holding the increase in the global average temperature to well below 2°C above pre-industrial levels and pursuing efforts to limit the temperature increase to 1.5°C above pre-industrial levels, recognizing that this would significantly reduce the risks and impacts of climate change;"(source)

The 2°C target helps facilitate coordination between nations, organisations, and individuals.

  • It provided a clear, measurable goal.
  • It provided a sense of urgency and severity.
  • It promoted a sense of shared responsibility.
  • It establishes common knowledge of stakeholder goals.
  • It helped to align efforts across different stakeholders.
  • It signals a technical practical mindset for solving the problem.
  • It created a shared understanding of what success would look like.

The 2°C target was the first step towards coordination, not the last step.

The AI governance community should converge around a similar target.

0.2 OOMs/year target

I propose a fixed target of 0.2 OOMs/year. "OOM" stands for "orders of magnitude" and corresponds to a ten-fold increase, so 0.2 OOMs/year corresponds to a 58% year-on-year growth. The 0.2 OOMs/year figure was recently suggested by Jaime Sevilla, which prompted me to write this article.

  • I do not propose any specific policy for achieving the 0.2 OOMs/year target, because the purpose of the target is to unify stakeholders even if they support different policies.
  • I do not propose any specific justification for the 0.2 OOMs/year target, because the purpose of the target is to unify stakeholders even if they have different justifications.

Here is the statement:

"Humanity — which includes all nations, organisations, and individuals — should limit the growth rate of machine learning training runs from 2020 until 2050 to below 0.2 OOMs/year."

The statement is intentionally ambiguous about how to measure "the growth rate of machine learning training runs". I suspect that a good proxy metric would be the effective training footprint (defined below) but I don't think the proxy metric should be included in the statement of the target itself.

Effective training footprint

What is the effective training footprint?

The effective training footprint, measured in FLOPs, is one proxy metric for the growth rate of machine learning training runs. The footprint of a model is defined, with caveats, as the total number of FLOPs used to train the model since initialisation.

Caveats:

  • A randomly initialised model has a footprint of 0 FLOPs.
  • If the model is trained from a randomly initialised model using SGD or a variant, then its footprint is the total number of FLOPs used in the training process.
  • If a pre-trained base model is used for the initialisation of another training process (such as unsupervised learning, supervised learning, fine-tuning, or reinforcement learning), then the footprint of the resulting model will include the footprint of the pre-trained model.
  • If multiple models are composed to form a single cohesive model, then the footprint of the resulting model is the sum of the footprints of each component model.
  • If there is a major algorithmic innovation which divides by a factor of  the FLOPs required to train a model to a particular score on downstream tasks, then the footprint of models trained with that innovation is multiplied by the same factor .
  • This list of caveats to the definition of Effective Training Footprint is non-exhaustive. Future consultations may yield additional caveats, or replace Effective Training Footprint with an entirely different proxy metric.

Fixing the y-axis

  • According to the 0.2 OOMs/year target, there cannot exist an ML model during the year  with a footprint exceeding , where . That means that  FLOPs for some fixed constant .
  • If we consult EpochAI's plot of compute training runs during the large-scale era of ML, we see that footprints have been growing with approximately 0.5 OOMs/year.
  • We can use this trend to fix the value of . In 2022, the footprint was approximately 1.0e+24. Therefore .
  • In other words, .
  • I have used 2022 as an anchor to fix the y-axis . If I had used an earlier date then the 0.2 OOMs/yr target would've been stricter, and if I had used a later date then the 0.2 OOMs/yr target would've been laxer. If the y-axis for the constraint is fixed to the day of the negotiation (the default schelling date), then stakeholders who want a laxer constraint are incentivised to delay negotiation. To avoid that hazard, I have picked "January 1st 2022" to fix the y-axis. I declare 1/1/2022 to be the schelling date for the 0.2 OOMs/year target.

Year-by-year limits

In year , all model must have a log10-footprint below .

YearMaximum training footprint (FLOPs) in logarithm base 10Maximum training footprint (FLOPs)
202023.63.98E+23
202123.86.31E+23
202224.01.00E+24
202324.21.58E+24
202424.42.51E+24
202524.63.98E+24
202624.86.31E+24
202725.01.00E+25
202825.21.58E+25
202925.42.51E+25
203025.63.98E+25
203125.86.31E+25
203226.01.00E+26
203326.21.58E+26
203426.42.51E+26
203526.63.98E+26
203626.86.31E+26
203727.01.00E+27
203827.21.58E+27
203927.42.51E+27
204027.63.98E+27
204127.86.31E+27
204228.01.00E+28
204328.21.58E+28
204428.42.51E+28
204528.63.98E+28
204628.86.31E+28
204729.01.00E+29
204829.21.58E+29
204929.42.51E+29
205029.63.98E+29

Implications of the 0.2 OOMs/year target

  • Because , this means that the maximum footprint would grow 58% every year.
  • 0.2 OOMs/year is equivalent to a doubling time of 18 months.
  • Every decade, the maximum permissible footprint increases by a factor of 100.
  • 0.2 OOMs/year was the pre-AlexNet growth rate in ML systems.
  • The current growth rate is 0.5 OOMs/year, which is 2.5 times faster than the target rate.
  • As the current 0.5 OOMs/year growth rate, after 10 years we would have ML training runs which are 100 000x larger than existing training runs. Under the 0.2 OOMs/year growth rate, this growth would be spread over 25 years instead.
  • Comparing 0.2 OOMs/year target to hardware growth-rates:
    • Moore's Law states that transitiors per integrated circuit doubles roughly every 2 years.
    • Koomey's Law states that the FLOPs-per-Joule doubled roughly every 1.57 years until 2000, whereupon it began doubling roughly every 2.6 years.
    • Huang's Law states that the growth-rate of GPU performance exceeds that of CPU performance. This is a somewhat dubious claim, but nonetheless I think the doubling time of GPUs is longer than 18 months.
    • In general, the 0.2 OOMs/year target is faster than the current hardware growth-rate.
  • On March 15 2023, OpenAI released GPT-4 which was trained with an estimated 2.8e+25 FLOPs. If OpenAI had followed the 0.2 OOMs/year target, then GPT-4 would've been released on March 29 2029. This is because if  then.
  • 0.2 OOMs/year target would therefore be an effective moratorium on models exceeding GPT-4 until 2029. Nonetheless, the moratorium would still allow an AI Summer Harvest — in which the impact of ChatGPT-3.5/4 steadily dissipates across the economy until a new general equilbirum is reached where...
  1. People have more money to spend.
  2. The products and services are more abundant, cheaper, and of a higher quality.
  3. People have more leisure to enjoy themselves.
New Comment
24 comments, sorted by Click to highlight new comments since: Today at 8:01 AM
[-]ESRogs1y21-8

0.2 OOMs/year was the pre-AlexNet growth rate in ML systems.

I think you'd want to set the limit to something slightly faster than Moore's law. Otherwise you have a constant large compute overhang.

Ultimately, we're going to be limited by Moore's law (or its successor) growth rates eventually anyway. We're on a kind of z-curve right now, where we're transitioning from ML compute being some small constant fraction of all compute to some much larger constant fraction of all compute. Before the transition it grows at the same speed as compute in general. After the transition it also grows at the same speed as compute in general. In the middle it grows faster as we rush to spend a much larger share of GWP on it.

From that perspective, Moore's law growth is the minimum growth rate you might have (unless annual spend on ML shrinks). And the question is just whether you transition from the small constant fraction of all compute to the large constant fraction of all compute slowly or quickly.

Trying to not do the transition at all (i.e. trying to growing at exactly the same rate as compute in general) seems potentially risky, because the resulting constant compute overhang means it's relatively easy for someone somewhere to rush ahead locally and build something much better than SOTA.

If on the other hand, you say full steam ahead and don't try to slow the transition at all, then on the plus side the compute overhang goes away, but on the minus side, you might rush into dangerous and destabilizing capabilities.

Perhaps a middle path makes sense, where you slow the growth rate down from current levels, but also slowly close the compute overhang gap over time.

Moore's law is a doubling every 2 years, while this proposes doubling every 18 months, so pretty much what you suggest (not sure if you were disagreeing tbh but seemed like you might be?)

Ah, good point!

Otherwise you have a constant large compute overhang.


I think we should strongly consider finding a way of dealing with that rather than only looking at solutions that produce no overhang. For all we know, total compute required for TAI (especially factoring in future algorithmic progress) isn't far away from where we are now. Dealing with the problem of preventing defectors from exploiting a compute overhang seems potentially easier than solving alignment on a very short timescale. 

I suppose a possible mistake in this analysis is that I'm treating Moore's law as the limit on compute growth rates, and this may not hold once we have stronger AIs helping to design and fabricate chips.

Even so, I think there's something to be said for trying to slowly close the compute overhang gap over time.

Upvoted for creative thinking. I'm having trouble deciding whether it's a good idea.

0.2 OOMs/year is equivalent to a doubling time of 8 months.

I think this is wrong, that's nearly 8 doublings in 5 years, should instead be doubling every 5 years, should instead be doubling every 5 / log2(10) = 1.5.. years

I think pushing GPT-4 out to 2029 would be a good level of slowdown from 2022, but assuming that we could achieve that level of impact, what's the case for having a fixed exponential increase? Is it to let of some level of 'steam' in the AI industry? So that we can still get AGI in our lifetimes? To make it seem more reasonable to policymakers?

I would still rather have a moratorium until some measure of progress of understanding personally. We don't have a fixed temperature increase per decade built into our climate targets.

The 0.2 OOMs/year target would be an effective moratorium until 2029, because GPT-4 overshot the target.

Yep, thanks! 0.2 OOMs/year is equivalent to a doubling time of 18 months. I think that was just a typo.

I wonder if a basket of SOTA benchmarks would make more sense. Allow no more than X% increase in performance across the average of the benchmarks per year. This would capture the FLOPS metric along with potential speedups, fine-tuning, or other strategies.

Conveniently, this is how the teams are already ranking their models against each other so there's ample evidence of past progress and researchers are incentivized to report accurately; there's no incentive to "cheat" if researchers are not allowed to publish greater increases on SOTA benchmarks than the limit allows (e.g. journals would say "shut it down" instead of publish the paper), unless an actor wanted to simply jump ahead of everyone else and go for a singleton on their own, which is already an unavoidable risk without EY-style coordinated hard stop.

Great idea! Let's measure algorithmic improvement in the same way economists measure inflation, with a basket-of-benchmarkets.

This basket can itself be adjusted over time so it continuously reflected the current use-cases of SOTA AI.

I haven't thought about it much, but my guess is the best thing to do is to limit training compute directly but adjust the limit using the basket-of-benchmarks.

One weakness I realized overnight is that this incentivizes branching out into new problem domains. One potential fix is to, when novel domains show up, shoehorn the big LLMs into solving that domain on the same benchmark and limit new types of models/training to what the LLMs can accomplish in that new domain. Basically setting an initially low SOTA that can grow at the same percentage as the rest of the basket. This might prevent leapfrogging the general models with narrow ones that are mostly mesa-optimizer or similar.

I think this is a really promising idea.

If the goal is to unify diverse stakeholders, including non-technical ones, I wonder if it would be better to use a less-wonky target (e.g. "50%" instead of ".002 OOMs")

I think you have to set this up in such a way that the current ceiling is where we already are, not back in time to before GPT-4. If you don't, then the chance it actually gets adopted seems vastly lower, since all adopters that didn't make their own GPT-4 already have to agree to be 2nd-class entities until 2029.

It's very difficult to talk about nuclear non-proliferation when a bunch of people already have nukes. If you can actually enforce it, that's a different story, but if you could actually enforce anything relating to this mess the rest just becomes details anyway.

  1. Nuclear proliferation worked despite the fact that many countries with nuclear weapons were "grandfathered in".

  2. If the y-axis for the constraint is fixed to the day of the negotiaiton, then stakeholders who want a laxer constraint are incentivised to delay negotiation. To avoid that hazard, I have picked a schelling date (2022) to fix the y-axis.
  3. The purpose of this article isn't to proposal any policy, strategy, treaty, agreement, law, etc which might achieve the 0.2 OOMs/year target. instead, the purpose of this article is to propose a target itself. This has inherent coordination benefits, c.f. the 2ºC target. 

Nuclear non-proliferation worked because the grandfathered-in countries had all the power and the ones who weren't were under the implicit threat of embargo, invasion, or even annihilation. Despite all its accomplishments, GPT-4 does not give Open AI the ability to enforce its monopoly with the threat of violence.

Not to mention that 3-4 of the 5 listed countries non-party to the treaty developed nukes anyway. If Meta decides to flagrantly ignore the 0.2 OOM limit and creates something actually dangerous it's not going to sit quietly in a silo waiting for further mistakes to be made before it kills us all.

I think you've misunderstood what we mean by "target". Similar issues applied to the 2°C target, which nonetheless yielded significant coordination benefits.

The 2°C target helps facilitate coordination between nations, organisations, and individuals.

  • It provided a clear, measurable goal.
  • It provided a sense of urgency and severity.
  • It promoted a sense of shared responsibility.
  • It helped to align efforts across different stakeholders.
  • It created a shared understanding of what success would look like.

The AI governance community should converge around a similar target.

Unfortunately we may already have enough compute, and it will be difficult to enforce a ban on decentralized training (which isn't competitive yet, but likely could be with more research).

This isn't a policy proposal, it's a target, like the 2°C climate target.

I think concrete ideas like this that take inspiration from past regulatory successes are quite good, esp. now that policymakers are discussing the issue.

I appreciate the concept. 

I wonder about how the hardware overhangs would look like after the moratorium ends or somebody bypasses it.

On first thought it doesnt look like a robuts solution, I assume a 15-20% improvement on Compute access per year. Would need to plot it against the moratorium treshold and see if over time one gets closer to the Yudkowski Airstrike treshold  but I assume no, 20% vs 54%,,

I dont know! maybe this is a good idea. 

I've added a section on hardware:

  • Comparing 0.2 OOMs/year target to hardware growth-rates:
    • Moore's Law states that transitiors per integrated circuit doubles roughly every 2 years.
    • Koomey's Law states that the FLOPs-per-Joule doubled roughly every 1.57 years until 2000, whereupon it began doubling roughly every 2.6 years.
    • Huang's Law states that the growth-rate of GPU performance exceeds that of CPU performance. This is a somewhat dubious claim, but nonetheless I think the doubling time of GPUs is longer than 18 months.
    • In general, the 0.2 OOMs/year target is faster than the current hardware growth-rate.

Suggestion: mention that OOM means "orders if magnitude". I was confused until I noticed the powers of 10.

Confused how?.. The only thing that comes to mind is that's FOOM sans F. Asking for 0.2 FOOMS limit seens reasonable given current trajectory 😅