Wiki Contributions


Read Section 6 titled “The Limit of the Predictability of Scaling Behavior” in this paper:

We describe how to go about fitting a BNSL to yield best extrapolation in the last paragraph of Appendix Section A.6 "Experimental details of fitting BNSL and determining the number of breaks" of the paper:

Ethan Caballero11moΩ-22-4

Sigmoids don't accurately extrapolate the scaling behavior(s) of the performance of artificial neural networks. 

Use a Broken Neural Scaling Law (BNSL) in order to obtain accurate extrapolations:

Did ARC try making a scaling plot with training compute on the x-axis and autonomous replication on the y-axis?

The setting was adversarial training and adversarial evaluation. During training, PGD attacker of 30 iterations is used to construct adversarial examples used for training. During testing, the evaluation test set is an adversarial test set that is constructed via PGD attacker of 20 iterations.

Experimental data of y-axis is obtained from Table 7 of; experimental data of x-axis is obtained from Figure 7 of

"However, to the best of our knowledge there are no quantitative scaling laws for robustness yet."

For scaling laws for adversarial robustness, see appendix A.15 of

See section 5.3 "Reinforcement Learning" of for more RL scaling laws with number of model parameters on the x-axis (and also RL scaling laws with the amount of compute used for training on the x-axis and RL scaling laws with training dataset size on the x-axis).

re: youtube estimates

You'll probably find some of this twitter discussion useful:

Load More