[AN #125]: Neural network scaling laws across multiple modalities

https://www.lesswrong.com/posts/XPqMbtpbku8aN55wd/an-125-neural-network-scaling-laws-across-multiple

Link post Contents

SECTIONS

HIGHLIGHTS TECHNICAL AI ALIGNMENT MESA OPTIMIZATION FORECASTING OTHER PROGRESS IN AI REINFORCEMENT LEARNING

HIGHLIGHTS

Scaling Laws for Autoregressive Generative Modeling (Tom Henighan, Jared Kaplan, Mor Katz et al) (summarized by Asya): This paper looks at scaling laws for generative Transformer models of images (predicting pixels or parts of image encodings), videos (predicting frames of image encodings), multimodal image <-> text (predicting captions based on images or images based on captions), and mathematical problem solving (predicting answers to auto-generated questions about algebra, arithmetic, calculus, comparisons, integer properties, measurement, polynomials, and probability). The authors find that:

TECHNICAL AI ALIGNMENT

MESA OPTIMIZATION

Confucianism in AI Alignment (John Wentworth) (summarized by Rohin): Suppose we trained our agent to behave well on some set of training tasks. Mesa optimization (AN #58) suggests that we may still have a problem: the agent might perform poorly during deployment, because it ends up optimizing for some misaligned mesa objective that only agrees with the base objective on the training distribution. This post suggests that in any training setup in which mesa optimizers would normally be incentivized, it is not sufficient to just prevent mesa optimization from happening. The fact that mesa optimizers could have arisen means that the incentives were bad. If you somehow removed mesa optimizers from the search space, there would still be a selection pressure for agents that without any malicious intent end up using heuristics that exploit the bad incentives. As a result, we should focus on fixing the incentives, rather than on excluding mesa optimizers from the search space. Clarifying inner alignment terminology (Evan Hubinger) (summarized by Rohin): This post clarifies the author’s definitions of various terms around inner alignment. Alignment is split into intent alignment and capability robustness, and then intent alignment is further subdivided into outer alignment and objective robustness. Inner alignment is one way of achieving objective robustness, in the specific case that you have a mesa optimizer. See the post for more details on the definitions. Rohin’s opinion: I’m glad that definitions are being made clear, especially since I usually use these terms differently than the author. In particular, as mentioned in my opinion on the highlighted paper, I expect performance to smoothly go up with additional compute, data, and model capacity, and there won’t be a clear divide between capability robustness and objective robustness. As a result, I prefer not to divide these as much as is done in this post.

FORECASTING

Measuring Progress in Deep Reinforcement Learning Sample Efficiency (Anonymous) (summarized by Asya) (H/​T Carl Shulman): This paper measures historic increases in sample efficiency by looking at the number of samples needed to reach some fixed performance level on Atari games and virtual continuous control tasks. The authors find exponential progress in sample efficiency, with estimated doubling times of 10 to 18 months on Atari, 5 to 24 months on state-based continuous control, and 4 to 9 months on pixel-based continuous control, depending on the specific task and performance level. They find that these gains were mainly driven by improvements in off-policy and model-based deep RL learning approaches, as well as the use of auxiliary learning objectives to speed up representation learning, and not by model size improvements. The authors stress that their study is limited in studying only the published training curves for only three tasks, not accounting for the extent to which hyperparameter tuning may have been responsible for historic gains. Asya’s opinion: Following in the footsteps of AI and Efficiency (AN #99), here we have a paper showing exponential gains in sample efficiency in particular. I’m really glad someone did this analysis—I think I’m surprised by how fast progress is, though as the paper notes it’s unclear exactly how to relate historic improvements on fixed task performance to a sense of overall improvement in continuous control (though several of the main contributors listed in the appendix seem fairly general). I also really appreciate how thorough the full paper is in listing limitations to this work. Since these papers are coming up in the same newsletter, I’ll note the contrast between the data-unlimited domains explored in the scaling laws paper and the severely data-limited domain of real-world robotics emphasized in this paper. In robotics, it seems we are definitely still constrained by algorithmic progress that lets us train on fewer samples (or do better transfer from simulations (AN #72)). Of course, maybe progress in data-unlimited domains will ultimately result in AIs that make algorithmic progress in data-limited domains faster than humans ever could.

OTHER PROGRESS IN AI

REINFORCEMENT LEARNING

DeepSpeed: Extreme-scale model training for everyone (DeepSpeed Team et al) (summarized by Asya): In this post, Microsoft announces updates to DeepSpeed, its open-source deep learning training optimization library. The new updates include:

Comment

https://www.lesswrong.com/posts/XPqMbtpbku8aN55wd/an-125-neural-network-scaling-laws-across-multiple?commentId=Pk23Yed2y7F6GudkD

As always, thanks to everyone involved for the newsletter! I’m usually particularly interested in the other RL/​Deep Learning papers, as those are the ones I have less chance to find on my own. On this newsletter, I especially enjoyed the summaries and opinions about the two scaling papers, and the comparison between the two.

https://www.lesswrong.com/posts/XPqMbtpbku8aN55wd/an-125-neural-network-scaling-laws-across-multiple?commentId=HkxScT4z8mjSHrJHd

Cool papers, Flo!

https://www.lesswrong.com/posts/XPqMbtpbku8aN55wd/an-125-neural-network-scaling-laws-across-multiple?commentId=KTmmwJGfnQnzNrP2Y

I don’t know what happened, but for me the visuals are a mess. Each part of the newsletter is separated by many screens of blank space, the comment section appears just after the "podcast" section, which makes sense, except that I see everything else below the "podcast" section, even if it should be the last section.

Comment

https://www.lesswrong.com/posts/XPqMbtpbku8aN55wd/an-125-neural-network-scaling-laws-across-multiple?commentId=j2fM3x4aTjXzodtTB

Known problem, should be fixed in the next few hours.

Comment

Okay, thanks!

Comment

Fixed.

Comment

Thanks!