Papers
arxiv:2405.19398

Neural Scaling Laws From Large-N Field Theory: Solvable Model Beyond the Ridgeless Limit

Published on May 29, 2024
Authors:

Abstract

Large-N field theory methods are used to solve a neural scaling law model, extending results to general ridge parameters and revealing a duality transformation underlying neural network scaling and quantum field theory simulation.

AI-generated summary

Many machine learning models based on neural networks exhibit scaling laws: their performance scales as power laws with respect to the sizes of the model and training data set. We use large-N field theory methods to solve a model recently proposed by Maloney, Roberts and Sully which provides a simplified setting to study neural scaling laws. Our solution extends the result in this latter paper to general nonzero values of the ridge parameter, which are essential to regularize the behavior of the model. In addition to obtaining new and more precise scaling laws, we also uncover a duality transformation at the diagrams level which explains the symmetry between model and training data set sizes. The same duality underlies recent efforts to design neural networks to simulate quantum field theories.

Community

Sign up or log in to comment

Get this paper in your agent:

hf papers read 2405.19398
Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2405.19398 in a model README.md to link it from this page.

Datasets citing this paper 1

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2405.19398 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.