Representations and generalization in artificial and brain neural networks
Humans and animals excel at generalizing from limited data, a capability yet to be fully
replicated in artificial intelligence. This perspective investigates generalization in biological …
replicated in artificial intelligence. This perspective investigates generalization in biological …
A statistical mechanics framework for Bayesian deep neural networks beyond the infinite-width limit
Despite the practical success of deep neural networks, a comprehensive theoretical
framework that can predict practically relevant scores, such as the test accuracy, from …
framework that can predict practically relevant scores, such as the test accuracy, from …
Self-consistent dynamical field theory of kernel evolution in wide neural networks
B Bordelon, C Pehlevan - Advances in Neural Information …, 2022 - proceedings.neurips.cc
We analyze feature learning in infinite-width neural networks trained with gradient flow
through a self-consistent dynamical field theory. We construct a collection of deterministic …
through a self-consistent dynamical field theory. We construct a collection of deterministic …
Dynamics of finite width kernel and prediction fluctuations in mean field neural networks
B Bordelon, C Pehlevan - Advances in Neural Information …, 2024 - proceedings.neurips.cc
We analyze the dynamics of finite width effects in wide but finite feature learning neural
networks. Starting from a dynamical mean field theory description of infinite width deep …
networks. Starting from a dynamical mean field theory description of infinite width deep …
How two-layer neural networks learn, one (giant) step at a time
We investigate theoretically how the features of a two-layer neural network adapt to the
structure of the target function through a few large batch gradient descent steps, leading to …
structure of the target function through a few large batch gradient descent steps, leading to …
Bayesian interpolation with deep linear networks
Characterizing how neural network depth, width, and dataset size jointly impact model
quality is a central problem in deep learning theory. We give here a complete solution in the …
quality is a central problem in deep learning theory. We give here a complete solution in the …
Feature-learning networks are consistent across widths at realistic scales
We study the effect of width on the dynamics of feature-learning neural networks across a
variety of architectures and datasets. Early in training, wide neural networks trained on …
variety of architectures and datasets. Early in training, wide neural networks trained on …
A theory of non-linear feature learning with one gradient step in two-layer neural networks
Feature learning is thought to be one of the fundamental reasons for the success of deep
neural networks. It is rigorously known that in two-layer fully-connected neural networks …
neural networks. It is rigorously known that in two-layer fully-connected neural networks …
Artificial Intelligence for Complex Network: Potential, Methodology and Application
Complex networks pervade various real-world systems, from the natural environment to
human societies. The essence of these networks is in their ability to transition and evolve …
human societies. The essence of these networks is in their ability to transition and evolve …
Neural network field theories: non-Gaussianity, actions, and locality
Both the path integral measure in field theory (FT) and ensembles of neural networks (NN)
describe distributions over functions. When the central limit theorem can be applied in the …
describe distributions over functions. When the central limit theorem can be applied in the …