Representations and generalization in artificial and brain neural networks

Q Li, B Sorscher, H Sompolinsky - Proceedings of the National Academy of …, 2024 - pnas.org
Humans and animals excel at generalizing from limited data, a capability yet to be fully
replicated in artificial intelligence. This perspective investigates generalization in biological …

A statistical mechanics framework for Bayesian deep neural networks beyond the infinite-width limit

R Pacelli, S Ariosto, M Pastore, F Ginelli… - Nature Machine …, 2023 - nature.com
Despite the practical success of deep neural networks, a comprehensive theoretical
framework that can predict practically relevant scores, such as the test accuracy, from …

Self-consistent dynamical field theory of kernel evolution in wide neural networks

B Bordelon, C Pehlevan - Advances in Neural Information …, 2022 - proceedings.neurips.cc
We analyze feature learning in infinite-width neural networks trained with gradient flow
through a self-consistent dynamical field theory. We construct a collection of deterministic …

Dynamics of finite width kernel and prediction fluctuations in mean field neural networks

B Bordelon, C Pehlevan - Advances in Neural Information …, 2024 - proceedings.neurips.cc
We analyze the dynamics of finite width effects in wide but finite feature learning neural
networks. Starting from a dynamical mean field theory description of infinite width deep …

How two-layer neural networks learn, one (giant) step at a time

Y Dandi, F Krzakala, B Loureiro, L Pesce… - arXiv preprint arXiv …, 2023 - arxiv.org
We investigate theoretically how the features of a two-layer neural network adapt to the
structure of the target function through a few large batch gradient descent steps, leading to …

Bayesian interpolation with deep linear networks

B Hanin, A Zlokapa - … of the National Academy of Sciences, 2023 - National Acad Sciences
Characterizing how neural network depth, width, and dataset size jointly impact model
quality is a central problem in deep learning theory. We give here a complete solution in the …

Feature-learning networks are consistent across widths at realistic scales

N Vyas, A Atanasov, B Bordelon… - Advances in …, 2024 - proceedings.neurips.cc
We study the effect of width on the dynamics of feature-learning neural networks across a
variety of architectures and datasets. Early in training, wide neural networks trained on …

A theory of non-linear feature learning with one gradient step in two-layer neural networks

B Moniri, D Lee, H Hassani, E Dobriban - arXiv preprint arXiv:2310.07891, 2023 - arxiv.org
Feature learning is thought to be one of the fundamental reasons for the success of deep
neural networks. It is rigorously known that in two-layer fully-connected neural networks …

Artificial Intelligence for Complex Network: Potential, Methodology and Application

J Ding, C Liu, Y Zheng, Y Zhang, Z Yu, R Li… - arXiv preprint arXiv …, 2024 - arxiv.org
Complex networks pervade various real-world systems, from the natural environment to
human societies. The essence of these networks is in their ability to transition and evolve …

Neural network field theories: non-Gaussianity, actions, and locality

M Demirtas, J Halverson, A Maiti… - Machine Learning …, 2024 - iopscience.iop.org
Both the path integral measure in field theory (FT) and ensembles of neural networks (NN)
describe distributions over functions. When the central limit theorem can be applied in the …