Statistical mechanics of deep learning

Y Bahri, J Kadmon, J Pennington… - Annual Review of …, 2020 - annualreviews.org
The recent striking success of deep neural networks in machine learning raises profound
questions about the theoretical principles underlying their success. For example, what can …

Scaling limits of wide neural networks with weight sharing: Gaussian process behavior, gradient independence, and neural tangent kernel derivation

G Yang - arXiv preprint arXiv:1902.04760, 2019 - arxiv.org
Several recent trends in machine learning theory and practice, from the design of state-of-
the-art Gaussian Process to the convergence analysis of deep neural nets (DNNs) under …

Does the brain behave like a (complex) network? I. Dynamics

D Papo, JM Buldú - Physics of Life Reviews, 2023 - Elsevier
Graph theory is now becoming a standard tool in system-level neuroscience. However,
endowing observed brain anatomy and dynamics with a complex network structure does not …

Self-consistent dynamical field theory of kernel evolution in wide neural networks

B Bordelon, C Pehlevan - Advances in Neural Information …, 2022 - proceedings.neurips.cc
We analyze feature learning in infinite-width neural networks trained with gradient flow
through a self-consistent dynamical field theory. We construct a collection of deterministic …

Dynamics of finite width kernel and prediction fluctuations in mean field neural networks

B Bordelon, C Pehlevan - Advances in Neural Information …, 2024 - proceedings.neurips.cc
We analyze the dynamics of finite width effects in wide but finite feature learning neural
networks. Starting from a dynamical mean field theory description of infinite width deep …

[图书][B] Statistical field theory for neural networks

M Helias, D Dahmen - 2020 - Springer
Many qualitative features of the emerging collective dynamics in neuronal networks, such as
correlated activity, stability, response to inputs, and chaotic and regular behavior, can be …

Incorporating heterogeneous interactions for ecological biodiversity

JI Park, DS Lee, SH Lee, HJ Park - Physical Review Letters, 2024 - APS
Understanding the behaviors of ecological systems is challenging given their multifaceted
complexity. To proceed, theoretical models such as Lotka-Volterra dynamics with random …

The effective noise of stochastic gradient descent

F Mignacco, P Urbani - Journal of Statistical Mechanics: Theory …, 2022 - iopscience.iop.org
Stochastic gradient descent (SGD) is the workhorse algorithm of deep learning technology.
At each step of the training phase, a mini batch of samples is drawn from the training dataset …

Second type of criticality in the brain uncovers rich multiple-neuron dynamics

D Dahmen, S Grün, M Diesmann… - Proceedings of the …, 2019 - National Acad Sciences
Cortical networks that have been found to operate close to a critical point exhibit joint
activations of large numbers of neurons. However, in motor cortex of the awake macaque …

Lyapunov spectra of chaotic recurrent neural networks

R Engelken, F Wolf, LF Abbott - Physical Review Research, 2023 - APS
This article is part of the Physical Review Research collection titled Physics of
Neuroscience. Recurrent networks are widely used as models of biological neural circuits …