Bayes-optimal learning of deep random networks of extensive-width

H Cui, F Krzakala, L Zdeborová - … Conference on Machine …, 2023 - proceedings.mlr.press
We consider the problem of learning a target function corresponding to a deep, extensive-
width, non-linear neural network with random Gaussian weights. We consider the asymptotic …

Universality laws for gaussian mixtures in generalized linear models

Y Dandi, L Stephan, F Krzakala… - Advances in …, 2024 - proceedings.neurips.cc
A recent line of work in high-dimensional statistics working under the Gaussian mixture
hypothesis has led to a number of results in the context of empirical risk minimization …

Spectral evolution and invariance in linear-width neural networks

Z Wang, A Engel, AD Sarwate… - Advances in Neural …, 2024 - proceedings.neurips.cc
We investigate the spectral properties of linear-width feed-forward neural networks, where
the sample size is asymptotically proportional to network width. Empirically, we show that the …

On double-descent in uncertainty quantification in overparametrized models

L Clarté, B Loureiro, F Krzakala… - International …, 2023 - proceedings.mlr.press
Uncertainty quantification is a central challenge in reliable and trustworthy machine
learning. Naive measures such as last-layer scores are well-known to yield overconfident …

Asymptotics of feature learning in two-layer networks after one gradient-step

H Cui, L Pesce, Y Dandi, F Krzakala, YM Lu… - arXiv preprint arXiv …, 2024 - arxiv.org
In this manuscript we investigate the problem of how two-layer neural networks learn
features from data, and improve over the kernel regime, after being trained with a single …

Nonlinear spiked covariance matrices and signal propagation in deep neural networks

Z Wang, D Wu, Z Fan - arXiv preprint arXiv:2402.10127, 2024 - arxiv.org
Many recent works have studied the eigenvalue spectrum of the Conjugate Kernel (CK)
defined by the nonlinear feature map of a feedforward neural network. However, existing …

Local Kernel Renormalization as a mechanism for feature learning in overparametrized Convolutional Neural Networks

R Aiudi, R Pacelli, A Vezzani, R Burioni… - arXiv preprint arXiv …, 2023 - arxiv.org
Feature learning, or the ability of deep neural networks to automatically learn relevant
features from raw data, underlies their exceptional capability to solve complex tasks …

Learning curves for deep structured Gaussian feature models

J Zavatone-Veth, C Pehlevan - Advances in Neural …, 2023 - proceedings.neurips.cc
In recent years, significant attention in deep learning theory has been devoted to analyzing
when models that interpolate their training data can still generalize well to unseen …

Universality of max-margin classifiers

A Montanari, F Ruan, B Saeed, Y Sohn - arXiv preprint arXiv:2310.00176, 2023 - arxiv.org
Maximum margin binary classification is one of the most fundamental algorithms in machine
learning, yet the role of featurization maps and the high-dimensional asymptotics of the …

Asymptotics of Learning with Deep Structured (Random) Features

D Schröder, D Dmitriev, H Cui, B Loureiro - arXiv preprint arXiv …, 2024 - arxiv.org
For a large class of feature maps we provide a tight asymptotic characterisation of the test
error associated with learning the readout layer, in the high-dimensional limit where the …