A farewell to the bias-variance tradeoff? an overview of the theory of overparameterized machine learning

Y Dar, V Muthukumar, RG Baraniuk - arXiv preprint arXiv:2109.02355, 2021 - arxiv.org
The rapid recent progress in machine learning (ML) has raised a number of scientific
questions that challenge the longstanding dogma of the field. One of the most important …

Universality of empirical risk minimization

A Montanari, BN Saeed - Conference on Learning Theory, 2022 - proceedings.mlr.press
Consider supervised learning from iid samples {(y_i, x_i)} _ {i≤ n} where x_i∈ R_p are
feature vectors and y_i∈ R are labels. We study empirical risk minimization over a class of …

Learning curves of generic features maps for realistic datasets with a teacher-student model

B Loureiro, C Gerbelot, H Cui, S Goldt… - Advances in …, 2021 - proceedings.neurips.cc
Teacher-student models provide a framework in which the typical-case performance of high-
dimensional supervised learning can be described in closed form. The assumptions of …

Bayes-optimal learning of deep random networks of extensive-width

H Cui, F Krzakala, L Zdeborová - … Conference on Machine …, 2023 - proceedings.mlr.press
We consider the problem of learning a target function corresponding to a deep, extensive-
width, non-linear neural network with random Gaussian weights. We consider the asymptotic …

Label-imbalanced and group-sensitive classification under overparameterization

GR Kini, O Paraskevas, S Oymak… - Advances in Neural …, 2021 - proceedings.neurips.cc
The goal in label-imbalanced and group-sensitive classification is to optimize relevant
metrics such as balanced error and equal opportunity. Classical methods, such as weighted …

Triple descent and the two kinds of overfitting: Where & why do they appear?

S d'Ascoli, L Sagun, G Biroli - Advances in neural …, 2020 - proceedings.neurips.cc
A recent line of research has highlighted the existence of a``double descent''phenomenon in
deep learning, whereby increasing the number of training examples N causes the …

Benign overfitting in multiclass classification: All roads lead to interpolation

K Wang, V Muthukumar… - Advances in Neural …, 2021 - proceedings.neurips.cc
The growing literature on" benign overfitting" in overparameterized models has been mostly
restricted to regression or binary classification settings; however, most success stories of …

Graph-based approximate message passing iterations

C Gerbelot, R Berthier - Information and Inference: A Journal of …, 2023 - academic.oup.com
Approximate message passing (AMP) algorithms have become an important element of high-
dimensional statistical inference, mostly due to their adaptability and concentration …

High-dimensional asymptotics of denoising autoencoders

H Cui, L Zdeborová - Advances in Neural Information …, 2024 - proceedings.neurips.cc
We address the problem of denoising data from a Gaussian mixture using a two-layer non-
linear autoencoder with tied weights and a skip connection. We consider the high …

Rigorous dynamical mean-field theory for stochastic gradient descent methods

C Gerbelot, E Troiani, F Mignacco, F Krzakala… - SIAM Journal on …, 2024 - SIAM
We prove closed-form equations for the exact high-dimensional asymptotics of a family of
first-order gradient-based methods, learning an estimator (eg, M-estimator, shallow neural …