A farewell to the bias-variance tradeoff? an overview of the theory of overparameterized machine learning
Y Dar, V Muthukumar, RG Baraniuk - arXiv preprint arXiv:2109.02355, 2021 - arxiv.org
The rapid recent progress in machine learning (ML) has raised a number of scientific
questions that challenge the longstanding dogma of the field. One of the most important …
questions that challenge the longstanding dogma of the field. One of the most important …
Universality of empirical risk minimization
A Montanari, BN Saeed - Conference on Learning Theory, 2022 - proceedings.mlr.press
Consider supervised learning from iid samples {(y_i, x_i)} _ {i≤ n} where x_i∈ R_p are
feature vectors and y_i∈ R are labels. We study empirical risk minimization over a class of …
feature vectors and y_i∈ R are labels. We study empirical risk minimization over a class of …
Learning curves of generic features maps for realistic datasets with a teacher-student model
Teacher-student models provide a framework in which the typical-case performance of high-
dimensional supervised learning can be described in closed form. The assumptions of …
dimensional supervised learning can be described in closed form. The assumptions of …
Bayes-optimal learning of deep random networks of extensive-width
We consider the problem of learning a target function corresponding to a deep, extensive-
width, non-linear neural network with random Gaussian weights. We consider the asymptotic …
width, non-linear neural network with random Gaussian weights. We consider the asymptotic …
Label-imbalanced and group-sensitive classification under overparameterization
The goal in label-imbalanced and group-sensitive classification is to optimize relevant
metrics such as balanced error and equal opportunity. Classical methods, such as weighted …
metrics such as balanced error and equal opportunity. Classical methods, such as weighted …
Triple descent and the two kinds of overfitting: Where & why do they appear?
A recent line of research has highlighted the existence of a``double descent''phenomenon in
deep learning, whereby increasing the number of training examples N causes the …
deep learning, whereby increasing the number of training examples N causes the …
Benign overfitting in multiclass classification: All roads lead to interpolation
K Wang, V Muthukumar… - Advances in Neural …, 2021 - proceedings.neurips.cc
The growing literature on" benign overfitting" in overparameterized models has been mostly
restricted to regression or binary classification settings; however, most success stories of …
restricted to regression or binary classification settings; however, most success stories of …
Graph-based approximate message passing iterations
C Gerbelot, R Berthier - Information and Inference: A Journal of …, 2023 - academic.oup.com
Approximate message passing (AMP) algorithms have become an important element of high-
dimensional statistical inference, mostly due to their adaptability and concentration …
dimensional statistical inference, mostly due to their adaptability and concentration …
High-dimensional asymptotics of denoising autoencoders
H Cui, L Zdeborová - Advances in Neural Information …, 2024 - proceedings.neurips.cc
We address the problem of denoising data from a Gaussian mixture using a two-layer non-
linear autoencoder with tied weights and a skip connection. We consider the high …
linear autoencoder with tied weights and a skip connection. We consider the high …
Rigorous dynamical mean-field theory for stochastic gradient descent methods
We prove closed-form equations for the exact high-dimensional asymptotics of a family of
first-order gradient-based methods, learning an estimator (eg, M-estimator, shallow neural …
first-order gradient-based methods, learning an estimator (eg, M-estimator, shallow neural …