A farewell to the bias-variance tradeoff? an overview of the theory of overparameterized machine learning

Y Dar, V Muthukumar, RG Baraniuk - arXiv preprint arXiv:2109.02355, 2021 - arxiv.org
The rapid recent progress in machine learning (ML) has raised a number of scientific
questions that challenge the longstanding dogma of the field. One of the most important …

[HTML][HTML] Deep learning for small and big data in psychiatry

G Koppe, A Meyer-Lindenberg… - …, 2021 - nature.com
Psychiatry today must gain a better understanding of the common and distinct
pathophysiological mechanisms underlying psychiatric disorders in order to deliver more …

Deep learning: a statistical viewpoint

PL Bartlett, A Montanari, A Rakhlin - Acta numerica, 2021 - cambridge.org
The remarkable practical success of deep learning has revealed some major surprises from
a theoretical perspective. In particular, simple gradient methods easily find near-optimal …

What neural networks memorize and why: Discovering the long tail via influence estimation

V Feldman, C Zhang - Advances in Neural Information …, 2020 - proceedings.neurips.cc
Deep learning algorithms are well-known to have a propensity for fitting the training data
very well and often fit even outliers and mislabeled data points. Such fitting requires …

Fit without fear: remarkable mathematical phenomena of deep learning through the prism of interpolation

M Belkin - Acta Numerica, 2021 - cambridge.org
In the past decade the mathematical theory of machine learning has lagged far behind the
triumphs of deep neural networks on practical challenges. However, the gap between theory …

Benign overfitting in linear regression

PL Bartlett, PM Long, G Lugosi… - Proceedings of the …, 2020 - National Acad Sciences
The phenomenon of benign overfitting is one of the key mysteries uncovered by deep
learning methodology: deep neural networks seem to predict well, even with a perfect fit to …

The generalization error of random features regression: Precise asymptotics and the double descent curve

S Mei, A Montanari - Communications on Pure and Applied …, 2022 - Wiley Online Library
Deep learning methods operate in regimes that defy the traditional statistical mindset.
Neural network architectures often contain more parameters than training samples, and are …

Reconciling modern machine-learning practice and the classical bias–variance trade-off

M Belkin, D Hsu, S Ma… - Proceedings of the …, 2019 - National Acad Sciences
Breakthroughs in machine learning are rapidly changing science and society, yet our
fundamental understanding of this technology has lagged far behind. Indeed, one of the …

Fine-grained analysis of optimization and generalization for overparameterized two-layer neural networks

S Arora, S Du, W Hu, Z Li… - … Conference on Machine …, 2019 - proceedings.mlr.press
Recent works have cast some light on the mystery of why deep nets fit any data and
generalize despite being very overparametrized. This paper analyzes training and …

[HTML][HTML] Surprises in high-dimensional ridgeless least squares interpolation

T Hastie, A Montanari, S Rosset, RJ Tibshirani - Annals of statistics, 2022 - ncbi.nlm.nih.gov
Interpolators—estimators that achieve zero training error—have attracted growing attention
in machine learning, mainly because state-of-the art neural networks appear to be models of …