Dataset distillation with infinitely wide convolutional networks

T Nguyen, R Novak, L Xiao… - Advances in Neural …, 2021 - proceedings.neurips.cc
The effectiveness of machine learning algorithms arises from being able to extract useful
features from large amounts of data. As model and dataset sizes increase, dataset …

Dataset meta-learning from kernel ridge-regression

T Nguyen, Z Chen, J Lee - arXiv preprint arXiv:2011.00050, 2020 - arxiv.org
One of the most fundamental aspects of any machine learning algorithm is the training data
used by the algorithm. We introduce the novel concept of $\epsilon $-approximation of …

Finite versus infinite neural networks: an empirical study

J Lee, S Schoenholz, J Pennington… - Advances in …, 2020 - proceedings.neurips.cc
We perform a careful, thorough, and large scale empirical study of the correspondence
between wide neural networks and kernel methods. By doing so, we resolve a variety of …

On implicit bias in overparameterized bilevel optimization

P Vicol, JP Lorraine, F Pedregosa… - International …, 2022 - proceedings.mlr.press
Many problems in machine learning involve bilevel optimization (BLO), including
hyperparameter optimization, meta-learning, and dataset distillation. Bilevel problems …

Bayesian deep ensembles via the neural tangent kernel

B He, B Lakshminarayanan… - Advances in neural …, 2020 - proceedings.neurips.cc
We explore the link between deep ensembles and Gaussian processes (GPs) through the
lens of the Neural Tangent Kernel (NTK): a recent development in understanding the …

Analytic theory for the dynamics of wide quantum neural networks

J Liu, K Najafi, K Sharma, F Tacchino, L Jiang… - Physical Review Letters, 2023 - APS
Parametrized quantum circuits can be used as quantum neural networks and have the
potential to outperform their classical counterparts when trained for addressing learning …

Representation learning via quantum neural tangent kernels

J Liu, F Tacchino, JR Glick, L Jiang, A Mezzacapo - PRX Quantum, 2022 - APS
Variational quantum circuits are used in quantum machine learning and variational quantum
simulation tasks. Designing good variational circuits or predicting how well they perform for …

Feature learning in infinite-width neural networks

G Yang, EJ Hu - arXiv preprint arXiv:2011.14522, 2020 - arxiv.org
As its width tends to infinity, a deep neural network's behavior under gradient descent can
become simplified and predictable (eg given by the Neural Tangent Kernel (NTK)), if it is …

Global convergence of deep networks with one wide layer followed by pyramidal topology

QN Nguyen, M Mondelli - Advances in Neural Information …, 2020 - proceedings.neurips.cc
Recent works have shown that gradient descent can find a global minimum for over-
parameterized neural networks where the widths of all the hidden layers scale polynomially …

A spectral condition for feature learning

G Yang, JB Simon, J Bernstein - arXiv preprint arXiv:2310.17813, 2023 - arxiv.org
The push to train ever larger neural networks has motivated the study of initialization and
training at large network width. A key challenge is to scale training so that a network's …