Advances in variational inference

C Zhang, J Bütepage, H Kjellström… - IEEE transactions on …, 2018 - ieeexplore.ieee.org
Many modern unsupervised or semi-supervised machine learning algorithms rely on
Bayesian probabilistic models. These models are usually intractable and thus require …

A review of the gumbel-max trick and its extensions for discrete stochasticity in machine learning

IAM Huijben, W Kool, MB Paulus… - IEEE transactions on …, 2022 - ieeexplore.ieee.org
The Gumbel-max trick is a method to draw a sample from a categorical distribution, given by
its unnormalized (log-) probabilities. Over the past years, the machine learning community …

NVAE: A deep hierarchical variational autoencoder

A Vahdat, J Kautz - Advances in neural information …, 2020 - proceedings.neurips.cc
Normalizing flows, autoregressive models, variational autoencoders (VAEs), and deep
energy-based models are among competing likelihood-based frameworks for deep …

Learning in the frequency domain

K Xu, M Qin, F Sun, Y Wang… - Proceedings of the …, 2020 - openaccess.thecvf.com
Deep neural networks have achieved remarkable success in computer vision tasks. Existing
neural networks mainly operate in the spatial domain with fixed input sizes. For practical …

Riemannian diffusion models

CW Huang, M Aghajohari, J Bose… - Advances in …, 2022 - proceedings.neurips.cc
Diffusion models are recent state-of-the-art methods for image generation and likelihood
estimation. In this work, we generalize continuous-time diffusion models to arbitrary …

Monte carlo gradient estimation in machine learning

S Mohamed, M Rosca, M Figurnov, A Mnih - Journal of Machine Learning …, 2020 - jmlr.org
This paper is a broad and accessible survey of the methods we have at our disposal for
Monte Carlo gradient estimation in machine learning and across the statistical sciences: the …

Learning Sparse Neural Networks through Regularization

C Louizos, M Welling, DP Kingma - arXiv preprint arXiv:1712.01312, 2017 - arxiv.org
We propose a practical method for $ L_0 $ norm regularization for neural networks: pruning
the network during training by encouraging weights to become exactly zero. Such …

Virtual adversarial training: a regularization method for supervised and semi-supervised learning

T Miyato, S Maeda, M Koyama… - IEEE transactions on …, 2018 - ieeexplore.ieee.org
We propose a new regularization method based on virtual adversarial loss: a new measure
of local smoothness of the conditional label distribution given input. Virtual adversarial loss …

Learning discrete structures for graph neural networks

L Franceschi, M Niepert, M Pontil… - … conference on machine …, 2019 - proceedings.mlr.press
Graph neural networks (GNNs) are a popular class of machine learning models that have
been successfully applied to a range of problems. Their major advantage lies in their ability …

Maskgan: better text generation via filling in the_

W Fedus, I Goodfellow, AM Dai - arXiv preprint arXiv:1801.07736, 2018 - arxiv.org
Neural text generation models are often autoregressive language models or seq2seq
models. These models generate text by sampling words sequentially, with each word …