Advances in variational inference
Many modern unsupervised or semi-supervised machine learning algorithms rely on
Bayesian probabilistic models. These models are usually intractable and thus require …
Bayesian probabilistic models. These models are usually intractable and thus require …
A review of the gumbel-max trick and its extensions for discrete stochasticity in machine learning
IAM Huijben, W Kool, MB Paulus… - IEEE transactions on …, 2022 - ieeexplore.ieee.org
The Gumbel-max trick is a method to draw a sample from a categorical distribution, given by
its unnormalized (log-) probabilities. Over the past years, the machine learning community …
its unnormalized (log-) probabilities. Over the past years, the machine learning community …
NVAE: A deep hierarchical variational autoencoder
Normalizing flows, autoregressive models, variational autoencoders (VAEs), and deep
energy-based models are among competing likelihood-based frameworks for deep …
energy-based models are among competing likelihood-based frameworks for deep …
Learning in the frequency domain
Deep neural networks have achieved remarkable success in computer vision tasks. Existing
neural networks mainly operate in the spatial domain with fixed input sizes. For practical …
neural networks mainly operate in the spatial domain with fixed input sizes. For practical …
Riemannian diffusion models
Diffusion models are recent state-of-the-art methods for image generation and likelihood
estimation. In this work, we generalize continuous-time diffusion models to arbitrary …
estimation. In this work, we generalize continuous-time diffusion models to arbitrary …
Monte carlo gradient estimation in machine learning
This paper is a broad and accessible survey of the methods we have at our disposal for
Monte Carlo gradient estimation in machine learning and across the statistical sciences: the …
Monte Carlo gradient estimation in machine learning and across the statistical sciences: the …
Learning Sparse Neural Networks through Regularization
We propose a practical method for $ L_0 $ norm regularization for neural networks: pruning
the network during training by encouraging weights to become exactly zero. Such …
the network during training by encouraging weights to become exactly zero. Such …
Virtual adversarial training: a regularization method for supervised and semi-supervised learning
We propose a new regularization method based on virtual adversarial loss: a new measure
of local smoothness of the conditional label distribution given input. Virtual adversarial loss …
of local smoothness of the conditional label distribution given input. Virtual adversarial loss …
Learning discrete structures for graph neural networks
Graph neural networks (GNNs) are a popular class of machine learning models that have
been successfully applied to a range of problems. Their major advantage lies in their ability …
been successfully applied to a range of problems. Their major advantage lies in their ability …
Maskgan: better text generation via filling in the_
Neural text generation models are often autoregressive language models or seq2seq
models. These models generate text by sampling words sequentially, with each word …
models. These models generate text by sampling words sequentially, with each word …