Non-convex learning via replica exchange stochastic gradient mcmc
Abstract Replica exchange Monte Carlo (reMC), also known as parallel tempering, is an
important technique for accelerating the convergence of the conventional Markov Chain …
important technique for accelerating the convergence of the conventional Markov Chain …
An adaptive empirical Bayesian method for sparse deep learning
We propose a novel adaptive empirical Bayesian (AEB) method for sparse deep learning,
where the sparsity is ensured via a class of self-adaptive spike-and-slab priors. The …
where the sparsity is ensured via a class of self-adaptive spike-and-slab priors. The …
An adaptively weighted stochastic gradient MCMC algorithm for Monte Carlo simulation and global optimization
We propose an adaptively weighted stochastic gradient Langevin dynamics (AWSGLD)
algorithm for Bayesian learning of big data problems. The proposed algorithm is scalable …
algorithm for Bayesian learning of big data problems. The proposed algorithm is scalable …
A contour stochastic gradient langevin dynamics algorithm for simulations of multi-modal distributions
We propose an adaptively weighted stochastic gradient Langevin dynamics algorithm
(SGLD), so-called contour stochastic gradient Langevin dynamics (CSGLD), for Bayesian …
(SGLD), so-called contour stochastic gradient Langevin dynamics (CSGLD), for Bayesian …
On convergence of federated averaging langevin dynamics
We propose a federated averaging Langevin algorithm (FA-LD) for uncertainty quantification
and mean predictions with distributed clients. In particular, we generalize beyond normal …
and mean predictions with distributed clients. In particular, we generalize beyond normal …
Nonconvex sampling with the Metropolis-adjusted Langevin algorithm
O Mangoubi, NK Vishnoi - Conference on learning theory, 2019 - proceedings.mlr.press
Abstract The Langevin Markov chain algorithms are widely deployed methods to sample
from distributions in challenging high-dimensional and non-convex statistics and machine …
from distributions in challenging high-dimensional and non-convex statistics and machine …
Does Hamiltonian Monte Carlo mix faster than a random walk on multimodal densities?
Hamiltonian Monte Carlo (HMC) is a very popular and generic collection of Markov chain
Monte Carlo (MCMC) algorithms. One explanation for the popularity of HMC algorithms is …
Monte Carlo (MCMC) algorithms. One explanation for the popularity of HMC algorithms is …
Replica exchange for non-convex optimization
Gradient descent (GD) is known to converge quickly for convex objective functions, but it can
be trapped at local minima. On the other hand, Langevin dynamics (LD) can explore the …
be trapped at local minima. On the other hand, Langevin dynamics (LD) can explore the …
Polygonal Unadjusted Langevin Algorithms: Creating stable and efficient adaptive algorithms for neural networks
DY Lim, S Sabanis - Journal of Machine Learning Research, 2024 - jmlr.org
We present a new class of Langevin-based algorithms, which overcomes many of the known
shortcomings of popular adaptive optimizers that are currently used for the fine tuning of …
shortcomings of popular adaptive optimizers that are currently used for the fine tuning of …
Approximate optimization of convex functions with outlier noise
We study the problem of minimizing a convex function given by a zeroth order oracle that is
possibly corrupted by {\em outlier noise}. Specifically, we assume the function values at …
possibly corrupted by {\em outlier noise}. Specifically, we assume the function values at …