Optimal experimental design for infinite-dimensional Bayesian inverse problems governed by PDEs: A review
A Alexanderian - Inverse Problems, 2021 - iopscience.iop.org
We present a review of methods for optimal experimental design (OED) for Bayesian inverse
problems governed by partial differential equations with infinite-dimensional parameters …
problems governed by partial differential equations with infinite-dimensional parameters …
Gpytorch: Blackbox matrix-matrix gaussian process inference with gpu acceleration
Despite advances in scalable models, the inference tools used for Gaussian processes
(GPs) have yet to fully capitalize on developments in computing hardware. We present an …
(GPs) have yet to fully capitalize on developments in computing hardware. We present an …
Pyhessian: Neural networks through the lens of the hessian
We present PYHESSIAN, a new scalable framework that enables fast computation of
Hessian (ie, second-order derivative) information for deep neural networks. PYHESSIAN …
Hessian (ie, second-order derivative) information for deep neural networks. PYHESSIAN …
Hawq-v2: Hessian aware trace-weighted quantization of neural networks
Quantization is an effective method for reducing memory footprint and inference time of
Neural Networks. However, ultra low precision quantization could lead to significant …
Neural Networks. However, ultra low precision quantization could lead to significant …
How to train your neural ODE: the world of Jacobian and kinetic regularization
Training neural ODEs on large datasets has not been tractable due to the necessity of
allowing the adaptive numerical ODE solver to refine its step size to very small values. In …
allowing the adaptive numerical ODE solver to refine its step size to very small values. In …
Gradient norm aware minimization seeks first-order flatness and improves generalization
Recently, flat minima are proven to be effective for improving generalization and sharpness-
aware minimization (SAM) achieves state-of-the-art performance. Yet the current definition of …
aware minimization (SAM) achieves state-of-the-art performance. Yet the current definition of …
Generative data augmentation for commonsense reasoning
Recent advances in commonsense reasoning depend on large-scale human-annotated
training data to achieve peak performance. However, manual curation of training examples …
training data to achieve peak performance. However, manual curation of training examples …
There are many consistent explanations of unlabeled data: Why you should average
Presently the most successful approaches to semi-supervised learning are based on
consistency regularization, whereby a model is trained to be robust to small perturbations of …
consistency regularization, whereby a model is trained to be robust to small perturbations of …
The hessian penalty: A weak prior for unsupervised disentanglement
Existing disentanglement methods for deep generative models rely on hand-picked priors
and complex encoder-based architectures. In this paper, we propose the Hessian Penalty, a …
and complex encoder-based architectures. In this paper, we propose the Hessian Penalty, a …
Network quantization with element-wise gradient scaling
Network quantization aims at reducing bit-widths of weights and/or activations, particularly
important for implementing deep neural networks with limited hardware resources. Most …
important for implementing deep neural networks with limited hardware resources. Most …