Biased stochastic first-order methods for conditional stochastic optimization and applications in meta learning
Conditional stochastic optimization covers a variety of applications ranging from invariant
learning and causal inference to meta-learning. However, constructing unbiased gradient …
learning and causal inference to meta-learning. However, constructing unbiased gradient …
MLMC techniques for discontinuous functions
MB Giles - International Conference on Monte Carlo and Quasi …, 2022 - Springer
Abstract The Multilevel Monte Carlo (MLMC) approach usually works well when estimating
the expected value of a quantity which is a Lipschitz function of intermediate quantities, but if …
the expected value of a quantity which is a Lipschitz function of intermediate quantities, but if …
[HTML][HTML] A review of efficient Multilevel Monte Carlo algorithms for derivative pricing and risk management
D Sinha, SP Chakrabarty - MethodsX, 2023 - Elsevier
In this article, we present a review of the recent developments on the topic of Multilevel
Monte Carlo (MLMC) algorithms, in the paradigm of applications in financial engineering …
Monte Carlo (MLMC) algorithms, in the paradigm of applications in financial engineering …
Risk-adaptive approaches to learning and decision making: A survey
JO Royset - arXiv preprint arXiv:2212.00856, 2022 - arxiv.org
Uncertainty is prevalent in engineering design, statistical learning, and decision making
broadly. Due to inherent risk-averseness and ambiguity about assumptions, it is common to …
broadly. Due to inherent risk-averseness and ambiguity about assumptions, it is common to …
Multilevel Monte Carlo approximation of functions
S Krumscheid, F Nobile - SIAM/ASA Journal on Uncertainty Quantification, 2018 - SIAM
Many applications across sciences and technologies require a careful quantification of
nondeterministic effects to a system output, for example, when evaluating the system's …
nondeterministic effects to a system output, for example, when evaluating the system's …
Multilevel Monte Carlo estimation of expected information gains
T Goda, T Hironaka, T Iwamoto - Stochastic Analysis and …, 2020 - Taylor & Francis
The expected information gain is an important quality criterion of Bayesian experimental
designs, which measures how much the information entropy about uncertain quantity of …
designs, which measures how much the information entropy about uncertain quantity of …
[PDF][PDF] Biased stochastic gradient descent for conditional stochastic optimization
Abstract Conditional Stochastic Optimization (CSO) covers a variety of applications ranging
from metalearning and causal inference to invariant learning. However, constructing …
from metalearning and causal inference to invariant learning. However, constructing …
Decision-making under uncertainty: using MLMC for efficient estimation of EVPPI
In this paper, we develop a very efficient approach to the Monte Carlo estimation of the
expected value of partial perfect information (EVPPI) that measures the average benefit of …
expected value of partial perfect information (EVPPI) that measures the average benefit of …
Multilevel double loop Monte Carlo and stochastic collocation methods with importance sampling for Bayesian optimal experimental design
An optimal experimental set‐up maximizes the value of data for statistical inferences. The
efficiency of strategies for finding optimal experimental set‐ups is particularly important for …
efficiency of strategies for finding optimal experimental set‐ups is particularly important for …
Unbiased multilevel Monte Carlo: Stochastic optimization, steady-state simulation, quantiles, and other applications
We present general principles for the design and analysis of unbiased Monte Carlo
estimators in a wide range of settings. Our estimators posses finite work-normalized …
estimators in a wide range of settings. Our estimators posses finite work-normalized …