On statistical bias in active learning: How and when to fix it
Active learning is a powerful tool when labelling data is expensive, but it introduces a bias
because the training data no longer follows the population distribution. We formalize this …
because the training data no longer follows the population distribution. We formalize this …
Target-aware Bayesian inference via generalized thermodynamic integration
F Llorente, L Martino, D Delgado - Computational Statistics, 2023 - Springer
In Bayesian inference, we are usually interested in the numerical approximation of integrals
that are posterior expectations or marginal likelihoods (aka, Bayesian evidence). In this …
that are posterior expectations or marginal likelihoods (aka, Bayesian evidence). In this …
Understanding and mitigating difficulties in posterior predictive evaluation
A Agrawal, J Domke - arXiv preprint arXiv:2405.19747, 2024 - arxiv.org
Predictive posterior densities (PPDs) are of interest in approximate Bayesian inference.
Typically, these are estimated by simple Monte Carlo (MC) averages using samples from the …
Typically, these are estimated by simple Monte Carlo (MC) averages using samples from the …
Daisee: Adaptive importance sampling by balancing exploration and exploitation
We study adaptive importance sampling (AIS) as an online learning problem and argue for
the importance of the trade‐off between exploration and exploitation in this adaptation …
the importance of the trade‐off between exploration and exploitation in this adaptation …
Expectation programming: Adapting probabilistic programming systems to estimate expectations efficiently
We show that the standard computational pipeline of probabilistic programming systems
(PPSs) can be inefficient for estimating expectations and introduce the concept of …
(PPSs) can be inefficient for estimating expectations and introduce the concept of …
Targeted active learning for bayesian decision-making
Active learning is usually applied to acquire labels of informative data points in supervised
learning, to maximize accuracy in a sample-efficient way. However, maximizing the accuracy …
learning, to maximize accuracy in a sample-efficient way. However, maximizing the accuracy …
Loss-calibrated expectation propagation for approximate Bayesian decision-making
Approximate Bayesian inference methods provide a powerful suite of tools for finding
approximations to intractable posterior distributions. However, machine learning …
approximations to intractable posterior distributions. However, machine learning …
Approximate Bayesian methods for optimal neural coding and decision-making
MJ Morais - 2021 - search.proquest.com
One fundamental goal of theoretical neuroscience is to understand the normative principles
governing the functional organization of neural circuits, and, in turn, to what extent they can …
governing the functional organization of neural circuits, and, in turn, to what extent they can …
[PDF][PDF] From Approximations to Decisions
J Sakaya - 2021 - helda.helsinki.fi
Bayesian models capture the intrinsic variability of a data-generating process as a posterior
distribution over the parameters of the model for the process. Decisions that are optimal for a …
distribution over the parameters of the model for the process. Decisions that are optimal for a …