[图书][B] Algorithms for decision making
A broad introduction to algorithms for decision making under uncertainty, introducing the
underlying mathematical problem formulations and the algorithms for solving them …
underlying mathematical problem formulations and the algorithms for solving them …
Online algorithms for POMDPs with continuous state, action, and observation spaces
Z Sunberg, M Kochenderfer - Proceedings of the International …, 2018 - ojs.aaai.org
Online solvers for partially observable Markov decision processes have been applied to
problems with large discrete state spaces, but continuous state, action, and observation …
problems with large discrete state spaces, but continuous state, action, and observation …
Approximate information state for approximate planning and reinforcement learning in partially observed systems
We propose a theoretical framework for approximate planning and learning in partially
observed systems. Our framework is based on the fundamental notion of information state …
observed systems. Our framework is based on the fundamental notion of information state …
Dynamic multi-robot task allocation under uncertainty and temporal constraints
We consider the problem of dynamically allocating tasks to multiple agents under time
window constraints and task completion uncertainty. Our objective is to minimize the number …
window constraints and task completion uncertainty. Our objective is to minimize the number …
Information particle filter tree: An online algorithm for pomdps with belief-based rewards on continuous domains
Abstract Planning in Partially Observable Markov Decision Processes (POMDPs) inherently
gathers the information necessary to act optimally under uncertainties. The framework can …
gathers the information necessary to act optimally under uncertainties. The framework can …
Reinforcement learning with state observation costs in action-contingent noiselessly observable markov decision processes
HJA Nam, S Fleming… - Advances in Neural …, 2021 - proceedings.neurips.cc
Many real-world problems that require making optimal sequences of decisions under
uncertainty involve costs when the agent wishes to obtain information about its environment …
uncertainty involve costs when the agent wishes to obtain information about its environment …
Bayesian optimized monte carlo planning
Online solvers for partially observable Markov decision processes have difficulty scaling to
problems with large action spaces. Monte Carlo tree search with progressive widening …
problems with large action spaces. Monte Carlo tree search with progressive widening …
Scalable information-theoretic path planning for a rover-helicopter team in uncertain environments
Mission-critical exploration of uncertain environments requires reliable and robust
mechanisms for achieving information gain. Typical measures of information gain such as …
mechanisms for achieving information gain. Typical measures of information gain such as …
Online planning for constrained POMDPs with continuous spaces through dual ascent
Rather than augmenting rewards with penalties for undesired behavior, Constrained
Partially Observable Markov Decision Processes (CPOMDPs) plan safely by imposing …
Partially Observable Markov Decision Processes (CPOMDPs) plan safely by imposing …
Online pomdp planning with anytime deterministic guarantees
M Barenboim, V Indelman - Advances in Neural Information …, 2024 - proceedings.neurips.cc
Autonomous agents operating in real-world scenarios frequently encounter uncertainty and
make decisions based on incomplete information. Planning under uncertainty can be …
make decisions based on incomplete information. Planning under uncertainty can be …