Bridging State and History Representations: Understanding Self-Predictive RL
Representations are at the core of all deep reinforcement learning (RL) methods for both
Markov decision processes (MDPs) and partially observable Markov decision processes …
Markov decision processes (MDPs) and partially observable Markov decision processes …
Connected and automated vehicles in mixed-traffic: Learning human driver behavior for effective on-ramp merging
Highway merging scenarios featuring mixed traffic conditions pose significant modeling and
control challenges for connected and automated vehicles (CAVs) interacting with incoming …
control challenges for connected and automated vehicles (CAVs) interacting with incoming …
Approximate information states for worst-case control and learning in uncertain systems
A Dave, N Venkatesh… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
In this paper, we investigate discrete-time decision-making problems in uncertain systems
with partially observed states. We consider a non-stochastic model, where uncontrolled …
with partially observed states. We consider a non-stochastic model, where uncontrolled …
Can direct latent model learning solve linear quadratic gaussian control?
We study the task of learning state representations from potentially high-dimensional
observations, with the goal of controlling an unknown partially observable system. We …
observations, with the goal of controlling an unknown partially observable system. We …
Agent-state based policies in POMDPs: Beyond belief-state MDPs
The traditional approach to POMDPs is to convert them into fully observed MDPs by
considering a belief state as an information state. However, a belief-state based approach …
considering a belief state as an information state. However, a belief-state based approach …
Information manipulation in partially observable markov decision processes
A common approach to solve partially observable Markov decision processes (POMDPs) is
transforming them into Makov decision processes (MDPs) defined on information states …
transforming them into Makov decision processes (MDPs) defined on information states …
Anticipating Oblivious Opponents in Stochastic Games
We present an approach for systematically anticipating the actions and policies employed
by\emph {oblivious} environments in concurrent stochastic games, while maximizing a …
by\emph {oblivious} environments in concurrent stochastic games, while maximizing a …
Asymmetric actor-critic with approximate information state
Reinforcement learning (RL) for partially observable Markov decision processes (POMDPs)
is a challenging problem because decisions need to be made based on the entire history of …
is a challenging problem because decisions need to be made based on the entire history of …
[图书][B] On Centralized and Decentralized Decision-Making Problems With Partial Information
AD Dave - 2023 - search.proquest.com
The advent of cyber-physical systems has revolutionized numerous applications, including
connected and automated vehicles, medicine and healthcare, the Internet of Things, social …
connected and automated vehicles, medicine and healthcare, the Internet of Things, social …
System-Scientific Foundations of Holonic Risk Analysis and Design in Socio-Cyber-Physical Systems
S Liu - 2024 - search.proquest.com
Within the context of decision-making, risk is a measure of potential harm to the decision-
maker's desired outcome, arising from the uncertainties associated with the likelihood of …
maker's desired outcome, arising from the uncertainties associated with the likelihood of …