Robust reinforcement learning using offline data

K Panaganti, Z Xu, D Kalathil… - Advances in neural …, 2022 - proceedings.neurips.cc
The goal of robust reinforcement learning (RL) is to learn a policy that is robust against the
uncertainty in model parameters. Parameter uncertainty commonly occurs in many real …

Policy gradient method for robust reinforcement learning

Y Wang, S Zou - International conference on machine …, 2022 - proceedings.mlr.press
This paper develops the first policy gradient method with global optimality guarantee and
complexity analysis for robust reinforcement learning under model mismatch. Robust …

The curious price of distributional robustness in reinforcement learning with a generative model

L Shi, G Li, Y Wei, Y Chen… - Advances in Neural …, 2024 - proceedings.neurips.cc
This paper investigates model robustness in reinforcement learning (RL) via the framework
of distributionally robust Markov decision processes (RMDPs). Despite recent efforts, the …

Distributionally robust model-based offline reinforcement learning with near-optimal sample complexity

L Shi, Y Chi - Journal of Machine Learning Research, 2024 - jmlr.org
This paper concerns the central issues of model robustness and sample efficiency in offline
reinforcement learning (RL), which aims to learn to perform decision making from history …

Seeing is not believing: Robust reinforcement learning against spurious correlation

W Ding, L Shi, Y Chi, D Zhao - Advances in Neural …, 2024 - proceedings.neurips.cc
Robustness has been extensively studied in reinforcement learning (RL) to handle various
forms of uncertainty such as random perturbations, rare events, and malicious attacks. In this …

Toward theoretical understandings of robust markov decision processes: Sample complexity and asymptotics

W Yang, L Zhang, Z Zhang - The Annals of Statistics, 2022 - projecteuclid.org
Toward theoretical understandings of robust Markov decision processes: Sample
complexity and asymptotics Page 1 The Annals of Statistics 2022, Vol. 50, No. 6, 3223–3248 …

Twice regularized MDPs and the equivalence between robustness and regularization

E Derman, M Geist, S Mannor - Advances in Neural …, 2021 - proceedings.neurips.cc
Abstract Robust Markov decision processes (MDPs) aim to handle changing or partially
known system dynamics. To solve them, one typically resorts to robust optimization methods …

Fast bellman updates for wasserstein distributionally robust MDPs

Z Yu, L Dai, S Xu, S Gao, CP Ho - Advances in Neural …, 2024 - proceedings.neurips.cc
Markov decision processes (MDPs) often suffer from the sensitivity issue under model
ambiguity. In recent years, robust MDPs have emerged as an effective framework to …

Policy gradient in robust mdps with global convergence guarantee

Q Wang, CP Ho, M Petrik - International Conference on …, 2023 - proceedings.mlr.press
Abstract Robust Markov decision processes (RMDPs) provide a promising framework for
computing reliable policies in the face of model errors. Many successful reinforcement …

Robust reinforcement learning using least squares policy iteration with provable performance guarantees

KP Badrinath, D Kalathil - International Conference on …, 2021 - proceedings.mlr.press
This paper addresses the problem of model-free reinforcement learning for Robust Markov
Decision Process (RMDP) with large state spaces. The goal of the RMDPs framework is to …