Rvs: What is essential for offline rl via supervised learning?

S Emmons, B Eysenbach, I Kostrikov… - arXiv preprint arXiv …, 2021 - arxiv.org
Recent work has shown that supervised learning alone, without temporal difference (TD)
learning, can be remarkably effective for offline RL. When does this hold true, and which …

Waypoint transformer: Reinforcement learning via supervised learning with intermediate targets

A Badrinath, Y Flet-Berliac, A Nie… - Advances in Neural …, 2024 - proceedings.neurips.cc
Despite the recent advancements in offline reinforcement learning via supervised learning
(RvS) and the success of the decision transformer (DT) architecture in various domains, DTs …

Model-based offline policy optimization with adversarial network

J Yang, X Chen, S Wang, B Zhang - ECAI 2023, 2023 - ebooks.iospress.nl
Abstract Model-based offline reinforcement learning (RL), which builds a supervised
transition model with logging dataset to avoid costly interactions with the online …

Bootstrapping Adaptive Human-Machine Interfaces with Offline Reinforcement Learning

J Gao, S Reddy, G Berseth, AD Dragan… - 2023 IEEE/RSJ …, 2023 - ieeexplore.ieee.org
Adaptive interfaces can help users perform sequential decision-making tasks like robotic
teleoperation given noisy, high-dimensional command signals (eg, from a brain-computer …

State-Aware Proximal Pessimistic Algorithms for Offline Reinforcement Learning

C Chen, H Tang, Y Ma, C Wang, Q Shen, D Li… - arXiv preprint arXiv …, 2022 - arxiv.org
Pessimism is of great importance in offline reinforcement learning (RL). One broad category
of offline RL algorithms fulfills pessimism by explicit or implicit behavior regularization …