A definition of continual reinforcement learning

D Abel, A Barreto, B Van Roy… - Advances in …, 2024 - proceedings.neurips.cc
In a standard view of the reinforcement learning problem, an agent's goal is to efficiently
identify a policy that maximizes long-term reward. However, this perspective is based on a …

Plastic: Improving input and label plasticity for sample efficient reinforcement learning

H Lee, H Cho, H Kim, D Gwak, J Kim… - Advances in …, 2024 - proceedings.neurips.cc
Abstract In Reinforcement Learning (RL), enhancing sample efficiency is crucial, particularly
in scenarios when data acquisition is costly and risky. In principle, off-policy RL algorithms …

Deep reinforcement learning with plasticity injection

E Nikishin, J Oh, G Ostrovski, C Lyle… - Advances in …, 2024 - proceedings.neurips.cc
A growing body of evidence suggests that neural networks employed in deep reinforcement
learning (RL) gradually lose their plasticity, the ability to learn from new data; however, the …

Improving language plasticity via pretraining with active forgetting

Y Chen, K Marchisio, R Raileanu… - Advances in …, 2023 - proceedings.neurips.cc
Pretrained language models (PLMs) are today the primary model for natural language
processing. Despite their impressive downstream performance, it can be difficult to apply …

Maintaining plasticity via regenerative regularization

S Kumar, H Marklund, B Van Roy - arXiv preprint arXiv:2308.11958, 2023 - arxiv.org
In continual learning, plasticity refers to the ability of an agent to quickly adapt to new
information. Neural networks are known to lose plasticity when processing non-stationary …

Drm: Mastering visual reinforcement learning through dormant ratio minimization

G Xu, R Zheng, Y Liang, X Wang, Z Yuan, T Ji… - arXiv preprint arXiv …, 2023 - arxiv.org
Visual reinforcement learning (RL) has shown promise in continuous control tasks. Despite
its progress, current algorithms are still unsatisfactory in virtually every aspect of the …

Disentangling the causes of plasticity loss in neural networks

C Lyle, Z Zheng, K Khetarpal, H van Hasselt… - arXiv preprint arXiv …, 2024 - arxiv.org
Underpinning the past decades of work on the design, initialization, and optimization of
neural networks is a seemingly innocuous assumption: that the network is trained on a\textit …

Continual learning as computationally constrained reinforcement learning

S Kumar, H Marklund, A Rao, Y Zhu, HJ Jeon… - arXiv preprint arXiv …, 2023 - arxiv.org
An agent that efficiently accumulates knowledge to develop increasingly sophisticated skills
over a long lifetime could advance the frontier of artificial intelligence capabilities. The …

Small batch deep reinforcement learning

J Obando Ceron, M Bellemare… - Advances in Neural …, 2024 - proceedings.neurips.cc
In value-based deep reinforcement learning with replay memories, the batch size parameter
specifies how many transitions to sample for each gradient update. Although critical to the …

COOM: a game benchmark for continual reinforcement learning

T Tomilin, M Fang, Y Zhang… - Advances in Neural …, 2024 - proceedings.neurips.cc
The advancement of continual reinforcement learning (RL) has been facing various
obstacles, including standardized metrics and evaluation protocols, demanding …