A survey on intrinsic motivation in reinforcement learning

A Aubret, L Matignon, S Hassas - arXiv preprint arXiv:1908.06976, 2019 - arxiv.org
The reinforcement learning (RL) research area is very active, with an important number of
new contributions; especially considering the emergent field of deep RL (DRL). However a …

Noveld: A simple yet effective exploration criterion

T Zhang, H Xu, X Wang, Y Wu… - Advances in …, 2021 - proceedings.neurips.cc
Efficient exploration under sparse rewards remains a key challenge in deep reinforcement
learning. Previous exploration methods (eg, RND) have achieved strong results in multiple …

Exploration in deep reinforcement learning: From single-agent to multiagent domain

J Hao, T Yang, H Tang, C Bai, J Liu… - … on Neural Networks …, 2023 - ieeexplore.ieee.org
Deep reinforcement learning (DRL) and deep multiagent reinforcement learning (MARL)
have achieved significant success across a wide range of domains, including game artificial …

Scaling map-elites to deep neuroevolution

C Colas, V Madhavan, J Huizinga, J Clune - Proceedings of the 2020 …, 2020 - dl.acm.org
Quality-Diversity (QD) algorithms, and MAP-Elites (ME) in particular, have proven very useful
for a broad range of applications including enabling real robots to recover quickly from joint …

Trajectory and communication design for cache-enabled UAVs in cellular networks: A deep reinforcement learning approach

J Ji, K Zhu, L Cai - IEEE Transactions on Mobile Computing, 2022 - ieeexplore.ieee.org
In this article, we investigate the content transmission in a heavy-crowded multiple access
cellular network, whose data traffic is offloaded through the combination of edge caching …

Made: Exploration via maximizing deviation from explored regions

T Zhang, P Rashidinejad, J Jiao… - Advances in …, 2021 - proceedings.neurips.cc
In online reinforcement learning (RL), efficient exploration remains particularly challenging
in high-dimensional environments with sparse rewards. In low-dimensional environments …

Bebold: Exploration beyond the boundary of explored regions

T Zhang, H Xu, X Wang, Y Wu, K Keutzer… - arXiv preprint arXiv …, 2020 - arxiv.org
Efficient exploration under sparse rewards remains a key challenge in deep reinforcement
learning. To guide exploration, previous work makes extensive use of intrinsic reward (IR) …

Unsupervised object interaction learning with counterfactual dynamics models

J Choi, S Lee, X Wang, S Sohn, H Lee - Proceedings of the AAAI …, 2024 - ojs.aaai.org
We present COIL (Counterfactual Object Interaction Learning), a novel way of learning skills
of object interactions on entity-centric environments. The goal is to learn primitive behaviors …

Variational dynamic for self-supervised exploration in deep reinforcement learning

C Bai, P Liu, K Liu, L Wang, Y Zhao… - IEEE Transactions on …, 2021 - ieeexplore.ieee.org
Efficient exploration remains a challenging problem in reinforcement learning, especially for
tasks where extrinsic rewards from environments are sparse or even totally disregarded …

Nuclear Norm Maximization Based Curiosity-Driven Reinforcement Learning

C Chen, Y Zhai, Z Gao, K Xu, S Yang… - IEEE Transactions …, 2023 - ieeexplore.ieee.org
Reinforcement learning (RL) has achieved promising results in solving numerous
challenging sequential decision problems. To address the issue of sparse extrinsic rewards …