作者
Théophane Weber, Sébastien Racanière, David P Reichert, Lars Buesing, Arthur Guez, Danilo Jimenez Rezende, Adria Puigdomenech Badia, Oriol Vinyals, Nicolas Heess, Yujia Li, Razvan Pascanu, Peter Battaglia, Demis Hassabis, David Silver, Daan Wierstra
发表日期
2017/7/19
期刊
arXiv preprint arXiv:1707.06203
简介
We introduce Imagination-Augmented Agents (I2As), a novel architecture for deep reinforcement learning combining model-free and model-based aspects. In contrast to most existing model-based reinforcement learning and planning methods, which prescribe how a model should be used to arrive at a policy, I2As learn to interpret predictions from a learned environment model to construct implicit plans in arbitrary ways, by using the predictions as additional context in deep policy networks. I2As show improved data efficiency, performance, and robustness to model misspecification compared to several baselines.
引用总数
2016201720182019202020212022202320242151021161261041107832
学术搜索中的文章
S Racanière, T Weber, D Reichert, L Buesing, A Guez… - Advances in neural information processing systems, 2017
T Weber, S Racaniere, DP Reichert, L Buesing, A Guez… - arXiv preprint arXiv:1707.06203, 2017