Selective Experience Replay for Lifelong Learning

Type: Article

Publication Date: 2018-04-29

Citations: 290

DOI: https://doi.org/10.1609/aaai.v32i1.11595

Abstract

Deep reinforcement learning has emerged as a powerful tool for a variety of learning tasks, however deep nets typically exhibit forgetting when learning multiple tasks in sequence. To mitigate forgetting, we propose an experience replay process that augments the standard FIFO buffer and selectively stores experiences in a long-term memory. We explore four strategies for selecting which experiences will be stored: favoring surprise, favoring reward, matching the global training distribution, and maximizing coverage of the state space. We show that distribution matching successfully prevents catastrophic forgetting, and is consistently the best approach on all domains tested. While distribution matching has better and more consistent performance, we identify one case in which coverage maximization is beneficial---when tasks that receive less trained are more important. Overall, our results show that selective experience replay, when suitable selection algorithms are employed, can prevent catastrophic forgetting.

Locations

  • Proceedings of the AAAI Conference on Artificial Intelligence - View - PDF
  • arXiv (Cornell University) - View - PDF

Similar Works

Action Title Year Authors
+ Selective Experience Replay for Lifelong Learning 2018 David Isele
Akansel Cosgun
+ Experience Replay for Continual Learning 2018 David Rolnick
Arun Ahuja
Jonathan Schwarz
Timothy Lillicrap
Greg Wayne
+ Experience Replay for Continual Learning 2019 David Rolnick
Arun Ahuja
Jonathan Schwarz
Timothy Lillicrap
Gregory Wayne
+ Complementary Learning for Overcoming Catastrophic Forgetting Using Experience Replay 2019 Mohammad Rostami
Soheil Kolouri
Praveen K. Pilly
+ Prioritized Sequence Experience Replay 2019 Marc Brittain
Joshua R. Bertram
Xuxi Yang
Peng Wei
+ Complementary Learning for Overcoming Catastrophic Forgetting Using Experience Replay 2019 Mohammad Rostami
Soheil Kolouri
Praveen K. Pilly
+ Complementary Learning for Overcoming Catastrophic Forgetting Using Experience Replay 2019 Mohammad Rostami
Soheil Kolouri
Praveen K. Pilly
+ Memory Efficient Experience Replay for Streaming Learning 2018 Tyler L. Hayes
Nathan D. Cahill
Christopher Kanan
+ PDF Chat Memory Efficient Experience Replay for Streaming Learning 2019 Tyler L. Hayes
Nathan D. Cahill
Christopher Kanan
+ Prioritized Experience Replay 2015 Tom Schaul
John Quan
Ioannis Antonoglou
David Silver
+ Prioritized Experience Replay 2015 Tom Schaul
John Quan
Ioannis Antonoglou
David Silver
+ Continual Learning: Tackling Catastrophic Forgetting in Deep Neural Networks with Replay Processes 2020 Timothée Lesort
+ Learning to Learn without Forgetting by Maximizing Transfer and Minimizing Interference 2018 Matthew Riemer
Ignacio Cases
Robert Ajemian
Miao Liu
Irina Rish
Yuhai Tu
Gerald Tesauro
+ Learning to Learn without Forgetting by Maximizing Transfer and Minimizing Interference 2018 Matthew Riemer
Ignacio Cases
Robert Ajemian
Miao Liu
Irina Rish
Yuhai Tu
Gerald Tesauro
+ Map-based Experience Replay: A Memory-Efficient Solution to Catastrophic Forgetting in Reinforcement Learning 2023 Muhammad Burhan Hafez
Tilman Immisch
Tom Weber
Stefan Wermter
+ Memory-efficient Reinforcement Learning with Value-based Knowledge Consolidation 2022 Qingfeng Lan
Yangchen Pan
Jun Luo
A. Rupam Mahmood
+ PDF Chat Map-based experience replay: a memory-efficient solution to catastrophic forgetting in reinforcement learning 2023 Muhammad Burhan Hafez
Tilman Immisch
Tom Weber
Stefan Wermter
+ Policy Consolidation for Continual Reinforcement Learning 2019 Christos Kaplanis
Murray Shanahan
Claudia Clopath
+ PDF Chat Augmenting Replay in World Models for Continual Reinforcement Learning 2024 Luke Yang
Levin Kuhlmann
Gideon Kowadlo
+ AdaER: An Adaptive Experience Replay Approach for Continual Lifelong Learning 2023 Xingyu Li
Bo Tang
Haifeng Li

Works Cited by This (13)

Action Title Year Authors
+ Improving neural networks by preventing co-adaptation of feature detectors 2012 Geoffrey E. Hinton
Nitish Srivastava
Alex Krizhevsky
Ilya Sutskever
Ruslan Salakhutdinov
+ Prioritized Experience Replay 2015 Tom Schaul
John Quan
Ioannis Antonoglou
David Silver
+ Combating Reinforcement Learning's Sisyphean Curse with Intrinsic Fear 2016 Zachary C. Lipton
Jianfeng Gao
Lihong Li
Jianshu Chen
Li Deng
+ Overcoming catastrophic forgetting in neural networks 2017 James Kirkpatrick
Razvan Pascanu
Neil C. Rabinowitz
Joel Veness
Guillaume Desjardins
Andrei A. Rusu
Kieran Milan
John Quan
Tiago Ramalho
Agnieszka Grabska‐Barwińska
+ PDF Chat Deep Reinforcement Learning with Double Q-Learning 2016 Hado van Hasselt
Arthur Guez
David Silver
+ Reinforcement Learning with Unsupervised Auxiliary Tasks 2016 Max Jaderberg
Volodymyr Mnih
Wojciech Marian Czarnecki
Tom Schaul
Joel Z. Leibo
David Silver
Koray Kavukcuoglu
+ Continuous control with deep reinforcement learning 2016 Timothy Lillicrap
Jonathan J. Hunt
Alexander Pritzel
Nicolas Heess
Tom Erez
Yuval Tassa
David Silver
Daan Wierstra
+ An Empirical Investigation of Catastrophic Forgetting in Gradient-Based Neural Networks 2014 Ian Goodfellow
Mehdi Mirza
Xiao Da
Aaron Courville
Yoshua Bengio
+ PDF Chat A Deep Hierarchical Approach to Lifelong Learning in Minecraft 2017 Chen Tessler
Shahar Givony
Tom Zahavy
Daniel J. Mankowitz
Shie Mannor
+ On Learning to Think: Algorithmic Information Theory for Novel Combinations of Reinforcement Learning Controllers and Recurrent Neural World Models 2015 Juergen Schmidhuber