Overcoming catastrophic forgetting in neural networks.

Journal: Proceedings of the National Academy of Sciences of the United States of America
Published Date:

Abstract

The ability to learn tasks in a sequential fashion is crucial to the development of artificial intelligence. Until now neural networks have not been capable of this and it has been widely thought that catastrophic forgetting is an inevitable feature of connectionist models. We show that it is possible to overcome this limitation and train networks that can maintain expertise on tasks that they have not experienced for a long time. Our approach remembers old tasks by selectively slowing down learning on the weights important for those tasks. We demonstrate our approach is scalable and effective by solving a set of classification tasks based on a hand-written digit dataset and by learning several Atari 2600 games sequentially.

Authors

  • James Kirkpatrick
    DeepMind, London EC4 5TW, United Kingdom; kirkpatrick@google.com.
  • Razvan Pascanu
    DeepMind, London EC4 5TW, United Kingdom.
  • Neil Rabinowitz
    DeepMind, London EC4 5TW, United Kingdom.
  • Joel Veness
    Google DeepMind, 5 New Street Square, London EC4A 3TW, UK.
  • Guillaume Desjardins
    DeepMind, London EC4 5TW, United Kingdom.
  • Andrei A Rusu
    Google DeepMind, 5 New Street Square, London EC4A 3TW, UK.
  • Kieran Milan
    DeepMind, London EC4 5TW, United Kingdom.
  • John Quan
    DeepMind, London EC4 5TW, United Kingdom.
  • Tiago Ramalho
    DeepMind, London EC4 5TW, United Kingdom.
  • Agnieszka Grabska-Barwinska
    DeepMind, London EC4 5TW, United Kingdom.
  • Demis Hassabis
    Google DeepMind, 5 New Street Square, London EC4A 3TW, UK.
  • Claudia Clopath
    Department of Bioengineering, Imperial College London, Royal School of Mines, London, SW7 2AZ, UK. c.clopath@imperial.ac.uk.
  • Dharshan Kumaran
    Google DeepMind, 5 New Street Square, London EC4A 3TW, UK.
  • Raia Hadsell
    DeepMind, London EC4 5TW, United Kingdom.