Learning to Learn without Forgetting By Maximizing Transfer and Minimizing Interference
Matthew Riemer
and
Ignacio Cases
and
Robert Ajemian
and
Miao Liu
and
Irina Rish
and
Yuhai Tu
and
Gerald Tesauro
arXiv e-Print archive - 2018 via Local arXiv
Keywords:
cs.LG, cs.AI, stat.ML
First published: 2018/10/29 (6 years ago) Abstract: Lack of performance when it comes to continual learning over non-stationary
distributions of data remains a major challenge in scaling neural network
learning to more human realistic settings. In this work we propose a new
conceptualization of the continual learning problem in terms of a trade-off
between transfer and interference. We then propose a new algorithm,
Meta-Experience Replay (MER), that directly exploits this view by combining
experience replay with optimization based meta-learning. This method learns
parameters that make interference based on future gradients less likely and
transfer based on future gradients more likely. We conduct experiments across
continual lifelong supervised learning benchmarks and non-stationary
reinforcement learning environments demonstrating that our approach
consistently outperforms recently proposed baselines for continual learning.
Our experiments show that the gap between the performance of MER and baseline
algorithms grows both as the environment gets more non-stationary and as the
fraction of the total experiences stored gets smaller.