资源论文Complementary Learning for Overcoming Catastrophic Forgetting Using Experience Replay

Complementary Learning for Overcoming Catastrophic Forgetting Using Experience Replay

2019-10-09 | |  39 |   36 |   0
Abstract Despite huge success, deep networks are unable to learn effectively in sequential multitask learning settings as they forget the past learned tasks after learning new tasks. Inspired from complementary learning systems theory, we address this challenge by learning a generative model that couples the current task to the past learned tasks through a discriminative embedding space. We learn an abstract generative distribution in the embedding that allows generation of data points to represent past experience. We sample from this distribution and utilize experience replay to avoid forgetting and simultaneously accumulate new knowledge to the abstract distribution in order to couple the current task with past experience. We demonstrate theoretically and empirically that our framework learns a distribution in the embedding, which is shared across all tasks, and as a result tackles catastrophic forgetting

上一篇:Coarse-to-Fine Image Inpainting via Region-wise Convolutions and Non-Local Correlation

下一篇:DeepCU: Integrating both Common and Unique Latent Information for Multimodal Sentiment Analysis

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...