论文标题
持续学习:通过重播过程来解决深层神经网络中的灾难性遗忘
Continual Learning: Tackling Catastrophic Forgetting in Deep Neural Networks with Replay Processes
论文作者
论文摘要
人类一生都学到了一生。他们从一系列学习经验中积累知识,并记住基本概念,而不会忘记他们以前学到的知识。人工神经网络很难学习。他们通常依靠严格的预处理数据来学习针对特定问题(例如分类或回归)的解决方案。特别是,如果接受新的学习经验,他们会忘记过去的学习经历。因此,人工神经网络常常无能为力,以应对现实生活中的环境,例如自主机器人,必须在线学习以适应新的情况并克服新问题,而又不会忘记其过去的学习经验。持续学习(CL)是解决此类问题的机器学习的一个分支。持续的算法旨在在学习经验的课程中积累和改善知识,而无需忘记。在本文中,我们建议通过重播过程探索连续的算法。重播过程聚集了彩排方法和生成重播方法。生成重播包括通过生成模型来再生过去的学习经验来记住它们。排练包括从过去的学习经验中节省一大批样本,以稍后再排练。重播过程使得在优化当前的学习目标与过去的学习目标之间实现学习,而无需忘记任务设置的序列。我们证明它们是持续学习的非常有前途的方法。值得注意的是,它们能够重新评估过去的数据,并以新知识和来自不同学习经验的数据对抗。我们证明了他们通过无监督的学习,监督学习和强化学习任务不断学习的能力。
Humans learn all their life long. They accumulate knowledge from a sequence of learning experiences and remember the essential concepts without forgetting what they have learned previously. Artificial neural networks struggle to learn similarly. They often rely on data rigorously preprocessed to learn solutions to specific problems such as classification or regression. In particular, they forget their past learning experiences if trained on new ones. Therefore, artificial neural networks are often inept to deal with real-life settings such as an autonomous-robot that has to learn on-line to adapt to new situations and overcome new problems without forgetting its past learning-experiences. Continual learning (CL) is a branch of machine learning addressing this type of problem. Continual algorithms are designed to accumulate and improve knowledge in a curriculum of learning-experiences without forgetting. In this thesis, we propose to explore continual algorithms with replay processes. Replay processes gather together rehearsal methods and generative replay methods. Generative Replay consists of regenerating past learning experiences with a generative model to remember them. Rehearsal consists of saving a core-set of samples from past learning experiences to rehearse them later. The replay processes make possible a compromise between optimizing the current learning objective and the past ones enabling learning without forgetting in sequences of tasks settings. We show that they are very promising methods for continual learning. Notably, they enable the re-evaluation of past data with new knowledge and the confrontation of data from different learning-experiences. We demonstrate their ability to learn continually through unsupervised learning, supervised learning and reinforcement learning tasks.