论文标题
TRGP:信任区域的渐进率投影持续学习
TRGP: Trust Region Gradient Projection for Continual Learning
论文作者
论文摘要
灾难性遗忘是持续学习的主要挑战之一。为了解决此问题,一些现有方法对新任务的优化空间施加了限制性限制,以最大程度地减少对旧任务的干扰。但是,这可能会导致新任务的性能不令人满意,尤其是当新任务与旧任务密切相关时。为了应对这一挑战,我们提出了信任区域梯度投影(TRGP),以持续学习,以促进基于任务相关性的有效表征的前向知识转移。特别是,我们介绍了“信任区域”的概念,以使用梯度投影的规范在任务输入跨越的子空间上选择新任务的最相关的旧任务。然后,提出了缩放权重投影,以巧妙地通过层缩放矩阵巧妙地重复信任区域中选定的旧任务的冷冻权重。通过共同优化缩放矩阵和模型,该模型沿着与旧任务子空间的方向更新,TRGP可以有效地提示知识传输而不会忘记。广泛的实验表明,我们的方法比相关的最新方法实现了显着改善。
Catastrophic forgetting is one of the major challenges in continual learning. To address this issue, some existing methods put restrictive constraints on the optimization space of the new task for minimizing the interference to old tasks. However, this may lead to unsatisfactory performance for the new task, especially when the new task is strongly correlated with old tasks. To tackle this challenge, we propose Trust Region Gradient Projection (TRGP) for continual learning to facilitate the forward knowledge transfer based on an efficient characterization of task correlation. Particularly, we introduce a notion of `trust region' to select the most related old tasks for the new task in a layer-wise and single-shot manner, using the norm of gradient projection onto the subspace spanned by task inputs. Then, a scaled weight projection is proposed to cleverly reuse the frozen weights of the selected old tasks in the trust region through a layer-wise scaling matrix. By jointly optimizing the scaling matrices and the model, where the model is updated along the directions orthogonal to the subspaces of old tasks, TRGP can effectively prompt knowledge transfer without forgetting. Extensive experiments show that our approach achieves significant improvement over related state-of-the-art methods.