论文标题
可解释的持续学习与分配变化的基础的混合
Mixture of basis for interpretable continual learning with distribution shifts
论文作者
论文摘要
对于几个现实世界应用程序,随着数据分布的转移环境的持续学习是一个具有挑战性的问题。在本文中,我们考虑了数据分布(任务)突然变化的设置,而这些转移的时机尚不清楚。此外,我们考虑了一个半监督的任务无关环境,其中学习算法可以访问以下任务细分和未分段的数据以进行离线培训。我们提出了一种新的方法,称为基本模型(MOB)的混合物,以解决此问题设置。核心思想是学习一小部分基础模型,并构建模型的动态,任务依赖性混合物,以预测当前任务。我们还提出了一种新方法,以检测有关现有基础模型过失并根据需要实例化新模型的观测值。我们在多个域中测试我们的方法,并表明它比现有方法在大多数情况下的预测错误更好,而使用模型少于其他多种模型方法。此外,我们分析了由MOB学到的潜在任务表示形式,并表明类似的任务倾向于聚集在潜在空间中,并且当任务不同时,潜在表示在任务边界上会转移。
Continual learning in environments with shifting data distributions is a challenging problem with several real-world applications. In this paper we consider settings in which the data distribution(task) shifts abruptly and the timing of these shifts are not known. Furthermore, we consider a semi-supervised task-agnostic setting in which the learning algorithm has access to both task-segmented and unsegmented data for offline training. We propose a novel approach called mixture of Basismodels (MoB) for addressing this problem setting. The core idea is to learn a small set of basis models and to construct a dynamic, task-dependent mixture of the models to predict for the current task. We also propose a new methodology to detect observations that are out-of-distribution with respect to the existing basis models and to instantiate new models as needed. We test our approach in multiple domains and show that it attains better prediction error than existing methods in most cases while using fewer models than other multiple model approaches. Moreover, we analyze the latent task representations learned by MoB and show that similar tasks tend to cluster in the latent space and that the latent representation shifts at the task boundaries when tasks are dissimilar.