论文标题
通过增强来改善元学习的概括
Improving Generalization in Meta-learning via Task Augmentation
论文作者
论文摘要
事实证明,元学习是一种有力的范式,可以将知识从以前的任务转移,以促进学习新任务的学习。当前的主导算法训练一个良好的模型初始化,该初始化通过支持集适用于每个任务。 CRUX在于优化初始化的概括能力,该功能是通过在每个任务的查询集上的改编模型的性能来衡量的。不幸的是,这种概括措施以经验结果证明,推动了初始化以使元训练任务过于努力,从而大大损害了对新任务的概括和适应性。为了解决这个问题,我们在评估概括时会积极增强使用“更多数据”的元训练任务。具体来说,我们提出了两种任务增强方法,包括metamix和Channel Shuffle。 MetAmix线性结合了来自支持和查询集的样品的功能和标签。对于每类样本,通道随机随机替换其通道的子集用来自其他类别的相应的通道。理论研究表明,任务增强如何改善元学习的概括。此外,MetAmix和Channel Shuffle的表现都超过了最先进的结果,但在许多数据集中的边距很大,并且与现有的元学习算法兼容。
Meta-learning has proven to be a powerful paradigm for transferring the knowledge from previous tasks to facilitate the learning of a novel task. Current dominant algorithms train a well-generalized model initialization which is adapted to each task via the support set. The crux lies in optimizing the generalization capability of the initialization, which is measured by the performance of the adapted model on the query set of each task. Unfortunately, this generalization measure, evidenced by empirical results, pushes the initialization to overfit the meta-training tasks, which significantly impairs the generalization and adaptation to novel tasks. To address this issue, we actively augment a meta-training task with "more data" when evaluating the generalization. Concretely, we propose two task augmentation methods, including MetaMix and Channel Shuffle. MetaMix linearly combines features and labels of samples from both the support and query sets. For each class of samples, Channel Shuffle randomly replaces a subset of their channels with the corresponding ones from a different class. Theoretical studies show how task augmentation improves the generalization of meta-learning. Moreover, both MetaMix and Channel Shuffle outperform state-of-the-art results by a large margin across many datasets and are compatible with existing meta-learning algorithms.