论文标题
大规模多任务学习系统中任务的动态引入的进化方法
An Evolutionary Approach to Dynamic Introduction of Tasks in Large-scale Multitask Learning Systems
论文作者
论文摘要
多任务学习假设能够从多个任务中学习的模型可以通过知识转移来实现更好的质量和效率,这是人类学习的关键特征。但是,最先进的ML模型依赖于每个任务的高自定义,并利用大小和数据量表,而不是扩展任务数量。同样,持续的学习将时间表添加到多任务中,通常集中于研究诸如灾难性遗忘之类的常见陷阱,而不是大规模研究,作为建立下一代人工智能的关键组成部分。我们提出了一种能够生成支持新任务的大型多任务模型的进化方法。生成的多任务模型被稀疏地激活,并集成了基于任务的路由,随着模型的扩展,可以保证有限的计算成本和更少的添加参数。该方法依赖于知识隔室化技术,以实现抗灾难性遗忘和其他常见陷阱(如渐变的渐进交流和负转移)的免疫力。我们从经验上证明,所提出的方法可以在69个公共图像分类任务上共同解决并取得竞争成果,例如,通过与公共数据培训的最佳模型相比,通过实现15%的相对误差来改善竞争性基准(例如CIFAR10),改善最新技术的状态。
Multitask learning assumes that models capable of learning from multiple tasks can achieve better quality and efficiency via knowledge transfer, a key feature of human learning. Though, state of the art ML models rely on high customization for each task and leverage size and data scale rather than scaling the number of tasks. Also, continual learning, that adds the temporal aspect to multitask, is often focused to the study of common pitfalls such as catastrophic forgetting instead of being studied at a large scale as a critical component to build the next generation artificial intelligence.We propose an evolutionary method capable of generating large scale multitask models that support the dynamic addition of new tasks. The generated multitask models are sparsely activated and integrates a task-based routing that guarantees bounded compute cost and fewer added parameters per task as the model expands.The proposed method relies on a knowledge compartmentalization technique to achieve immunity against catastrophic forgetting and other common pitfalls such as gradient interference and negative transfer. We demonstrate empirically that the proposed method can jointly solve and achieve competitive results on 69public image classification tasks, for example improving the state of the art on a competitive benchmark such as cifar10 by achieving a 15% relative error reduction compared to the best model trained on public data.