论文标题

具有成本效益的互动注意力学习与神经注意力的过程

Cost-effective Interactive Attention Learning with Neural Attention Processes

论文作者

Heo, Jay, Park, Junhyeon, Jeong, Hyewon, Kim, Kwang Joon, Lee, Juho, Yang, Eunho, Hwang, Sung Ju

论文摘要

我们提出了一个新颖的互动学习框架,我们称之为交互式注意力学习(IAL),其中人类主管互动操纵分配的注意力,以通过更新引起注意力的网络来纠正模型的行为。但是,由于人类注释的稀缺性,这种模型很容易过度拟合,并且需要昂贵的重新训练。此外,人类注释者几乎不可避免地检查对大量实例和特征的关注。我们通过提出样本有效的注意机制和具有成本效益的重读算法来应对这些挑战。首先,我们提出神经注意力过程(NAP),这是一个注意力产生者,可以通过纳入新的注意力级别的监督而无需任何重新训练来更新其行为。其次,我们提出了一种算法,该算法通过其负面影响来优先考虑实例和特征,以便该模型可以通过最少的人类反馈产生大量改进。我们从多个领域(医疗保健,房地产和计算机视觉)上对各种时间序列数据集进行了验证,在这些数据集中,它在上面大大超过了基本线,具有常规的关注机制,或者没有具有成本效益的重新依赖性,其重新培训和人类模型的相互作用成本要少得多。

We propose a novel interactive learning framework which we refer to as Interactive Attention Learning (IAL), in which the human supervisors interactively manipulate the allocated attentions, to correct the model's behavior by updating the attention-generating network. However, such a model is prone to overfitting due to scarcity of human annotations, and requires costly retraining. Moreover, it is almost infeasible for the human annotators to examine attentions on tons of instances and features. We tackle these challenges by proposing a sample-efficient attention mechanism and a cost-effective reranking algorithm for instances and features. First, we propose Neural Attention Process (NAP), which is an attention generator that can update its behavior by incorporating new attention-level supervisions without any retraining. Secondly, we propose an algorithm which prioritizes the instances and the features by their negative impacts, such that the model can yield large improvements with minimal human feedback. We validate IAL on various time-series datasets from multiple domains (healthcare, real-estate, and computer vision) on which it significantly outperforms baselines with conventional attention mechanisms, or without cost-effective reranking, with substantially less retraining and human-model interaction cost.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源