论文标题
通过变异信息瓶颈进行事件参数提取的多格式转移学习模型
A Multi-Format Transfer Learning Model for Event Argument Extraction via Variational Information Bottleneck
论文作者
论文摘要
事件参数提取(EAE)的目的是从文本中提取具有特定角色的参数,这些论点已在自然语言处理中广泛研究。以前的大多数作品在具有专用神经体系结构的特定EAE数据集中取得了良好的性能。鉴于,这些架构通常很难适应具有各种注释模式或格式的新数据集/方案。此外,他们依靠大规模标记的数据进行培训,由于大多数情况下的标签成本高,因此无法获得培训。在本文中,我们提出了一个具有变异信息瓶颈的多格式转移学习模型,该模型利用了信息,尤其是新数据集中EAE现有数据集中的常识。具体而言,我们引入了一个共享特定的及时框架,以从具有不同格式的数据集中学习格式共享和格式特定的知识。为了进一步吸收EAE的常识并消除了无关的噪音,我们将变分信息的瓶颈整合到我们的体系结构中以完善共享表示。我们在三个基准数据集上进行了广泛的实验,并在EAE上获得了新的最先进的性能。
Event argument extraction (EAE) aims to extract arguments with given roles from texts, which have been widely studied in natural language processing. Most previous works have achieved good performance in specific EAE datasets with dedicated neural architectures. Whereas, these architectures are usually difficult to adapt to new datasets/scenarios with various annotation schemas or formats. Furthermore, they rely on large-scale labeled data for training, which is unavailable due to the high labelling cost in most cases. In this paper, we propose a multi-format transfer learning model with variational information bottleneck, which makes use of the information especially the common knowledge in existing datasets for EAE in new datasets. Specifically, we introduce a shared-specific prompt framework to learn both format-shared and format-specific knowledge from datasets with different formats. In order to further absorb the common knowledge for EAE and eliminate the irrelevant noise, we integrate variational information bottleneck into our architecture to refine the shared representation. We conduct extensive experiments on three benchmark datasets, and obtain new state-of-the-art performance on EAE.