论文标题

对话中的推理:通过上下文改善响应生成阅读理解

Reasoning in Dialog: Improving Response Generation by Context Reading Comprehension

论文作者

Chen, Xiuying, Cui, Zhi, Zhang, Jiayi, Wei, Chen, Cui, Jianwei, Wang, Bin, Zhao, Dongyan, Yan, Rui

论文摘要

在多转话的对话框中,话语并不总是采取句子的完整形式\ cite {Carbonell1983discoursepa},这自然会使理解对话框上下文更加困难。但是,必须完全掌握对话框上下文以产生合理的响应。因此,在本文中,我们建议通过检查模型回答阅读理解问题的能力来提高响应生成性能,而该问题集中在对话框中的省略信息上。在多任务学习方案的启发下,我们提出了一个统一这两个任务的联合框架,共享相同的编码器,以与不同的解码器一起提取常见和任务不变的功能,以学习特定于任务的功能。为了更好地从问题和编码部分中的对话框历史记录中融合信息,我们建议使用内存更新器增强变压器体系结构,该内存旨在选择性地存储和更新历史记录对话框信息,以支持下游任务。对于实验,我们采用人类注释来编写和检查大规模对话框阅读理解数据集。在此数据集上进行了广泛的实验,结果表明,所提出的模型对这两个任务的几个强大基准都进行了实质性改进。这样,我们证明了推理确实可以帮助更好的响应产生,反之亦然。我们发布大规模数据集以进行进一步研究。

In multi-turn dialog, utterances do not always take the full form of sentences \cite{Carbonell1983DiscoursePA}, which naturally makes understanding the dialog context more difficult. However, it is essential to fully grasp the dialog context to generate a reasonable response. Hence, in this paper, we propose to improve the response generation performance by examining the model's ability to answer a reading comprehension question, where the question is focused on the omitted information in the dialog. Enlightened by the multi-task learning scheme, we propose a joint framework that unifies these two tasks, sharing the same encoder to extract the common and task-invariant features with different decoders to learn task-specific features. To better fusing information from the question and the dialog history in the encoding part, we propose to augment the Transformer architecture with a memory updater, which is designed to selectively store and update the history dialog information so as to support downstream tasks. For the experiment, we employ human annotators to write and examine a large-scale dialog reading comprehension dataset. Extensive experiments are conducted on this dataset, and the results show that the proposed model brings substantial improvements over several strong baselines on both tasks. In this way, we demonstrate that reasoning can indeed help better response generation and vice versa. We release our large-scale dataset for further research.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源