论文标题

MLR:具有多任务学习的两阶段对话查询重写模型

MLR: A Two-stage Conversational Query Rewriting Model with Multi-task Learning

论文作者

Song, Shuangyong, Wang, Chao, Xie, Qianqian, Zu, Xinxing, Chen, Huan, Chen, Haiqing

论文摘要

会话上下文理解旨在认识到对话历史记录中用户的真实意图,这对于建立对话系统至关重要。但是,在开放型领域中的多转交谈理解仍然很具有挑战性,这需要系统提取重要信息并解决各种开放主题之间的依赖关系。在本文中,我们提出了对话性查询重写模型-MLR,它是序列标记和查询重写的多任务模型。 MLR将多转化的对话查询重新定义为单个转弯查询,该查询传达了用户简洁的真正意图,并减轻了多转向对话建模的困难。在模型中,我们将查询重写作为序列生成问题,并通过辅助Word类别标签预测任务介绍单词类别信息。为了培训我们的模型,我们构建了一个新的中国查询重写数据集并在其上进行实验。实验结果表明,我们的模型优于比较模型,并证明了单词类别信息在改善重写性能方面的有效性。

Conversational context understanding aims to recognize the real intention of user from the conversation history, which is critical for building the dialogue system. However, the multi-turn conversation understanding in open domain is still quite challenging, which requires the system extracting the important information and resolving the dependencies in contexts among a variety of open topics. In this paper, we propose the conversational query rewriting model - MLR, which is a Multi-task model on sequence Labeling and query Rewriting. MLR reformulates the multi-turn conversational queries into a single turn query, which conveys the true intention of users concisely and alleviates the difficulty of the multi-turn dialogue modeling. In the model, we formulate the query rewriting as a sequence generation problem and introduce word category information via the auxiliary word category label predicting task. To train our model, we construct a new Chinese query rewriting dataset and conduct experiments on it. The experimental results show that our model outperforms compared models, and prove the effectiveness of the word category information in improving the rewriting performance.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源