论文标题
分析和从用户互动中学习以进行搜索澄清
Analyzing and Learning from User Interactions for Search Clarification
论文作者
论文摘要
提出澄清问题以响应搜索查询已被认为是揭示查询基本意图的有用技术。澄清在具有不同界面的检索系统中具有应用程序,从传统的Web搜索接口到有限的带宽接口,例如在语音和小屏幕设备中。最近在文献中研究了澄清问题的生成和评估。但是,使用澄清问题的用户互动相对尚未探索。在本文中,我们通过分析主要的Web搜索引擎中的澄清问题来进行全面的研究。更详细地,我们通过根据搜索查询的不同属性,澄清问题及其候选答案来澄清问题来分析收到的用户参与。我们进一步研究数据中的点击偏见,并表明,即使阅读澄清的问题和候选答案并没有付出重大努力,但数据中仍然存在某种位置和表现偏见。我们还提出了一个模型,用于学习表示基于用户互动数据作为隐式反馈的问题。该模型用于重新排列许多自动生成的澄清问题,以供给定查询。对点击数据和人类标记数据的评估都证明了所提出的方法的高质量。
Asking clarifying questions in response to search queries has been recognized as a useful technique for revealing the underlying intent of the query. Clarification has applications in retrieval systems with different interfaces, from the traditional web search interfaces to the limited bandwidth interfaces as in speech-only and small screen devices. Generation and evaluation of clarifying questions have been recently studied in the literature. However, user interaction with clarifying questions is relatively unexplored. In this paper, we conduct a comprehensive study by analyzing large-scale user interactions with clarifying questions in a major web search engine. In more detail, we analyze the user engagements received by clarifying questions based on different properties of search queries, clarifying questions, and their candidate answers. We further study click bias in the data, and show that even though reading clarifying questions and candidate answers does not take significant efforts, there still exist some position and presentation biases in the data. We also propose a model for learning representation for clarifying questions based on the user interaction data as implicit feedback. The model is used for re-ranking a number of automatically generated clarifying questions for a given query. Evaluation on both click data and human labeled data demonstrates the high quality of the proposed method.