论文标题
ListWise学习通过探索独特的评分来排名
Listwise Learning to Rank by Exploring Unique Ratings
论文作者
论文摘要
在本文中,我们提出了新的Listwise学习到级别模型,以减轻现有的模型。现有的ListWise学习到级模型通常来自经典的Plackett-Luce模型,该模型具有三个主要限制。 (1)其置换概率忽略了纽带,即,一个以上文档在查询方面具有相同的评分的情况。这可能导致不精确的排列概率和效率低下的培训,因为一一选择文档。 (2)它不利于具有很高相关性的文件。 (3)它有一个松散的假设,即不同步骤的采样文档是独立的。为了克服前两个限制,我们将基于唯一评分级别的降低顺序的唯一评分级别从候选设置中选择文档进行模型。训练的步骤数取决于唯一评分级别的数量。我们通过向所选文档分配高权重以优化标准化的折扣累积增益(NDCG),为整个加权分类任务序列提出了一个新的损失函数和相关的四个模型。为了克服最终限制,我们进一步提出了一种新颖而有效的方法,通过将改编的香草复发性神经网络(RNN)模型与以前步骤合并给定的选定文档的合并。我们编码由RNN模型选择的所有文档。在一个步骤中,我们使用RNN的最后一个单元格多次对所有文档进行相同的评分对所有文档进行排名。我们已经使用三个设置实施了模型:神经网络,具有梯度增强的神经网络以及具有梯度增强的回归树。我们已经在四个公共数据集上进行了实验。实验表明,这些模型尤其优于最先进的学习对级模型。
In this paper, we propose new listwise learning-to-rank models that mitigate the shortcomings of existing ones. Existing listwise learning-to-rank models are generally derived from the classical Plackett-Luce model, which has three major limitations. (1) Its permutation probabilities overlook ties, i.e., a situation when more than one document has the same rating with respect to a query. This can lead to imprecise permutation probabilities and inefficient training because of selecting documents one by one. (2) It does not favor documents having high relevance. (3) It has a loose assumption that sampling documents at different steps is independent. To overcome the first two limitations, we model ranking as selecting documents from a candidate set based on unique rating levels in decreasing order. The number of steps in training is determined by the number of unique rating levels. We propose a new loss function and associated four models for the entire sequence of weighted classification tasks by assigning high weights to the selected documents with high ratings for optimizing Normalized Discounted Cumulative Gain (NDCG). To overcome the final limitation, we further propose a novel and efficient way of refining prediction scores by combining an adapted Vanilla Recurrent Neural Network (RNN) model with pooling given selected documents at previous steps. We encode all of the documents already selected by an RNN model. In a single step, we rank all of the documents with the same ratings using the last cell of the RNN multiple times. We have implemented our models using three settings: neural networks, neural networks with gradient boosting, and regression trees with gradient boosting. We have conducted experiments on four public datasets. The experiments demonstrate that the models notably outperform state-of-the-art learning-to-rank models.