论文标题
无偏见的TOP-K学习以因果可能分解的排名
Unbiased Top-k Learning to Rank with Causal Likelihood Decomposition
论文作者
论文摘要
已经提出了公正的学习排名,以减轻搜索排名中的偏见,从而可以使用用户交互数据训练排名模型。在实际应用程序中,搜索引擎旨在仅显示检索到的候选人集中最相关的K文档。休息的候选人被丢弃。结果,位置偏差和样本选择偏差通常同时发生。现有的无偏学习将排名的方法侧重于一种类型的偏见(例如,位置偏见),或者通过单独的组件减轻位置偏差和样本选择偏差,从而俯瞰其关联。在这项研究中,我们首先从因果图的角度分析了位置偏差和样本选择偏差的机制和关联。基于分析,我们提出了因果可能性分解(CLD),这是一种统一的方法,可以同时减轻TOP-K学习排名中的这两个偏见。通过将偏置数据的对数可能性分解为仅与相关性相关的公正术语,以及与偏见有关的其他术语,CLD成功地将相关性从位置偏差和样本选择偏见中分离出来。通过最大化整个可能性,可以从无偏的项中获得公正的排名模型。还开发了成对神经排名的扩展。 CLD的优势包括理论声音和一个统一的框架,用于排名排名。广泛的实验结果验证了CLD(包括其成对神经延伸),通过减轻位置偏置和样品选择偏差来超过基准。经验研究还表明,CLD对偏置严重程度和点击噪声的变化具有鲁棒性。
Unbiased learning to rank has been proposed to alleviate the biases in the search ranking, making it possible to train ranking models with user interaction data. In real applications, search engines are designed to display only the most relevant k documents from the retrieved candidate set. The rest candidates are discarded. As a consequence, position bias and sample selection bias usually occur simultaneously. Existing unbiased learning to rank approaches either focus on one type of bias (e.g., position bias) or mitigate the position bias and sample selection bias with separate components, overlooking their associations. In this study, we first analyze the mechanisms and associations of position bias and sample selection bias from the viewpoint of a causal graph. Based on the analysis, we propose Causal Likelihood Decomposition (CLD), a unified approach to simultaneously mitigating these two biases in top-k learning to rank. By decomposing the log-likelihood of the biased data as an unbiased term that only related to relevance, plus other terms related to biases, CLD successfully detaches the relevance from position bias and sample selection bias. An unbiased ranking model can be obtained from the unbiased term, via maximizing the whole likelihood. An extension to the pairwise neural ranking is also developed. Advantages of CLD include theoretical soundness and a unified framework for pointwise and pairwise unbiased top-k learning to rank. Extensive experimental results verified that CLD, including its pairwise neural extension, outperformed the baselines by mitigating both the position bias and the sample selection bias. Empirical studies also showed that CLD is robust to the variation of bias severity and the click noise.