Learning a Deep Listwise Context Model for Ranking Refinement
release_occryvm5rzhcbjijfpue23m43e
by
Qingyao Ai, Keping Bi, Jiafeng Guo, W. Bruce Croft
2018
Abstract
Learning to rank has been intensively studied and widely applied in
information retrieval. Typically, a global ranking function is learned from a
set of labeled data, which can achieve good performance on average but may be
suboptimal for individual queries by ignoring the fact that relevant documents
for different queries may have different distributions in the feature space.
Inspired by the idea of pseudo relevance feedback where top ranked documents,
which we refer as the local ranking context, can provide important
information about the query's characteristics, we propose to use the inherent
feature distributions of the top results to learn a Deep Listwise Context Model
that helps us fine tune the initial ranked list. Specifically, we employ a
recurrent neural network to sequentially encode the top results using their
feature vectors, learn a local context model and use it to re-rank the top
results. There are three merits with our model: (1) Our model can capture the
local ranking context based on the complex interactions between top results
using a deep neural network; (2) Our model can be built upon existing
learning-to-rank methods by directly using their extracted feature vectors; (3)
Our model is trained with an attention-based loss function, which is more
effective and efficient than many existing listwise methods. Experimental
results show that the proposed model can significantly improve the
state-of-the-art learning to rank methods on benchmark retrieval corpora.
In text/plain
format
Archived Files and Locations
application/pdf 1.6 MB
file_ypcn3xxg7bazlj36fmtr6tebou
|
arxiv.org (repository) web.archive.org (webarchive) |
1804.05936v2
access all versions, variants, and formats of this works (eg, pre-prints)