Learning to Rank by Optimizing NDCG Measure

Abstract

Learning to rank is a relatively new field of study, aiming to learn a ranking function from a set of training data with relevancy labels. The ranking algorithms are often evaluated using information retrieval measures, such as Normalized Discounted Cumulative Gain (NDCG) [1] and Mean Average Precision (MAP) [2]. Until recently, most learning to rank algorithms were not using a loss function related to the above mentioned evaluation measures. The main difficulty in direct optimization of these measures is that they depend on the ranks of documents, not the numerical values output by the ranking function. We propose a probabilistic framework that addresses this challenge by optimizing the expectation of NDCG over all the possible permutations of documents. A relaxation strategy is used to approximate the average of NDCG over the space of permutation, and a bound optimization approach is proposed to make the computation efficient. Extensive experiments show that the proposed algorithm outperforms state-of-the-art ranking algorithms on several benchmark data sets.

Extracted Key Phrases

1 Figure or Table

010203020102011201220132014201520162017
Citations per Year

104 Citations

Semantic Scholar estimates that this publication has 104 citations based on the available data.

See our FAQ for additional information.

Cite this paper

@inproceedings{Valizadegan2009LearningTR, title={Learning to Rank by Optimizing NDCG Measure}, author={Hamed Valizadegan and Rong Jin and Ruofei Zhang and Jianchang Mao}, booktitle={NIPS}, year={2009} }