Multi-partite ranking with multi-class AdaBoost algorithm

The algorithms on learning to rank can traditionally be categorized as three classes including point-wise, pair-wise and list-wise. In our work, we focus on the regression-based method for the multi-partite ranking problems due to the efficiency of the point-wise methods. We proposed two ranking algorithms with the real AdaBoost and the discrete AdaBoost, which compute the expectation of the ratings with the estimation of the pseudo posterior probabilities. We found that it can be explained in the framework of the regression with the squared loss. It is more easily implemented than the previous McRank method since the algorithm adopts the decision stump as the weak leaner instead of the regression tree. In the fifteen benchmark datasets, our methods achieve better performance than the pair-wise method RankBoost under the C-index, NDCG and variant of NDCG measures. It has the lower training time complexity than RankBoost but the identical test time complexity.

[1]  Koby Crammer,et al.  Pranking with Ranking , 2001, NIPS.

[2]  Sunil J Rao,et al.  Regression Modeling Strategies: With Applications to Linear Models, Logistic Regression, and Survival Analysis , 2003 .

[3]  Eyke Hüllermeier,et al.  Binary Decomposition Methods for Multipartite Ranking , 2009, ECML/PKDD.

[4]  Rich Caruana,et al.  Obtaining Calibrated Probabilities from Boosting , 2005, UAI.

[5]  Radford M. Neal Pattern Recognition and Machine Learning , 2007, Technometrics.

[6]  Yoram Singer,et al.  Improved Boosting Algorithms Using Confidence-rated Predictions , 1998, COLT' 98.

[7]  Yoram Singer,et al.  BoosTexter: A Boosting-based System for Text Categorization , 2000, Machine Learning.

[8]  Thorsten Joachims,et al.  Optimizing search engines using clickthrough data , 2002, KDD.

[9]  Frank E. Harrell,et al.  Regression Modeling Strategies: With Applications to Linear Models, Logistic Regression, and Survival Analysis , 2001 .

[10]  Thomas Hofmann,et al.  Learning to Rank with Nonsmooth Cost Functions , 2006, NIPS.

[11]  Yoram Singer,et al.  An Efficient Boosting Algorithm for Combining Preferences by , 2013 .

[12]  Rong Jin,et al.  Learning to Rank by Optimizing NDCG Measure , 2009, NIPS.

[13]  Xinchang Zhang,et al.  Evaluating Web Content Quality via Multi-scale Features , 2013, ArXiv.

[14]  Balázs Kégl,et al.  Ranking by calibrated AdaBoost , 2010, Yahoo! Learning to Rank Challenge.

[15]  Qiang Wu,et al.  McRank: Learning to Rank Using Multiple Classification and Gradient Boosting , 2007, NIPS.