Multi-objective Relevance Ranking

In this paper, we introduce an Augmented Lagrangian basedmethod in a search relevance ranking algorithm to incorporate the multidimensional nature of relevance and business constraints, both of which are the requirements for building relevance ranking models in production. The off-the-shelf solutions cannot handle such complex objectives and therefore, modelers are left hand-tuning of parameters that have only indirect impact to the objectives, attempting to incorporate multiple objectives (MO) in a model. This process is time-consuming and tends to face sub-optimality. The proposed method is designed to systematically solve the MO problem in a constrained optimization framework, which is integrated with a popular Boosting algorithm and is, by all means, a novel contribution. Furthermore, we propose a procedure to specify the constraints to achieve business goals and the exploration scales linearly in the number of constraints, while existing methodology can scale exponentially. The experimental results show that the method successfully builds models that achieve MO criteria much more efficiently than existing methods. The potential impact includes significant reduction in model development time and allows for automation of model refresh even with presence of several MO criteria, in real world production system scale with hundreds of millions of records.

[1]  Christopher J. C. Burges,et al.  From RankNet to LambdaRank to LambdaMART: An Overview , 2010 .

[2]  Quoc V. Le,et al.  Learning to Rank with Nonsmooth Cost Functions , 2006, Neural Information Processing Systems.

[3]  Brian D. Davison,et al.  Learning to rank for freshness and relevance , 2011, SIGIR.

[4]  Stefano Mizzaro,et al.  How many relevances in information retrieval? , 1998, Interact. Comput..

[5]  Wei-Yin Loh,et al.  Classification and regression trees , 2011, WIREs Data Mining Knowl. Discov..

[6]  Hongyuan Zha,et al.  A General Boosting Method and its Application to Learning Ranking Functions for Web Search , 2007, NIPS.

[7]  Maya R. Gupta,et al.  Launch and Iterate: Reducing Prediction Churn , 2016, NIPS.

[8]  Maksims Volkovs,et al.  Learning to rank with multiple objective functions , 2011, WWW.

[9]  J. Friedman Greedy function approximation: A gradient boosting machine. , 2001 .

[10]  John Dines,et al.  A Multi-Objective Learning to re-Rank Approach to Optimize Online Marketplaces for Multiple Stakeholders , 2017, ArXiv.

[11]  Pia Borlund,et al.  The concept of relevance in IR , 2003, J. Assoc. Inf. Sci. Technol..

[12]  Daria Sorokina,et al.  Amazon Search: The Joy of Ranking Products , 2016, SIGIR.

[13]  Yi Chang,et al.  Learning to rank with multi-aspect relevance for vertical search , 2012, WSDM '12.

[14]  Tie-Yan Liu,et al.  LightGBM: A Highly Efficient Gradient Boosting Decision Tree , 2017, NIPS.

[15]  Matthew W. Hoffman,et al.  A General Framework for Constrained Bayesian Optimization using Information-based Search , 2015, J. Mach. Learn. Res..

[16]  Paul N. Bennett,et al.  Robust ranking models via risk-sensitive optimization , 2012, SIGIR '12.

[17]  Gilad Mishne,et al.  Towards recency ranking in web search , 2010, WSDM '10.

[18]  Tianqi Chen,et al.  XGBoost: A Scalable Tree Boosting System , 2016, KDD.

[19]  Matt J. Kusner,et al.  Bayesian Optimization with Inequality Constraints , 2014, ICML.