Learning-To-Ensemble by Contextual Rank Aggregation in E-Commerce

Ensemble models in E-commerce combine predictions from multiple sub-models for ranking and revenue improvement. Industrial ensemble models are typically deep neural networks, following the supervised learning paradigm to infer conversion rate given inputs from sub-models. However, this process has the following two problems. Firstly, the point-wise scoring approach disregards the relationships between items and leads to homogeneous displayed results, while diversified display benefits user experience and revenue. Secondly, the learning paradigm focuses on the ranking metrics and does not directly optimize the revenue. In our work, we propose a new Learning-To-Ensemble (LTE) framework RAEGO, which replaces the ensemble model with a contextual Rank Aggregator (RA) and explores the best weights of sub-models by the Evaluator-Generator Optimization (EGO). To achieve the best online performance, we propose a new rank aggregation algorithm TournamentGreedy as a refinement of classic rank aggregators, which also produces the best average weighted Kendall Tau Distance (KTD) amongst all the considered algorithms with quadratic time complexity. Under the assumption that the best output list should be Pareto Optimal on the KTD metric for sub-models, we show that our RA algorithm has higher efficiency and coverage in exploring the optimal weights. Combined with the idea of Bayesian Optimization and gradient descent, we solve the online contextual Black-Box Optimization task that finds the optimal weights for sub-models given a chosen RA model. RA-EGO has been deployed in our online system and has improved the revenue significantly.

[1]  Anke van Zuylen,et al.  Rank Aggregation: Together We're Strong , 2009, ALENEX.

[2]  Tie-Yan Liu,et al.  Listwise approach to learning to rank: theory and algorithm , 2008, ICML '08.

[3]  Gregory N. Hullender,et al.  Learning to rank using gradient descent , 2005, ICML.

[4]  Fabio Stella,et al.  Contrasting Offline and Online Results when Evaluating Recommendation Algorithms , 2016, RecSys.

[5]  Pieter Spronck,et al.  Monte-Carlo Tree Search: A New Framework for Game AI , 2008, AIIDE.

[6]  Thorsten Joachims,et al.  Unbiased Learning-to-Rank with Biased Feedback , 2016, WSDM.

[7]  Ky Khac Vu,et al.  Surrogate-based methods for black-box optimization , 2017, Int. Trans. Oper. Res..

[8]  Arya Mazumdar,et al.  Efficient Rank Aggregation via Lehmer Codes , 2017, AISTATS.

[9]  Tie-Yan Liu,et al.  Learning to rank: from pairwise approach to listwise approach , 2007, ICML '07.

[10]  Cheng Soon Ong,et al.  Multivariate Spearman's rho for Aggregating Ranks Using Copulas , 2014, J. Mach. Learn. Res..

[11]  Tom Coleman,et al.  Ranking tournaments: Local search and a new algorithm , 2008, JEAL.

[12]  Huifeng Guo,et al.  PAL: a position-bias aware learning framework for CTR prediction in live recommender systems , 2019, RecSys.

[13]  Jöran Beel,et al.  A comparative analysis of offline and online evaluations and discussion of research paper recommender system evaluation , 2013, RepSys '13.

[14]  Sean M. McNee,et al.  Being accurate is not enough: how accuracy metrics have hurt recommender systems , 2006, CHI Extended Abstracts.

[15]  Gang Fu,et al.  Deep & Cross Network for Ad Click Predictions , 2017, ADKDD@KDD.

[16]  Joshua B. Tenenbaum,et al.  Structure Discovery in Nonparametric Regression through Compositional Kernel Search , 2013, ICML.

[17]  Qing Da,et al.  AliExpress Learning-to-Rank: Maximizing Online Model Performance Without Going Online , 2023, IEEE Transactions on Knowledge and Data Engineering.

[18]  Christopher J. C. Burges,et al.  From RankNet to LambdaRank to LambdaMART: An Overview , 2010 .

[19]  Alexandros Karatzoglou,et al.  RecoGym: A Reinforcement Learning Environment for the problem of Product Recommendation in Online Advertising , 2018, ArXiv.

[20]  Kevin Leyton-Brown,et al.  Sequential Model-Based Optimization for General Algorithm Configuration , 2011, LION.

[21]  Jian Tan,et al.  CobBO: Coordinate Backoff Bayesian Optimization , 2021, ArXiv.

[22]  Yuandong Tian,et al.  Learning Search Space Partition for Black-box Optimization using Monte Carlo Tree Search , 2020, NeurIPS.

[23]  Xiao Ma,et al.  Entire Space Multi-Task Model: An Effective Approach for Estimating Post-Click Conversion Rate , 2018, SIGIR.

[24]  Leslie Pack Kaelbling,et al.  Monte Carlo Tree Search in Continuous Spaces Using Voronoi Optimistic Optimization with Regret Bounds , 2020, AAAI.

[25]  Jasper Snoek,et al.  Practical Bayesian Optimization of Machine Learning Algorithms , 2012, NIPS.

[26]  D. Goldberg,et al.  BOA: the Bayesian optimization algorithm , 1999 .

[27]  O. Teytaud,et al.  Black-Box Optimization Revisited: Improving Algorithm Selection Wizards Through Massive Benchmarking , 2020, IEEE Transactions on Evolutionary Computation.

[28]  Amos J. Storkey,et al.  Exploration by Random Network Distillation , 2018, ICLR.

[29]  Yunming Ye,et al.  DeepFM: A Factorization-Machine based Neural Network for CTR Prediction , 2017, IJCAI.

[30]  Martin Pelikan,et al.  Bayesian Optimization Algorithm , 2005 .

[31]  Prabhat,et al.  Scalable Bayesian Optimization Using Deep Neural Networks , 2015, ICML.

[32]  Moni Naor,et al.  Rank aggregation methods for the Web , 2001, WWW '01.

[33]  Pengcheng Li,et al.  Improving Multi-Scenario Learning to Rank in E-commerce by Exploiting Task Relationships in the Label Space , 2020, CIKM.

[34]  David P. Williamson,et al.  Rank Aggregation: New Bounds for MCx , 2015, Discret. Appl. Math..

[35]  Jieming Zhu,et al.  Ensembled CTR Prediction via Knowledge Distillation , 2020, CIKM.