Short-Term Satisfaction and Long-Term Coverage: Understanding How Users Tolerate Algorithmic Exploration

Any learning algorithm for recommendation faces a fundamental trade-off between exploiting partial knowledge of a user»s interests to maximize satisfaction in the short term and discovering additional user interests to maximize satisfaction in the long term. To enable discovery, a machine learning algorithm typically elicits feedback on items it is uncertain about, which is termed algorithmic exploration in machine learning. This exploration comes with a cost to the user, since the items an algorithm chooses for exploration frequently turn out to not match the user»s interests. In this paper, we study how users tolerate such exploration and how presentation strategies can mitigate the exploration cost. To this end, we conduct a behavioral study with over 600 people, where we vary how algorithmic exploration is mixed into the set of recommendations. We find that users respond non-linearly to the amount of exploration, where some exploration mixed into the set of recommendations has little effect on short-term satisfaction and behavior. For long-term satisfaction, the overall goal is to learn via exploration about the items presented. We therefore also analyze the quantity and quality of implicit feedback signals such as clicks and hovers, and how they vary with different amounts of mix-in exploration. Our findings provide insights into how to design presentation strategies for algorithmic exploration in interactive recommender systems, mitigating the short-term costs of algorithmic exploration while aiming to elicit informative feedback data for learning.

[1]  Harald Steck,et al.  Interactive Recommender Systems: Tutorial , 2015, RecSys.

[2]  Dietmar Jannach,et al.  Item Familiarity Effects in User-Centric Evaluations of Recommender Systems , 2015, RecSys Posters.

[3]  Mark Claypool,et al.  Implicit interest indicators , 2001, IUI '01.

[4]  Jonathan L. Herlocker,et al.  Evaluating collaborative filtering recommender systems , 2004, TOIS.

[5]  F. Maxwell Harper,et al.  The MovieLens Datasets: History and Context , 2016, TIIS.

[6]  Thorsten Joachims,et al.  Counterfactual Risk Minimization: Learning from Logged Bandit Feedback , 2015, ICML.

[7]  John Riedl,et al.  The Tag Genome: Encoding Community Knowledge to Support Novel Interaction , 2012, TIIS.

[8]  Andrew Turpin,et al.  Why batch and user evaluations do not give the same results , 2001, SIGIR '01.

[9]  Dietmar Jannach,et al.  Adaptation and Evaluation of Recommendations for Short-term Shopping Goals , 2015, RecSys.

[10]  Oren Somekh,et al.  Budget-Constrained Item Cold-Start Handling in Collaborative Filtering Recommenders via Optimal Design , 2014, WWW.

[11]  Li Chen,et al.  A user-centric evaluation framework for recommender systems , 2011, RecSys '11.

[12]  Alexandros Karatzoglou,et al.  Parallel Recurrent Neural Network Architectures for Feature-rich Session-based Recommendations , 2016, RecSys.

[13]  Ryen W. White,et al.  No clicks, no problem: using cursor movements to understand and improve search , 2011, CHI.

[14]  Eugene Agichtein,et al.  Beyond dwell time: estimating document relevance from cursor movements and other post-click searcher behavior , 2012, WWW.

[15]  Filip Radlinski,et al.  Evaluating the accuracy of implicit feedback from clicks and query reformulations in Web search , 2007, TOIS.

[16]  Òscar Celma,et al.  A new approach to evaluating novel recommendations , 2008, RecSys '08.

[17]  John Riedl,et al.  Is seeing believing?: how recommender system interfaces affect users' opinions , 2003, CHI '03.

[18]  Filip Radlinski,et al.  How does clickthrough data reflect retrieval quality? , 2008, CIKM '08.

[19]  Alexander J. Smola,et al.  Using Navigation to Improve Recommendations in Real-Time , 2016, RecSys.

[20]  Wei Chu,et al.  Unbiased offline evaluation of contextual-bandit-based news article recommendation algorithms , 2010, WSDM '11.

[21]  Paul Resnick,et al.  Recommender systems , 1997, CACM.

[22]  Lior Rokach,et al.  Introduction to Recommender Systems Handbook , 2011, Recommender Systems Handbook.

[23]  Aniket Kittur,et al.  Crowdsourcing user studies with Mechanical Turk , 2008, CHI.

[24]  Martin Hepp,et al.  Effects of the Placement of Diverse Items in Recommendation Lists , 2012, ICEIS.

[25]  Joaquin Quiñonero Candela,et al.  Counterfactual reasoning and learning systems: the example of computational advertising , 2013, J. Mach. Learn. Res..

[26]  Kirsten Swearingen,et al.  Beyond Algorithms: An HCI Perspective on Recommender Systems , 2001 .

[27]  Tobias Höllerer,et al.  TasteWeights: a visual interactive hybrid recommender system , 2012, RecSys.

[28]  Franca Garzotto,et al.  Looking for "Good" Recommendations: A Comparative Evaluation of Recommender Systems , 2011, INTERACT.

[29]  Valerie J. Trifts,et al.  Consumer Decision Making in Online Shopping Environments: The Effects of Interactive Decision Aids , 2000 .

[30]  Sean M. McNee,et al.  Getting to know you: learning new user preferences in recommender systems , 2002, IUI '02.

[31]  F. Maxwell Harper,et al.  User perception of differences in recommender algorithms , 2014, RecSys '14.

[32]  Steve Fox,et al.  Evaluating implicit measures to improve web search , 2005, TOIS.

[33]  Franca Garzotto,et al.  User-Centric vs. System-Centric Evaluation of Recommender Systems , 2013, INTERACT.

[34]  Susan T. Dumais,et al.  Using Shortlists to Support Decision Making and Improve Recommender System Performance , 2015, WWW.

[35]  Bart P. Knijnenburg,et al.  Explaining the user experience of recommender systems , 2012, User Modeling and User-Adapted Interaction.

[36]  Sophie Ahrens,et al.  Recommender Systems , 2012 .

[37]  Oren Somekh,et al.  ExcUseMe: Asking Users to Help in Item Cold-Start Recommendations , 2015, RecSys.

[38]  Shai Shalev-Shwartz,et al.  Online Learning and Online Convex Optimization , 2012, Found. Trends Mach. Learn..

[39]  Sean M. McNee,et al.  Improving recommendation lists through topic diversification , 2005, WWW '05.

[40]  Bracha Shapira,et al.  Recommender Systems Handbook , 2015, Springer US.

[41]  Pearl Pu,et al.  User Technology Adoption Issues in Recommender Systems , 2007 .

[42]  Siddharth Suri,et al.  Conducting behavioral research on Amazon’s Mechanical Turk , 2010, Behavior research methods.

[43]  Rashmi R. Sinha,et al.  The role of transparency in recommender systems , 2002, CHI Extended Abstracts.

[44]  Katrien Verbert,et al.  Interactive recommender systems: A survey of the state of the art and future research challenges and opportunities , 2016, Expert Syst. Appl..

[45]  Leslie Pack Kaelbling,et al.  Algorithms for multi-armed bandit problems , 2014, ArXiv.

[46]  Yiming Yang,et al.  Personalized active learning for collaborative filtering , 2008, SIGIR '08.