How to Find the Best Rated Items on a Likert Scale and How Many Ratings Are Enough

The collection and exploitation of ratings from users are modern pillars of collaborative filtering. Likert scale is a psychometric quantifier of ratings popular among the electronic commerce sites. In this paper, we consider the tasks of collecting Likert scale ratings of items and of finding the n-k best-rated items, i.e., the n items that are most likely to be the top-k in a ranking constructed from these ratings. We devise an algorithm, Pundit, that computes the n-k best-rated items. Pundit uses the probability-generating function constructed from the Likert scale responses to avoid the combinatorial exploration of the possible outcomes and to compute the result efficiently. Selection of the best-rated items meets, in practice, the major obstacle of the scarcity of ratings. We propose an approach that learns from the available data how many ratings are enough to meet a prescribed error. We empirically validate with real datasets the effectiveness of our method to recommend the collection of additional ratings.

[1]  Ronald Fagin,et al.  Comparing top k lists , 2003, SODA '03.

[2]  A. W. Kemp,et al.  Univariate Discrete Distributions , 1993 .

[3]  Luo Si,et al.  Collaborative filtering with decoupled models for preferences and ratings , 2003, CIKM '03.

[4]  Jongwuk Lee,et al.  Improving the accuracy of top-N recommendation using a preference model , 2016, Inf. Sci..

[5]  A. Kiureghian,et al.  Aleatory or epistemic? Does it matter? , 2009 .

[6]  Ihab F. Ilyas,et al.  Supporting ranking queries on uncertain and incomplete data , 2010, The VLDB Journal.

[7]  Guoliang Li,et al.  Crowdsourced Top-k Algorithms: An Experimental Evaluation , 2016, Proc. VLDB Endow..

[8]  R. A. Leibler,et al.  On Information and Sufficiency , 1951 .

[9]  Paul N. Bennett,et al.  Pairwise ranking aggregation in a crowdsourced setting , 2013, WSDM.

[10]  Lei Yu,et al.  Listwise Approach for Rank Aggregation in Crowdsourcing , 2015, WSDM.

[11]  Jian Li,et al.  Ranking continuous probabilistic datasets , 2010, Proc. VLDB Endow..

[12]  R. Likert “Technique for the Measurement of Attitudes, A” , 2022, The SAGE Encyclopedia of Research Design.

[13]  Stéphane Bressan,et al.  Top-k Queries Over Uncertain Scores , 2016, OTM Conferences.

[14]  Aditya G. Parameswaran,et al.  So who won?: dynamic max discovery with the crowd , 2012, SIGMOD Conference.

[15]  Neoklis Polyzotis,et al.  Max algorithms in crowdsourcing environments , 2012, WWW.

[16]  S. Jamieson Likert scales: how to (ab)use them , 2004, Medical education.

[17]  John Riedl,et al.  Collaborative Filtering Recommender Systems , 2011, Found. Trends Hum. Comput. Interact..

[18]  D. Doermann,et al.  Combining preference and absolute judgements in a crowd-sourced setting , 2013 .

[19]  Anton van den Hengel,et al.  Image-Based Recommendations on Styles and Substitutes , 2015, SIGIR.

[20]  Christian P. Robert,et al.  Monte Carlo Statistical Methods , 2005, Springer Texts in Statistics.

[21]  Jian Pei,et al.  Ranking queries on uncertain data: a probabilistic threshold approach , 2008, SIGMOD Conference.

[22]  J. Tukey,et al.  An algorithm for the machine calculation of complex Fourier series , 1965 .

[23]  Jure Leskovec,et al.  Inferring Networks of Substitutable and Complementary Products , 2015, KDD.