Introduction to the Special Issue on Recommender System Benchmarking

Recommender systems addvalue to vast content resources by matching users with items of interest. In recent years, immense progress has been made in recommendation techniques. The evaluation of these systems is still based on traditional information retrieval and statistics metrics (e.g., precision, recall, RMSE), often not taking the use case and situation of the system into consideration. However, the rapid evolution of recommender systems in both their goals and their application domains fosters the need for new evaluation methodologies and environments. This special issue serves as a venue for work on novel, recommendation-centric benchmarking approaches taking the users’ utility, the business values, and the technical constraints into consideration. Building on the success of the Recommendation Utility Evaluation Workshop [Amatriain et al. 2012] held at Recsys 2012, the Workshop on Benchmarking Adaptive Retrieval and Recommender Systems [Castells et al. 2013a, 2013b] held at SIGIR 2013, the Workshop on Reproducibility and Replication in Recommender System Evaluation [Bellogı́n et al. 2013, 2014], the various Recommender System Challenges [Adomavicius et al. 2010; Said et al. 2011; Manouselis et al. 2012; Blomo et al. 2013; Said et al. 2014; Ben-Shimon et al. 2015], and other similar events, this special issue collects articles focusing on various aspects of challenges in benchmarking and evaluation or recommender systems. In the decade since the inception of the ACM RecSys conference and the decades since the first papers on this topic started to appear [Resnick et al. 1994], the field has

[1]  Toon De Pessemier,et al.  A Framework for Dataset Benchmarking and Its Application to a New Movie Rating Dataset , 2016, ACM Trans. Intell. Syst. Technol..

[2]  Martin Ester,et al.  RecSys challenge 2013 , 2013, RecSys.

[3]  Lior Rokach,et al.  RecSys Challenge 2015 and the YOOCHOOSE Dataset , 2015, RecSys.

[4]  Katrien Verbert,et al.  Recommender systems challenge 2012 , 2012, RecSys '12.

[5]  Frank Hopfgartner,et al.  Workshop on benchmarking adaptive retrieval and recommender systems: BARS 2013 , 2013, SIGIR.

[6]  Gediminas Adomavicius,et al.  Context-awareness in recommender systems: research workshop and movie recommendation challenge , 2010, RecSys '10.

[7]  Alan Said,et al.  Report on the workshop on reproducibility and replication in recommender systems evaluation (RepSys) , 2014, SIGF.

[8]  Hai Yang,et al.  ACM Transactions on Intelligent Systems and Technology - Special Section on Urban Computing , 2014 .

[9]  Alan Said,et al.  Challenge on context-aware movie recommendation: CAMRa2011 , 2011, RecSys '11.

[10]  John Riedl,et al.  GroupLens: an open architecture for collaborative filtering of netnews , 1994, CSCW '94.

[11]  Guy Shani,et al.  Anytime Algorithms for Recommendation Service Providers , 2016, ACM Trans. Intell. Syst. Technol..

[12]  Frank Hopfgartner,et al.  Report on the SIGIR 2013 workshop on benchmarking adaptive retrieval and recommender systems , 2013, SIGIR Forum.

[13]  Simon Dooms,et al.  Recommender systems challenge 2014 , 2014, RecSys '14.

[14]  David H. Glass,et al.  A Novel Classification Framework for Evaluating Individual and Aggregate Diversity in Top-N Recommendations , 2016, ACM Trans. Intell. Syst. Technol..

[15]  Gerd Stumme,et al.  The Role of Cores in Recommender Benchmarking for Social Bookmarking Systems , 2016, ACM Trans. Intell. Syst. Technol..