The Diverse Cohort Selection Problem: Multi-Armed Bandits with Varied Pulls

How should a firm allocate its limited interviewing resources to select the optimal cohort of new employees from a large set of job applicants? How should that firm allocate cheap but noisy résumé screenings and expensive but in-depth in-person interviews? We view this problem through the lens of combinatorial pure exploration (CPE) in the multi-armed bandit setting, where a central learning agent performs costly exploration of a set of arms before selecting a final subset with some combinatorial structure. We generalize a recent CPE algorithm to the setting where arm pulls can have different cost, but return different levels of information, and prove theoretical upper bounds for a general class of arm-pulling strategies in this new setting. We then apply our general algorithm to a real-world problem with combinatorial structure: incorporating diversity into university admissions. We take real data from admissions at one of the largest US-based computer science graduate programs and show that a simulation of our algorithm produced more diverse student cohorts at low cost to individual student quality, and does so by spending comparable budget to the current admissions process at that university.

[1]  Wei Chen,et al.  Combinatorial Pure Exploration of Multi-Armed Bandits , 2014, NIPS.

[2]  Klaus Rifbjerg,et al.  Sex Roles , 2011 .

[3]  Mark Fuge,et al.  Diverse Weighted Bipartite b-Matching , 2017, IJCAI.

[4]  Xiaoyan Zhu,et al.  Promoting Diversity in Recommendation by Entropy Regularizer , 2013, IJCAI.

[5]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[6]  Yair Zick,et al.  Diversity Constraints in Public Housing Allocation , 2017, AAMAS.

[7]  Robert D. Nowak,et al.  Top Arm Identification in Multi-Armed Bandits with Batch Arm Pulls , 2016, AISTATS.

[8]  Hui Lin,et al.  A Class of Submodular Functions for Document Summarization , 2011, ACL.

[9]  Andreas Krause,et al.  Noisy Submodular Maximization via Adaptive Sampling with Applications to Crowdsourced Image Collection Summarization , 2015, AAAI.

[10]  Dan Roth,et al.  Will I Get in? Modeling the Graduate Admission Process for American Universities , 2016, 2016 IEEE 16th International Conference on Data Mining Workshops (ICDMW).

[11]  Pamela S Douglas,et al.  Diversity Matters. , 2017, Journal of the American College of Cardiology.

[12]  M. L. Fisher,et al.  An analysis of approximations for maximizing submodular set functions—I , 1978, Math. Program..

[13]  Nasir D. Memon,et al.  A robust model for paper reviewer assignment , 2014, RecSys '14.

[14]  Sébastien Bubeck,et al.  Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems , 2012, Found. Trends Mach. Learn..

[15]  Junyu Niu,et al.  A Framework for Recommending Relevant and Diverse Items , 2016, IJCAI.

[16]  Aaron Roth,et al.  Fairness in Learning: Classic and Contextual Bandits , 2016, NIPS.

[17]  Michel W. Pharand Human Creativity , 2018 .

[18]  Steven Bird,et al.  NLTK: The Natural Language Toolkit , 2002, ACL.

[19]  Andreas Krause,et al.  Learning to Hire Teams , 2015, HCOMP.

[20]  Filip Radlinski,et al.  Learning diverse rankings with multi-armed bandits , 2008, ICML '08.

[21]  T. Schmader,et al.  A Linguistic Comparison of Letters of Recommendation for Male and Female Chemistry and Biochemistry Job Applicants , 2007, Sex roles.

[22]  Wei Cao,et al.  On Top-k Selection in Multi-Armed Bandits and Hidden Bipartite Graphs , 2015, NIPS.

[23]  P. Desrochers,et al.  Local Diversity, Human Creativity, and Technological Innovation , 2001 .

[24]  Michael I. Jordan,et al.  Latent Dirichlet Allocation , 2001, J. Mach. Learn. Res..

[25]  Sujit Gujar,et al.  A quality assuring multi-armed bandit crowdsourcing mechanism with incentive compatible learning , 2014, AAMAS.

[26]  Nenghai Yu,et al.  Budgeted Multi-Armed Bandits with Multiple Plays , 2016, IJCAI.

[27]  Anil M. Shende,et al.  Applications of supervised learning techniques on undergraduate admissions data , 2016, Conf. Computing Frontiers.

[28]  Tao Qin,et al.  Multi-Armed Bandit with Budget Constraint and Variable Costs , 2013, AAAI.

[29]  Risto Miikkulainen,et al.  GRADE: Machine Learning Support for Graduate Admissions , 2013, AI Mag..

[30]  P. Cochat,et al.  Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.

[31]  Alexandra Carpentier,et al.  An optimal algorithm for the Thresholding Bandit Problem , 2016, ICML.