Batch Active Learning Using Determinantal Point Processes

Data collection and labeling is one of the main challenges in employing machine learning algorithms in a variety of real-world applications with limited data. While active learning methods attempt to tackle this issue by labeling only the data samples that give high information, they generally suffer from large computational costs and are impractical in settings where data can be collected in parallel. Batch active learning methods attempt to overcome this computational burden by querying batches of samples at a time. To avoid redundancy between samples, previous works rely on some ad hoc combination of sample quality and diversity. In this paper, we present a new principled batch active learning method using Determinantal Point Processes, a repulsive point process that enables generating diverse batches of samples. We develop tractable algorithms to approximate the mode of a DPP distribution, and provide theoretical guarantees on the degree of approximation. We further demonstrate that an iterative greedy method for DPP maximization, which has lower computational costs but worse theoretical guarantees, still gives competitive results for batch active learning. Our experiments show the value of our methods on several datasets against state-of-the-art baselines.

[1]  Suvrit Sra,et al.  Fast Mixing Markov Chains for Strongly Rayleigh Measures, DPPs, and Constrained Sampling , 2016, NIPS.

[2]  Robert D. Nowak,et al.  Active Ranking using Pairwise Comparisons , 2011, NIPS.

[3]  Sanjoy Dasgupta,et al.  Hierarchical sampling for active learning , 2008, ICML '08.

[4]  Pengtao Xie,et al.  Deep Determinantal Point Process for Large-Scale Multi-label Classification , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[5]  Nan Ye,et al.  Active Learning for Probabilistic Hypotheses Using the Maximum Gibbs Error Criterion , 2013, NIPS.

[6]  Ulrike von Luxburg,et al.  Feasibility of Active Machine Learning for Multiclass Compound Classification , 2016, J. Chem. Inf. Model..

[7]  Charles Blundell,et al.  Simple and Scalable Predictive Uncertainty Estimation using Deep Ensembles , 2016, NIPS.

[8]  Ben Taskar,et al.  k-DPPs: Fixed-Size Determinantal Point Processes , 2011, ICML.

[9]  I-Cheng Yeh,et al.  Knowledge discovery on RFM model using Bernoulli sequence , 2009, Expert Syst. Appl..

[10]  Jerry Alan Fails,et al.  Interactive machine learning , 2003, IUI '03.

[11]  Malik Magdon-Ismail,et al.  On selecting a maximum volume sub-matrix of a matrix and related problems , 2009, Theor. Comput. Sci..

[12]  Hedvig Kjellstrom,et al.  Determinantal Point Processes for Mini-Batch Diversification , 2017, UAI 2017.

[13]  Suvrit Sra,et al.  Exponentiated Strongly Rayleigh Distributions , 2018, NeurIPS.

[14]  Thorsten Joachims,et al.  Learning preferences for manipulation tasks from online coactive feedback , 2015, Int. J. Robotics Res..

[15]  Ben Taskar,et al.  Determinantal Point Processes for Machine Learning , 2012, Found. Trends Mach. Learn..

[16]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[17]  Luís Torgo,et al.  OpenML: networked science in machine learning , 2014, SKDD.

[18]  Anil K. Jain,et al.  Statistical Pattern Recognition: A Review , 2000, IEEE Trans. Pattern Anal. Mach. Intell..

[19]  Thorsten Joachims,et al.  Coactive Learning , 2015, J. Artif. Intell. Res..

[20]  Zoubin Ghahramani,et al.  Bayesian Active Learning for Classification and Preference Learning , 2011, ArXiv.

[21]  Justin Salez,et al.  Modified log-Sobolev inequalities for strong-Rayleigh measures , 2019, The Annals of Applied Probability.

[22]  Dale Schuurmans,et al.  Discriminative Batch Mode Active Learning , 2007, NIPS.

[23]  Xiaowei Xu,et al.  Representative Sampling for Text Classification Using Support Vector Machines , 2003, ECIR.

[24]  Michèle Sebag,et al.  APRIL: Active Preference-learning based Reinforcement Learning , 2012, ECML/PKDD.

[25]  Dorsa Sadigh,et al.  Learning Reward Functions by Integrating Human Demonstrations and Preferences , 2019, Robotics: Science and Systems.

[26]  Anca D. Dragan,et al.  Active Preference-Based Learning of Reward Functions , 2017, Robotics: Science and Systems.

[27]  Dorsa Sadigh,et al.  The Green Choice: Learning and Influencing Human Decisions on Shared Roads , 2019, 2019 IEEE 58th Conference on Decision and Control (CDC).

[28]  Pushmeet Kohli,et al.  Batched Gaussian Process Bandit Optimization via Determinantal Point Processes , 2016, NIPS.

[29]  Aleksandar Nikolov Randomized Rounding for the Largest Simplex Problem , 2015, STOC.

[30]  Lorenzo Bruzzone,et al.  A Batch-Mode Active Learning Technique Based on Multiple Uncertainty for SVM Classifier , 2012, IEEE Geoscience and Remote Sensing Letters.

[31]  Glencora Borradaile,et al.  Batch Active Learning via Coordinated Matching , 2012, ICML.

[32]  Cheng Zhang,et al.  Active Mini-Batch Sampling using Repulsive Point Processes , 2018, AAAI.

[33]  Nima Anari,et al.  Monte Carlo Markov Chain Algorithms for Sampling Strongly Rayleigh Distributions and Determinantal Point Processes , 2016, COLT.

[34]  Dorsa Sadigh,et al.  Batch Active Preference-Based Learning of Reward Functions , 2018, CoRL.

[35]  Silvio Savarese,et al.  Active Learning for Convolutional Neural Networks: A Core-Set Approach , 2017, ICLR.

[36]  Fedor Zhdanov,et al.  Diverse mini-batch Active Learning , 2019, ArXiv.

[37]  Percy Liang,et al.  Uncertainty Sampling is Preconditioned Stochastic Gradient Descent on Zero-One Loss , 2018, NeurIPS.

[38]  John Langford,et al.  Importance weighted active learning , 2008, ICML '09.

[39]  Hsuan-Tien Lin,et al.  Active Learning by Learning , 2015, AAAI.

[40]  Thomas Hofmann,et al.  Active learning for misspecified generalized linear models , 2007 .

[41]  Ben Taskar,et al.  Discovering Diverse and Salient Threads in Document Collections , 2012, EMNLP.

[42]  Daniel King,et al.  Fetch & Freight : Standard Platforms for Service Robot Applications , 2016 .

[43]  Dong Yu,et al.  Maximizing global entropy reduction for active learning in speech recognition , 2009, 2009 IEEE International Conference on Acoustics, Speech and Signal Processing.

[44]  Robert Tibshirani,et al.  A Comparison of Some Error Estimates for Neural Network Models , 1996, Neural Computation.

[45]  Andreas Krause,et al.  Submodular Function Maximization , 2014, Tractability.

[46]  Anca D. Dragan,et al.  Planning for Autonomous Cars that Leverage Effects on Human Actions , 2016, Robotics: Science and Systems.

[47]  Maurice Queyranne,et al.  An Exact Algorithm for Maximum Entropy Sampling , 1995, Oper. Res..

[48]  Malik Magdon-Ismail,et al.  Exponential Inapproximability of Selecting a Maximum Volume Sub-matrix , 2011, Algorithmica.

[49]  Guilherme A. Barreto,et al.  Short-term memory mechanisms in neural network learning of robot navigation tasks: A case study , 2009, 2009 6th Latin American Robotics Symposium (LARS 2009).

[50]  Nima Anari,et al.  Log-concave polynomials II: high-dimensional walks and an FPRAS for counting bases of a matroid , 2018, STOC.

[51]  Finale Doshi-Velez,et al.  Decomposition of Uncertainty in Bayesian Deep Learning for Efficient and Risk-sensitive Learning , 2017, ICML.

[52]  Leslie Pack Kaelbling,et al.  Active Model Learning and Diverse Action Sampling for Task and Motion Planning , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[53]  Maria-Florina Balcan,et al.  Margin Based Active Learning , 2007, COLT.

[54]  Ben Taskar,et al.  Structured Determinantal Point Processes , 2010, NIPS.

[55]  Nikolaos Papanikolopoulos,et al.  Multi-class active learning for image classification , 2009, CVPR.

[56]  Yuval Tassa,et al.  MuJoCo: A physics engine for model-based control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[57]  Paul N. Bennett,et al.  Active Comparison Based Learning Incorporating User Uncertainty and Noise , 2016 .

[58]  Shane Legg,et al.  Deep Reinforcement Learning from Human Preferences , 2017, NIPS.

[59]  Roland Vollgraf,et al.  Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.

[60]  Burr Settles,et al.  Active Learning Literature Survey , 2009 .

[61]  W. B. Johnson,et al.  Extensions of Lipschitz mappings into Hilbert space , 1984 .

[62]  Yi Yang,et al.  Multi-Class Active Learning by Uncertainty Sampling with Diversity Maximization , 2015, International Journal of Computer Vision.

[63]  Andreas Krause,et al.  Near-optimal Batch Mode Active Learning and Adaptive Submodular Optimization , 2013, ICML.

[64]  E. Rains,et al.  Eynard–Mehta Theorem, Schur Process, and their Pfaffian Analogs , 2004, math-ph/0409059.

[65]  Francis R. Bach,et al.  Learning Determinantal Point Processes in Sublinear Time , 2016, AISTATS.

[66]  Bernt Schiele,et al.  RALF: A reinforced active learning formulation for object class recognition , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[67]  Hugo Larochelle,et al.  Meta-Learning for Batch Mode Active Learning , 2018, ICLR.

[68]  Marcos André Gonçalves,et al.  Ranked batch-mode active learning , 2017, Inf. Sci..