Fast Parallel Algorithms for Feature Selection

In this paper, we propose a new framework for designing fast parallel algorithms for fundamental statistical subset selection tasks that include feature selection and experimental design. Such tasks are known to be weakly submodular and are amenable to optimization via the standard greedy algorithm. Despite its desirable approximation guarantees, however, the greedy algorithm is inherently sequential and in the worst case, its parallel runtime is linear in the size of the data. Recently, there has been a surge of interest in a parallel optimization technique called adaptive sampling which produces solutions with desirable approximation guarantees for submodular maximization in exponentially faster parallel runtime. Unfortunately, we show that for general weakly submodular functions such accelerations are impossible. The major contribution in this paper is a novel relaxation of submodularity which we call differential submodularity. We first prove that differential submodularity characterizes objectives like feature selection and experimental design. We then design an adaptive sampling algorithm for differentially submodular functions whose parallel runtime is logarithmic in the size of the data and achieves strong approximation guarantees. Through experiments, we show the algorithm's performance is competitive with state-of-the-art methods and obtains dramatic speedups for feature selection and experimental design problems.

[1]  Alexandros G. Dimakis,et al.  Streaming Weak Submodularity: Interpreting Neural Networks on the Fly , 2017, NIPS.

[2]  Eric Balkanski,et al.  An Exponential Speedup in Parallel Running Time for Submodular Maximization without Loss in Approximation , 2018, SODA.

[3]  Yaron Singer,et al.  Maximization of Approximately Submodular Functions , 2016, NIPS.

[4]  Morteza Zadimoghaddam,et al.  Submodular Maximization with Optimal Approximation, Adaptivity and Query Complexity , 2018, ArXiv.

[5]  Abhimanyu Das,et al.  Submodular meets Spectral: Greedy Algorithms for Subset Selection, Sparse Approximation and Dictionary Selection , 2011, ICML.

[6]  Amin Karbasi,et al.  Unconstrained submodular maximization with constant adaptive complexity , 2019, STOC.

[7]  Andreas Krause,et al.  Submodular Dictionary Selection for Sparse Representation , 2010, ICML.

[8]  Huy L. Nguyen,et al.  Submodular maximization with matroid and packing constraints in parallel , 2018, STOC.

[9]  Alexandros G. Dimakis,et al.  Restricted Strong Convexity Implies Weak Submodularity , 2016, The Annals of Statistics.

[10]  Julien Mairal,et al.  Complexity Analysis of the Lasso Regularization Path , 2012, ICML.

[11]  Eric Balkanski,et al.  The adaptive complexity of maximizing a submodular function , 2018, STOC.

[12]  Yuanzhi Li,et al.  The Well Tempered Lasso , 2018, ICML.

[13]  Morteza Zadimoghaddam,et al.  Non-monotone Submodular Maximization with Nearly Optimal Adaptivity Complexity , 2018, ArXiv.

[14]  Andreas Krause,et al.  Guarantees for Greedy Maximization of Non-submodular Functions with Applications , 2017, ICML.

[15]  Kent Quanrud,et al.  Submodular Function Maximization in Parallel via the Multilinear Relaxation , 2019, SODA.

[16]  Eric Balkanski,et al.  An optimal approximation for submodular maximization under a matroid constraint in the adaptive complexity model , 2018, STOC.

[17]  Eric Balkanski,et al.  Approximation Guarantees for Adaptive Sampling , 2018, ICML.

[18]  Huy L. Nguyen,et al.  A Parallel Double Greedy Algorithm for Submodular Maximization , 2018, ArXiv.

[19]  Eric Balkanski,et al.  Non-monotone Submodular Maximization in Exponentially Fewer Iterations , 2018, NeurIPS.

[20]  N. L. Johnson,et al.  Multivariate Analysis , 1958, Nature.

[21]  Kent Quanrud,et al.  Parallelizing greedy for submodular set function maximization in matroids and beyond , 2018, STOC.

[22]  Abhimanyu Das,et al.  Selecting Diverse Features via Spectral Regularization , 2012, NIPS.

[23]  Andreas Krause,et al.  Near-Optimal Sensor Placements in Gaussian Processes: Theory, Efficient Algorithms and Empirical Studies , 2008, J. Mach. Learn. Res..

[24]  Morteza Zadimoghaddam,et al.  Non-monotone Submodular Maximization with Nearly Optimal Adaptivity and Query Complexity , 2018, ICML.

[25]  Huy L. Nguyen,et al.  Submodular Maximization with Packing Constraints in Parallel , 2018, ArXiv.

[26]  Huy L. Nguyen,et al.  Submodular Maximization with Nearly-optimal Approximation and Adaptivity in Nearly-linear Time , 2018, SODA.