Learning-Based Low-Rank Approximations

We introduce a “learning-based” algorithm for the low-rank decomposition problem: given an $n \times d$ matrix $A$, and a parameter $k$, compute a rank-$k$ matrix $A'$ that minimizes the approximation loss $\|A-A'\|_F$. The algorithm uses a training set of input matrices in order to optimize its performance. Specifically, some of the most efficient approximate algorithms for computing low-rank approximations proceed by computing a projection $SA$, where $S$ is a sparse random $m \times n$ “sketching matrix”, and then performing the singular value decomposition of $SA$. We show how to replace the random matrix $S$ with a “learned” matrix of the same sparsity to reduce the error. Our experiments show that, for multiple types of data sets, a learned sketch matrix can substantially reduce the approximation loss compared to a random matrix $S$, sometimes up to one order of magnitude. We also study mixed matrices where only some of the rows are trained and the remaining ones are random, and show that matrices still offer improved performance while retaining worst-case guarantees. Finally, to understand the theoretical aspects of our approach, we study the special case of $m=1$. In particular, we give an approximation algorithm for minimizing the empirical loss, with approximation factor depending on the stable rank of matrices in the training set. We also show generalization bounds for the sketch matrix learning problem.

[1]  Alexandros G. Dimakis,et al.  Compressed Sensing using Generative Models , 2017, ICML.

[2]  Jakob Hoydis,et al.  Adaptive Neural Signal Detection for Massive MIMO , 2019, IEEE Transactions on Wireless Communications.

[3]  Richard G. Baraniuk,et al.  A deep learning approach to structured signal recovery , 2015, 2015 53rd Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[4]  Michael Mitzenmacher,et al.  A Model for Learned Bloom Filters and Optimizing by Sandwiching , 2018, NeurIPS.

[5]  Vladislav Voroninski,et al.  Global Guarantees for Enforcing Deep Generative Priors by Empirical Risk , 2017, IEEE Transactions on Information Theory.

[6]  Michael B. Cohen,et al.  Dimensionality Reduction for k-Means Clustering and Low Rank Approximation , 2014, STOC.

[7]  Wei Liu,et al.  Learning to Hash for Indexing Big Data—A Survey , 2015, Proceedings of the IEEE.

[8]  David P. Woodruff,et al.  Numerical linear algebra in the streaming model , 2009, STOC '09.

[9]  Edo Liberty,et al.  Simple and deterministic matrix sketching , 2012, KDD.

[10]  Christos Boutsidis,et al.  Improved Matrix Algorithms via the Subsampled Randomized Hadamard Transform , 2012, SIAM J. Matrix Anal. Appl..

[11]  Yuanzhi Li,et al.  Even Faster SVD Decomposition Yet Without Agonizing Pain , 2016, NIPS.

[12]  Tim Kraska,et al.  The Case for Learned Index Structures , 2018 .

[13]  V. Rokhlin,et al.  A randomized algorithm for the approximation of matrices , 2006 .

[14]  Sreenivas Gollapudi,et al.  Online Algorithms for Rent-Or-Buy with Expert Advice , 2019, ICML.

[15]  Piotr Indyk,et al.  Approximate Nearest Neighbor: Towards Removing the Curse of Dimensionality , 2012, Theory Comput..

[16]  Maria-Florina Balcan,et al.  Learning to Branch , 2018, ICML.

[17]  David P. Woodru Sketching as a Tool for Numerical Linear Algebra , 2014 .

[18]  David L Donoho,et al.  Compressed sensing , 2006, IEEE Transactions on Information Theory.

[19]  Sergei Vassilvitskii,et al.  Competitive caching with machine learned advice , 2018, ICML.

[20]  Google,et al.  Improving Online Algorithms via ML Predictions , 2024, NeurIPS.

[21]  Huy L. Nguyen,et al.  OSNAP: Faster Numerical Linear Algebra Algorithms via Sparser Subspace Embeddings , 2012, 2013 IEEE 54th Annual Symposium on Foundations of Computer Science.

[22]  Tamás Sarlós,et al.  Improved Approximation Algorithms for Large Matrices via Random Projections , 2006, 2006 47th Annual IEEE Symposium on Foundations of Computer Science (FOCS'06).

[23]  Nathan Halko,et al.  Finding Structure with Randomness: Probabilistic Algorithms for Constructing Approximate Matrix Decompositions , 2009, SIAM Rev..

[24]  V. Rokhlin,et al.  A fast randomized algorithm for the approximation of matrices ✩ , 2007 .

[25]  Yu Oishi,et al.  Hyperspectral Image Dataset for Benchmarking on Salient Object Detection , 2018, 2018 Tenth International Conference on Quality of Multimedia Experience (QoMEX).

[26]  Le Song,et al.  2 Common Formulation for Greedy Algorithms on Graphs , 2018 .

[27]  David P. Woodruff,et al.  Frequent Directions: Simple and Deterministic Matrix Sketching , 2015, SIAM J. Comput..

[28]  David P. Woodruff Sketching as a Tool for Numerical Linear Algebra , 2014, Found. Trends Theor. Comput. Sci..

[29]  Richard G. Baraniuk,et al.  Learned D-AMP: Principled Neural Network based Compressive Image Recovery , 2017, NIPS.

[30]  Jeff M. Phillips,et al.  Relative Errors for Deterministic Low-Rank Matrix Approximations , 2013, SODA.

[31]  Volkan Cevher,et al.  Learning-Based Compressive Subsampling , 2015, IEEE Journal of Selected Topics in Signal Processing.

[32]  David P. Woodruff,et al.  Low rank approximation and regression in input sparsity time , 2013, STOC '13.

[33]  Emmanuel J. Candès,et al.  Robust uncertainty principles: exact signal reconstruction from highly incomplete frequency information , 2004, IEEE Transactions on Information Theory.

[34]  Evgeniy Gabrilovich,et al.  Parameterized generation of labeled datasets for text categorization based on a hierarchical directory , 2004, SIGIR '04.

[35]  Michael W. Mahoney,et al.  Low-distortion subspace embeddings in input-sparsity time and applications to robust linear regression , 2012, STOC '13.

[36]  Piotr Indyk,et al.  Learning-Based Frequency Estimation Algorithms , 2018, ICLR.

[37]  Shai Ben-David,et al.  Understanding Machine Learning: From Theory to Algorithms , 2014 .