Selective multiple power iteration: from tensor PCA to gradient-based exploration of landscapes

We propose Selective Multiple Power Iterations (SMPI), a new algorithm to address the important Tensor PCA problem that consists in recovering a spike v⊗k 0 corrupted by a Gaussian noise tensor Z ∈ (Rn)⊗k such that T = √ nβv⊗k 0 +Z where β is the signal-to-noise ratio. SMPI consists in generating a polynomial number of random initializations, performing a polynomial number of symmetrized tensor power iterations on each initialization, then selecting the one that maximizes 〈T,v〉. Various numerical simulations for k = 3 in the conventionally considered range n ≤ 1000, where existent algorithms exhibit negligible finite size effects, show that the experimental performances of SMPI improve drastically upon existent algorithms and becomes comparable to the theoretical optimal recovery. We show that these unexpected performances are due to a powerful mechanism in which the noise plays a key role for the signal recovery and that takes place at low β. Furthermore, this mechanism results from five essential features of SMPI that distinguish it from previous algorithms based on power iteration. These remarkable results may have strong impact on both practical and theoretical applications of Tensor PCA. (i) We provide in the supplementary material multiple variants of this algorithm to tackle low-rank CP tensor decomposition. These proposed algorithms also outperforms existent methods even on real data which shows a huge potential impact for practical applications. (ii) We present new theoretical insights on the behavior of SMPI and gradient descent methods for the optimization in high-dimensional non-convex landscapes that are present in various machine learning problems. (iii) We expect that these results may help the discussion concerning the existence of the conjectured statistical-algorithmic gap.

[1]  Hossein Mobahi,et al.  Homotopy Analysis for Tensor PCA , 2016, COLT.

[3]  Nasser M. Nasrabadi,et al.  Hyperspectral Target Detection : An Overview of Current and Future Challenges , 2014, IEEE Signal Processing Magazine.

[4]  Maja Pantic,et al.  TensorLy: Tensor Learning in Python , 2016, J. Mach. Learn. Res..

[5]  Afonso S. Bandeira,et al.  Average-Case Integrality Gap for Non-Negative Principal Component Analysis , 2020, MSML.

[6]  David R. Thompson,et al.  ABoVE: Hyperspectral Imagery from AVIRIS-NG, Alaskan and Canadian Arctic, 2017-2019 , 2018 .

[7]  Caroline Fossati,et al.  Denoising of Hyperspectral Images Using the PARAFAC Model and Statistical Performance Analysis , 2012, IEEE Transactions on Geoscience and Remote Sensing.

[8]  Seung-Ik Lee,et al.  CP-decomposition with Tensor Power Method for Convolutional Neural Networks compression , 2017, 2017 IEEE International Conference on Big Data and Smart Computing (BigComp).

[9]  Alexis Decurninge,et al.  Tensor-Based Modulation for Unsourced Massive Random Access , 2021, IEEE Wireless Communications Letters.

[10]  Afonso S. Bandeira,et al.  Notes on Computational Hardness of Hypothesis Testing: Predictions using the Low-Degree Likelihood Ratio , 2019, ArXiv.

[11]  Florent Krzakala,et al.  Statistical and computational phase transitions in spiked tensor estimation , 2017, 2017 IEEE International Symposium on Information Theory (ISIT).

[12]  Florent Krzakala,et al.  Who is Afraid of Big Bad Minima? Analysis of Gradient-Flow in a Spiked Matrix-Tensor Model , 2019, NeurIPS.

[13]  M. B. Hastings,et al.  Classical and Quantum Algorithms for Tensor Principal Component Analysis , 2019, Quantum.

[14]  G. B. Arous,et al.  The Landscape of the Spiked Tensor Model , 2017, Communications on Pure and Applied Mathematics.

[15]  Anima Anandkumar,et al.  Learning Overcomplete Latent Variable Models through Tensor Methods , 2014, COLT.

[16]  Cristopher Moore,et al.  The Kikuchi Hierarchy and Tensor PCA , 2019, 2019 IEEE 60th Annual Symposium on Foundations of Computer Science (FOCS).

[17]  Mohamed Tamaazousti,et al.  Field Theoretical Approach for Signal Detection in Nearly Continuous Positive Spectra II: Tensorial Data , 2020, Entropy.

[18]  Daniel Z. Huang,et al.  Power Iteration for Tensor PCA , 2020, J. Mach. Learn. Res..

[19]  Tselil Schramm,et al.  Fast spectral algorithms from sum-of-squares proofs: tensor decomposition and planted sparse vectors , 2015, STOC.

[20]  Guy Bresler,et al.  Reducibility and Statistical-Computational Gaps from Secret Leakage , 2020, COLT.

[21]  Richard D. Braatz,et al.  Opportunities in tensorial data analytics for chemical and biological manufacturing processes , 2020, Comput. Chem. Eng..

[22]  G. B. Arous,et al.  Algorithmic thresholds for tensor PCA , 2018, The Annals of Probability.

[23]  Anima Anandkumar,et al.  Tensor decompositions for learning latent variable models , 2012, J. Mach. Learn. Res..

[24]  Florent Krzakala,et al.  Passed & Spurious: Descent Algorithms and Local Minima in Spiked Matrix-Tensor Models , 2019, ICML.

[25]  Anru R. Zhang,et al.  Tensor SVD: Statistical and Computational Limits , 2017, IEEE Transactions on Information Theory.

[26]  G. Biroli,et al.  Complex Energy Landscapes in Spiked-Tensor and Simple Glassy Models: Ruggedness, Arrangements of Local Minima, and Phase Transitions , 2018, Physical Review X.

[27]  Yann LeCun,et al.  The Loss Surfaces of Multilayer Networks , 2014, AISTATS.

[28]  Afonso S. Bandeira,et al.  Statistical limits of spiked tensor models , 2016, Annales de l'Institut Henri Poincaré, Probabilités et Statistiques.

[29]  Florent Krzakala,et al.  Marvels and Pitfalls of the Langevin Algorithm in Noisy High-dimensional Inference , 2018, Physical Review X.

[30]  Anima Anandkumar,et al.  Online and Differentially-Private Tensor Decomposition , 2016, NIPS.

[31]  Federico Ricci-Tersenghi,et al.  How to iron out rough landscapes and get optimal performances: averaged gradient descent and its application to tensor PCA , 2019, Journal of Physics A: Mathematical and Theoretical.

[32]  Andrea Montanari,et al.  Non-Negative Principal Component Analysis: Message Passing Algorithms and Sharp Asymptotics , 2014, IEEE Transactions on Information Theory.

[33]  Yuetian Luo,et al.  Open Problem: Average-Case Hardness of Hypergraphic Planted Clique Detection , 2020, COLT 2020.

[34]  Tamara G. Kolda,et al.  Tensor Decompositions and Applications , 2009, SIAM Rev..

[35]  Anima Anandkumar,et al.  A Tensor Spectral Approach to Learning Mixed Membership Community Models , 2013, COLT.

[36]  Andrea Montanari,et al.  A statistical model for tensor PCA , 2014, NIPS.

[37]  Weihong Guo,et al.  CPAC-Conv: CP-decomposition to Approximately Compress Convolutional Layers in Deep Learning , 2020, ArXiv.

[38]  Aukosh Jagannath,et al.  Statistical thresholds for tensor PCA , 2018, The Annals of Applied Probability.