Anchored Regression: Solving Random Convex Equations via Convex Programming

We consider the problem of estimating a solution to (random) systems of equations that involve convex nonlinearities which has applications in machine learning and signal processing. Conventional estimators based on empirical risk minimization generally lead to non-convex programs that are often computationally intractable. We propose anchored regression, a new approach that utilizes an anchor vector to formulate an estimator based on a simple convex program. We analyze accuracy of this method and specify the required sample complexity. The proposed convex program is formulated in the natural space of the problem rather than a lifted domain, which makes it computationally favorable. This feature of anchored regression also provides great flexibility as structural priors (e.g., sparsity) can be seamlessly incorporated through convex regularization. We also provide recipes for constructing the anchor vector from the data.

[1]  Joel A. Tropp,et al.  Convex recovery of a structured signal from independent random linear measurements , 2014, ArXiv.

[2]  M. Talagrand,et al.  Probability in Banach Spaces: Isoperimetry and Processes , 1991 .

[3]  Justin Romberg,et al.  Phase Retrieval Meets Statistical Learning Theory: A Flexible Convex Relaxation , 2016, AISTATS.

[4]  Vladislav Voroninski,et al.  An Elementary Proof of Convex Phase Retrieval in the Natural Parameter Space via the Linear Program PhaseMax , 2016, ArXiv.

[5]  Simon Haykin,et al.  Neural Networks and Learning Machines , 2010 .

[6]  Arkadi Nemirovski,et al.  Topics in Non-Parametric Statistics , 2000 .

[7]  Vladislav Voroninski,et al.  Corruption Robust Phase Retrieval via Linear Programming , 2016, ArXiv.

[8]  P. Gänssler Weak Convergence and Empirical Processes - A. W. van der Vaart; J. A. Wellner. , 1997 .

[9]  Vladislav Voroninski,et al.  Compressed Sensing from Phaseless Gaussian Measurements via Linear Programming in the Natural Parameter Space , 2016, ArXiv.

[10]  S. Mendelson,et al.  Regularization and the small-ball method I: sparse recovery , 2016, 1601.05584.

[11]  Shahar Mendelson,et al.  Learning without Concentration , 2014, COLT.

[12]  Shahar Mendelson,et al.  Regularization and the small-ball method II: complexity dependent error rates , 2016, J. Mach. Learn. Res..

[13]  Yaniv Plan,et al.  The Generalized Lasso With Non-Linear Observations , 2015, IEEE Transactions on Information Theory.

[14]  Colin McDiarmid,et al.  Surveys in Combinatorics, 1989: On the method of bounded differences , 1989 .

[15]  V. Koltchinskii,et al.  Bounding the smallest singular value of a random matrix without concentration , 2013, 1312.3580.

[16]  H. Ichimura,et al.  SEMIPARAMETRIC LEAST SQUARES (SLS) AND WEIGHTED SLS ESTIMATION OF SINGLE-INDEX MODELS , 1993 .

[17]  Tengyao Wang,et al.  A useful variant of the Davis--Kahan theorem for statisticians , 2014, 1405.0680.

[18]  Tom Goldstein,et al.  PhaseMax: Convex Phase Retrieval via Basis Pursuit , 2016, IEEE Transactions on Information Theory.

[19]  Y. Plan,et al.  High-dimensional estimation with geometric constraints , 2014, 1404.3749.

[20]  Susan A. Murphy,et al.  Monographs on statistics and applied probability , 1990 .

[21]  R. Tibshirani,et al.  Sparse Principal Component Analysis , 2006 .

[22]  I. Johnstone,et al.  On Consistency and Sparsity for Principal Components Analysis in High Dimensions , 2009, Journal of the American Statistical Association.

[23]  P. McCullagh,et al.  Monograph on Statistics and Applied Probability , 1989 .

[24]  Shai Ben-David,et al.  Understanding Machine Learning: From Theory to Algorithms , 2014 .

[25]  S. Mendelson Learning without concentration for general loss functions , 2014, 1410.3192.