SILVar: Single Index Latent Variable Models

A semiparametric, nonlinear regression model in the presence of latent variables is introduced. These latent variables can correspond to unmodeled phenomena or unmeasured agents in a complex networked system. This new formulation allows joint estimation of certain nonlinearities in the system, the direct interactions between measured variables and the effects of unmodeled elements on the observed system. The particular form of the model adopted is justified, and learning is posed as a regularized empirical risk minimization. This leads to classes of structured convex optimization problems with a “sparse plus low-rank” flavor. Relations between the proposed model and several common model paradigms, such as those of robust principal component analysis and vector autoregression (VAR), are established. Particularly in the VAR setting, the low-rank contributions can come from broad trends exhibited in the time series. Details of the algorithm for learning the model are presented. Experiments demonstrate the performance of the model and the estimation algorithm on simulated and real data.

[1]  José M. F. Moura,et al.  Discrete Signal Processing on Graphs , 2012, IEEE Transactions on Signal Processing.

[2]  Yi Ma,et al.  Robust principal component analysis? , 2009, JACM.

[3]  Pablo A. Parrilo,et al.  Rank-Sparsity Incoherence for Matrix Decomposition , 2009, SIAM J. Optim..

[4]  José M. F. Moura,et al.  Signal Processing on Graphs: Causal Modeling of Unstructured Data , 2015, IEEE Transactions on Signal Processing.

[5]  Joydeep Ghosh,et al.  Parameter Estimation of Generalized Linear Models without Assuming their Link Function , 2015, AISTATS.

[6]  W. Wilbur,et al.  Isotonic Regression under Lipschitz Constraint , 2009, Journal of optimization theory and applications.

[7]  H. Ichimura,et al.  SEMIPARAMETRIC LEAST SQUARES (SLS) AND WEIGHTED SLS ESTIMATION OF SINGLE-INDEX MODELS , 1993 .

[8]  C. Granger Investigating causal relations by econometric models and cross-spectral methods , 1969 .

[9]  José M. F. Moura,et al.  Signal processing on graphs: Estimating the structure of a graph , 2015, 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[10]  Rebecca Willett,et al.  Matrix Completion Under Monotonic Single Index Models , 2015, NIPS.

[11]  Robert D. Nowak,et al.  Learning Single Index Models in High Dimensions , 2015, ArXiv.

[12]  G. Michailidis,et al.  Regularized estimation in sparse high-dimensional time series models , 2013, 1311.4175.

[13]  Richard L. Dykstra An isotonic regression algorithm , 1981 .

[14]  C. Granger Causality, cointegration, and control , 1988 .

[15]  Mark W. Schmidt,et al.  Optimizing Costly Functions with Simple Constraints: A Limited-Memory Projected Quasi-Newton Algorithm , 2009, AISTATS.

[16]  Inderjit S. Dhillon,et al.  Clustering with Bregman Divergences , 2005, J. Mach. Learn. Res..

[17]  Robert D. Nowak,et al.  Causal Network Inference Via Group Sparse Regularization , 2011, IEEE Transactions on Signal Processing.

[18]  Stephen J. Wright,et al.  Numerical Optimization , 2018, Fundamental Statistical Inference.

[19]  Pablo A. Parrilo,et al.  Latent variable graphical model selection via convex optimization , 2010, 2010 48th Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[20]  José M. F. Moura,et al.  Discrete Signal Processing on Graphs: Frequency Analysis , 2013, IEEE Transactions on Signal Processing.

[21]  R. Tibshirani,et al.  Sparse inverse covariance estimation with the graphical lasso. , 2008, Biostatistics.

[22]  Yves Lucet,et al.  Faster than the Fast Legendre Transform, the Linear-time Legendre Transform , 1997, Numerical Algorithms.

[23]  Martin J. Wainwright,et al.  A unified framework for high-dimensional analysis of $M$-estimators with decomposable regularizers , 2009, NIPS.

[24]  L. Bregman The relaxation method of finding the common point of convex sets and its application to the solution of problems in convex programming , 1967 .

[25]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[26]  Emmanuel J. Candès,et al.  Matrix Completion With Noise , 2009, Proceedings of the IEEE.

[27]  Eric P. Xing,et al.  Fast structure learning in generalized stochastic processes with latent factors , 2013, KDD.

[28]  Adam Tauman Kalai,et al.  Efficient Learning of Generalized Linear and Single Index Models with Isotonic Regression , 2011, NIPS.

[29]  John N. Tsitsiklis,et al.  Parallel and distributed computation , 1989 .

[30]  Stephen P. Boyd,et al.  Generalized Low Rank Models , 2014, Found. Trends Mach. Learn..

[31]  Emmanuel J. Candès,et al.  Adaptive Restart for Accelerated Gradient Schemes , 2012, Foundations of Computational Mathematics.

[32]  Ali Jalali,et al.  Learning the Dependence Graph of Time Series with Latent Factors , 2011, ICML.