Sparse Volterra and Polynomial Regression Models: Recoverability and Estimation

Volterra and polynomial regression models play a major role in nonlinear system identification and inference tasks. Exciting applications ranging from neuroscience to genome-wide association analysis build on these models with the additional requirement of parsimony. This requirement has high interpretative value, but unfortunately cannot be met by least-squares based or kernel regression methods. To this end, compressed sampling (CS) approaches, already successful in linear regression settings, can offer a viable alternative. The viability of CS for sparse Volterra and polynomial models is the core theme of this work. A common sparse regression task is initially posed for the two models. Building on (weighted) Lasso-based schemes, an adaptive RLS-type algorithm is developed for sparse polynomial regressions. The identifiability of polynomial models is critically challenged by dimensionality. However, following the CS principle, when these models are sparse, they could be recovered by far fewer measurements. To quantify the sufficient number of measurements for a given level of sparsity, restricted isometry properties (RIP) are investigated in commonly met polynomial regression settings, generalizing known results for their linear counterparts. The merits of the novel (weighted) adaptive CS algorithms to sparse polynomial modeling are verified through synthetic as well as real data tests for genotype-phenotype analysis.

[1]  J. Franklin,et al.  The elements of statistical learning: data mining, inference and prediction , 2005 .

[2]  Shizhong Xu,et al.  An Empirical Bayes Method for Estimating Epistatic Effects of Quantitative Trait Loci , 2007, Biometrics.

[3]  Theodore W. Berger,et al.  The Neurobiological Basis of Cognition: Identification by Multi-Input, Multioutput Nonlinear Dynamic Modeling , 2010, Proceedings of the IEEE.

[4]  Theodore W. Berger,et al.  Estimating sparse Volterra models using group L1-regularization , 2010, 2010 Annual International Conference of the IEEE Engineering in Medicine and Biology.

[5]  Sergio Benedetto,et al.  Nonlinear Equalization of Digital Satellite Channels , 1982, IEEE J. Sel. Areas Commun..

[6]  Shizhong Xu,et al.  Genomewide Analysis of Epistatic Effects for Quantitative Traits in Barley , 2007, Genetics.

[7]  Emmanuel J. Cand The Restricted Isometry Property and Its Implications for Compressed Sensing , 2008 .

[8]  Holger Rauhut,et al.  Compressive Sensing with structured random matrices , 2012 .

[9]  Michael A. Saunders,et al.  Atomic Decomposition by Basis Pursuit , 1998, SIAM J. Sci. Comput..

[10]  T. Poggio,et al.  The Volterra Representation and the Wiener Expansion: Validity and Pitfalls , 1977 .

[11]  H. Zou The Adaptive Lasso and Its Oracle Properties , 2006 .

[12]  Anthony Widjaja,et al.  Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond , 2003, IEEE Transactions on Neural Networks.

[13]  Patrick M. Hayes,et al.  Regions of the genome that affect agronomic performance in two-row barley , 1996 .

[14]  Balas K. Natarajan,et al.  Sparse Approximate Solutions to Linear Systems , 1995, SIAM J. Comput..

[15]  Massimo Fornasier,et al.  Theoretical Foundations and Numerical Methods for Sparse Recovery , 2010, Radon Series on Computational and Applied Mathematics.

[16]  M. Yuan,et al.  Model selection and estimation in regression with grouped variables , 2006 .

[17]  H. Cordell Detecting gene–gene interactions that underlie human diseases , 2009, Nature Reviews Genetics.

[18]  H. Rauhut Compressive Sensing and Structured Random Matrices , 2009 .

[19]  Trevor Hastie,et al.  The Elements of Statistical Learning , 2001 .

[20]  Robert Tibshirani,et al.  The Elements of Statistical Learning: Data Mining, Inference, and Prediction, 2nd Edition , 2001, Springer Series in Statistics.

[21]  Robert D. Nowak,et al.  Sparse interactions: Identifying high-dimensional multilinear systems via compressed sensing , 2010, 2010 48th Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[22]  Vahid Tarokh,et al.  An Adaptive Greedy Algorithm With Application to Nonlinear Communications , 2010, IEEE Transactions on Signal Processing.

[23]  Walter Kellermann,et al.  Fast and Robust Adaptation of DFT-Domain Volterra Filters in Diagonal Coordinates Using Iterated Coefficient Updates , 2010, IEEE Transactions on Signal Processing.

[24]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[25]  Jianqing Fan,et al.  Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties , 2001 .

[26]  Georgios B. Giannakis,et al.  Online Adaptive Estimation of Sparse Signals: Where RLS Meets the $\ell_1$ -Norm , 2010, IEEE Transactions on Signal Processing.

[27]  Robert D. Nowak,et al.  Toeplitz Compressed Sensing Matrices With Applications to Sparse Channel Estimation , 2010, IEEE Transactions on Information Theory.

[28]  Georgios B. Giannakis,et al.  Sparsity-aware estimation of nonlinear Volterra kernels , 2009, 2009 3rd IEEE International Workshop on Computational Advances in Multi-Sensor Adaptive Processing (CAMSAP).

[29]  P. Bickel,et al.  SIMULTANEOUS ANALYSIS OF LASSO AND DANTZIG SELECTOR , 2008, 0801.1095.

[30]  V. J. Mathews,et al.  Polynomial Signal Processing , 2000 .

[31]  Bernhard Schölkopf,et al.  A Unifying View of Wiener and Volterra Theory and Polynomial Kernel Regression , 2006, Neural Computation.

[32]  Emmanuel J. Candès,et al.  Decoding by linear programming , 2005, IEEE Transactions on Information Theory.

[33]  Sriram V. Pemmaraju,et al.  Equitable colorings extend Chernoff-Hoeffding bounds , 2001, SODA '01.

[34]  Terence Tao,et al.  The Dantzig selector: Statistical estimation when P is much larger than n , 2005, math/0506081.

[35]  Robert D. Nowak,et al.  Invertibility of higher order moment matrices , 1995, IEEE Trans. Signal Process..

[36]  E. Candès The restricted isometry property and its implications for compressed sensing , 2008 .

[37]  Trevor J. Hastie,et al.  Genome-wide association analysis by lasso penalized logistic regression , 2009, Bioinform..

[38]  Massimo Fornasier,et al.  Compressive Sensing and Structured Random Matrices , 2010 .

[39]  R. Tibshirani,et al.  PATHWISE COORDINATE OPTIMIZATION , 2007, 0708.1485.