A Comprehensive Approach to Universal Piecewise Nonlinear Regression Based on Trees

In this paper, we investigate adaptive nonlinear regression and introduce tree based piecewise linear regression algorithms that are highly efficient and provide significantly improved performance with guaranteed upper bounds in an individual sequence manner. We use a tree notion in order to partition the space of regressors in a nested structure. The introduced algorithms adapt not only their regression functions but also the complete tree structure while achieving the performance of the “best” linear mixture of a doubly exponential number of partitions, with a computational complexity only polynomial in the number of nodes of the tree. While constructing these algorithms, we also avoid using any artificial “weighting” of models (with highly data dependent parameters) and, instead, directly minimize the final regression error, which is the ultimate performance goal. The introduced methods are generic such that they can readily incorporate different tree construction methods such as random trees in their framework and can use different regressor or partitioning functions as demonstrated in the paper.

[1]  Damiana Lazzaro,et al.  Nonlinear Filtering for Sparse Signal Recovery From Incomplete Measurements , 2009, IEEE Transactions on Signal Processing.

[2]  Tamás Linder,et al.  Efficient adaptive algorithms and minimax bounds for zero-delay lossy source coding , 2004, IEEE Transactions on Signal Processing.

[3]  R. de Figueiredo The Volterra and Wiener theories of nonlinear systems , 1982, Proceedings of the IEEE.

[4]  Ann B. Lee,et al.  Treelets--An adaptive multi-scale basis for sparse unordered data , 2007, 0707.0481.

[5]  Jon Louis Bentley,et al.  Multidimensional Binary Search Trees in Database Applications , 1979, IEEE Transactions on Software Engineering.

[6]  Gil David,et al.  Hierarchical data organization , clustering and denoising via localized diffusion folders , 2011 .

[7]  Robert E. Schapire,et al.  Predicting Nearly As Well As the Best Pruning of a Decision Tree , 1995, COLT '95.

[8]  Rémi Munos,et al.  Linear regression with random projections , 2012, J. Mach. Learn. Res..

[9]  Andrew C. Singer,et al.  Universal linear least squares prediction: Upper and lower bounds , 2002, IEEE Trans. Inf. Theory.

[10]  Tareq Y. Al-Naffouri,et al.  Transient analysis of data-normalized adaptive filters , 2003, IEEE Trans. Signal Process..

[11]  Jiashu Zhang,et al.  A Novel Adaptive Nonlinear Filter-Based Pipelined Feedforward Second-Order Volterra Architecture , 2009, IEEE Transactions on Signal Processing.

[12]  Ali H. Sayed,et al.  On the learning mechanism of adaptive filters , 2000, IEEE Trans. Signal Process..

[13]  Eweda Eweda,et al.  Comparison of RLS, LMS, and sign algorithms for tracking randomly time-varying channels , 1994, IEEE Trans. Signal Process..

[14]  Marco F. Huber Chebyshev polynomial Kalman filter , 2013, Digit. Signal Process..

[15]  Aníbal R. Figueiras-Vidal,et al.  New algorithms for improved adaptive convex combination of LMS transversal filters , 2005, IEEE Transactions on Instrumentation and Measurement.

[16]  Elad Hazan,et al.  Logarithmic regret algorithms for online convex optimization , 2006, Machine Learning.

[17]  Tsachy Weissman,et al.  Universal FIR MMSE Filtering , 2009, IEEE Transactions on Signal Processing.

[18]  Andrew C. Singer,et al.  Universal linear prediction by model order weighting , 1999, IEEE Trans. Signal Process..

[19]  Lei Wang,et al.  A Novel Nonlinear Regression Approach for Efficient and Accurate Image Matting , 2013, IEEE Signal Processing Letters.

[20]  Damiana Lazzaro,et al.  Fast Sparse Image Reconstruction Using Adaptive Nonlinear Filtering , 2011, IEEE Transactions on Image Processing.

[21]  A. Figueiras-Vidal,et al.  Multiple plant identifier via adaptive LMS convex combination , 2003, IEEE International Symposium on Intelligent Signal Processing, 2003.

[22]  Georgios B. Giannakis,et al.  Sparse Volterra and Polynomial Regression Models: Recoverability and Estimation , 2011, IEEE Transactions on Signal Processing.

[23]  Andrew C. Singer,et al.  Nonlinear Autoregressive Modeling and Estimation in the Presence of Noise , 1994 .

[24]  Frans M. J. Willems,et al.  The context-tree weighting method: basic properties , 1995, IEEE Trans. Inf. Theory.

[25]  Roman Rosipal,et al.  Kernel Partial Least Squares Regression in Reproducing Kernel Hilbert Space , 2002, J. Mach. Learn. Res..

[26]  Tamás Linder,et al.  A zero-delay sequential scheme for lossy coding of individual sequences , 2001, IEEE Trans. Inf. Theory.

[27]  Zhigang Zeng,et al.  H∞ Filtering for Neutral Systems With Mixed Delays and Multiplicative Noises , 2012, IEEE Trans. Circuits Syst. II Express Briefs.

[28]  Wei Yang,et al.  H∞ filtering for nonlinear stochastic systems with sensor saturation, quantization and random packet losses , 2012, Signal Process..

[29]  Bor-Sen Chen,et al.  Robust H∞ filtering for nonlinear stochastic systems , 2005 .

[30]  Georg Zeitler,et al.  Universal Piecewise Linear Prediction Via Context Trees , 2007, IEEE Transactions on Signal Processing.

[31]  Andy J. Keane,et al.  Some Greedy Learning Algorithms for Sparse Regression and Classification with Mercer Kernels , 2003, J. Mach. Learn. Res..

[32]  Ali H. Sayed,et al.  Steady-State MSE Performance Analysis of Mixture Approaches to Adaptive Filtering , 2010, IEEE Transactions on Signal Processing.

[33]  Suleyman Serdar Kozat,et al.  Competitive Randomized Nonlinear Prediction Under Additive Noise , 2010, IEEE Signal Processing Letters.

[34]  N. Sloane,et al.  Some Doubly Exponential Sequences , 1973, The Fibonacci Quarterly.

[35]  Jun Hu,et al.  Robust variance-constrained filtering for a class of nonlinear stochastic systems with missing measurements , 2010, Signal Process..

[36]  Alfred O. Hero,et al.  Tree-structured nonlinear signal modeling and prediction , 1999, IEEE Trans. Signal Process..

[37]  Rodney X. Sturdivant,et al.  Applied Logistic Regression: Hosmer/Applied Logistic Regression , 2005 .

[38]  E. Lorenz Deterministic nonperiodic flow , 1963 .

[39]  Zhigang Zeng,et al.  $H_{\infty}$ Filtering for Neutral Systems With Mixed Delays and Multiplicative Noises , 2012, IEEE Transactions on Circuits and Systems II: Express Briefs.

[40]  Tareq Y. Al-Naffouri,et al.  Transient analysis of adaptive filters with error nonlinearities , 2003, IEEE Trans. Signal Process..

[41]  Eric L. Miller,et al.  Nonlinear Filtering Using a New Proposal Distribution and the Improved Fast Gauss Transform With Tighter Performance Bounds , 2008, IEEE Transactions on Signal Processing.

[42]  Vladimir Vovk,et al.  Predicting nearly as well as the best pruning of a decision tree through dynamic programming scheme , 2001, Theor. Comput. Sci..

[43]  Manfred K. Warmuth,et al.  Predicting nearly as well as the best pruning of a planar decision graph , 2002, Theor. Comput. Sci..

[44]  Ali H. Sayed,et al.  Mean-square performance of a convex combination of two adaptive filters , 2006, IEEE Transactions on Signal Processing.

[45]  Jesús Alcalá-Fdez,et al.  KEEL Data-Mining Software Tool: Data Set Repository, Integration of Algorithms and Experimental Analysis Framework , 2011, J. Multiple Valued Log. Soft Comput..

[46]  Yingmin Jia,et al.  H-infinity filtering for a class of nonlinear discrete-time systems based on unscented transform , 2010, Signal Process..

[47]  David W. Hosmer,et al.  Applied Logistic Regression , 1991 .

[48]  Paulo Sergio Ramirez,et al.  Fundamentals of Adaptive Filtering , 2002 .

[49]  Andrew C. Singer,et al.  Constrained Complexity Generalized Context-Tree Algorithms , 2007, 2007 IEEE/SP 14th Workshop on Statistical Signal Processing.

[50]  Danilo Comminiello,et al.  Nonlinear spline adaptive filtering , 2013, Signal Process..

[51]  Sanjoy Dasgupta,et al.  Random projection trees for vector quantization , 2008, 2008 46th Annual Allerton Conference on Communication, Control, and Computing.

[52]  Alberto Carini,et al.  Fourier nonlinear filters , 2014, Signal Process..