Boosting and Neural Networks for the Prediction of Heterokedastic Time Series

This work develops and evaluates new algorithms based on neural networks and boosting techniques, designed to model and predict heteroskedastic time series. The main novel elements of these new algorithms are as follows: a) in regard to neural networks, the simultaneous estimation of trend and volatility through the maximization of the likelihood; b) in regard to boosting, its simultaneous application to the likelihood trend and volatility components, and the use of likelihood-based models (e.g., GARCH) as the base hypothesis rather than gradient adjustment techniques using least squares. The behavior of the proposed algorithms is evaluated over simulated data and over the Standard & Poor's 500 Index returns series, resulting in frequent and signi cant improvements in relation to the ARMA-GARCH models.

[1]  Peter Bühlmann,et al.  Volatility and risk estimation with linear and nonlinear methods based on high frequency data , 2001 .

[2]  A. Neil Burgess,et al.  Neural networks in financial engineering: a study in methodology , 1997, IEEE Trans. Neural Networks.

[3]  Yoav Freund,et al.  Boosting the margin: A new explanation for the effectiveness of voting methods , 1997, ICML.

[4]  M. Medeiros,et al.  Building Neural Network Models for Time Series: A Statistical Approach , 2002 .

[5]  Jonathan D. Cryer,et al.  Time Series Analysis , 1986 .

[6]  Peter Bühlmann,et al.  An algorithm for nonparametric GARCH modelling , 2002 .

[7]  Stephen Taylor,et al.  Forecasting Economic Time Series , 1979 .

[8]  Peter Tiño,et al.  Financial volatility trading using recurrent neural networks , 2001, IEEE Trans. Neural Networks.

[9]  Gunnar Rätsch,et al.  An Introduction to Boosting and Leveraging , 2002, Machine Learning Summer School.

[10]  T. Bollerslev,et al.  Generalized autoregressive conditional heteroskedasticity , 1986 .

[11]  Giorgio Valentini,et al.  Ensembles of Learning Machines , 2002, WIRN.

[12]  Allan Pinkus,et al.  Multilayer Feedforward Networks with a Non-Polynomial Activation Function Can Approximate Any Function , 1991, Neural Networks.

[13]  Georg Dorffner,et al.  Neural Networks for Time Series Processing , 1996 .

[14]  Amanda J. C. Sharkey,et al.  Multi-Net Systems , 1999 .

[15]  P. Bühlmann,et al.  Volatility estimation with functional gradient descent for very high-dimensional financial time series , 2003 .

[16]  N. Burgess,et al.  Modelling market volatilities: the neural network perspective , 1997 .

[17]  N. Wermuth,et al.  Nonlinear Time Series : Nonparametric and Parametric Methods , 2005 .

[18]  Leo Breiman,et al.  Prediction Games and Arcing Algorithms , 1999, Neural Computation.

[19]  E. Dockner,et al.  Forecasting Time-dependent Conditional Densities: A Semi-non- parametric Neural Network Approach , 2000 .

[20]  C. Granger,et al.  Modelling Nonlinear Economic Relationships , 1995 .

[21]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.

[22]  Andreas S. Weigend,et al.  Time Series Prediction: Forecasting the Future and Understanding the Past , 1994 .

[23]  D. B. Preston Spectral Analysis and Time Series , 1983 .

[24]  P. Young,et al.  Time series analysis, forecasting and control , 1972, IEEE Transactions on Automatic Control.

[25]  R. Schapire The Strength of Weak Learnability , 1990, Machine Learning.

[26]  Richard D. De Veaux,et al.  Estimating Prediction Intervals for Arti cial Neural Networks , 2022 .

[27]  S. Hyakin,et al.  Neural Networks: A Comprehensive Foundation , 1994 .

[28]  J. Friedman Special Invited Paper-Additive logistic regression: A statistical view of boosting , 2000 .

[29]  Halbert White,et al.  Improved Rates and Asymptotic Normality for Nonparametric Neural Network Estimators , 1999, IEEE Trans. Inf. Theory.

[30]  Carlo Novara,et al.  Nonlinear Time Series , 2003 .

[31]  Jooyoung Park,et al.  Approximation and Radial-Basis-Function Networks , 1993, Neural Computation.

[32]  J. Friedman Greedy function approximation: A gradient boosting machine. , 2001 .

[33]  Kurt Hornik,et al.  Stationary and Integrated Autoregressive Neural Network Processes , 2000, Neural Computation.

[34]  Adam Krzyzak,et al.  Nonparametric estimation and classification using radial basis function nets and empirical risk minimization , 1996, IEEE Trans. Neural Networks.

[35]  Federico Girosi,et al.  On the Relationship between Generalization Error, Hypothesis Complexity, and Sample Complexity for Radial Basis Functions , 1996, Neural Computation.

[36]  Eric R. Ziegel,et al.  Analysis of Financial Time Series , 2002, Technometrics.

[37]  Yoram Singer,et al.  Improved Boosting Algorithms Using Confidence-rated Predictions , 1998, COLT' 98.

[38]  Terrence L. Fine,et al.  Parameter Convergence and Learning Curves for Neural Networks , 1999, Neural Computation.

[39]  D. Signorini,et al.  Neural networks , 1995, The Lancet.