Predicting Conditional Probability Densities of Stationary Stochastic Time Series

Abstract Feedforward neural networks applied to time series prediction are usually trained to predict the next time step x(t + 1) as a function of m previous values, x(t) := (x(t), x(t + 1),…, x(t − m + 1)), which, if a sum-of-squares error function is chosen, results in predicting the conditional mean 〈y|x(t)〉. However, further information about the distribution is lost, which is a serious drawback especially in the case of multimodality, where the conditional mean alone turns out to be a rather insufficient or even misleading quantity. The only satisfactory approach in the general case is therefore to predict the whole conditional probability density for the time series, P(x(t + 1)|x(t), x(t − m 1),…, x(t − + 1)). We deduce here a two-hidden-layer universal approximator network for modelling this function, and develop a training algorithm from maximum likelihood. The method is tested on three time series of different nature, which will demonstrate how state-space dependent variances and multimodal transitions can be learned. We will finally note comparisons with other recent neural network approaches to this problem, and will state results on a benchmark problem. © 1997 Elsevier Science Ltd. All Rights Reserved.

[1]  I N Bronstein,et al.  Taschenbuch der Mathematik , 1966 .

[2]  David J. C. MacKay,et al.  A Practical Bayesian Framework for Backpropagation Networks , 1992, Neural Computation.

[3]  Donald F. Specht,et al.  Probabilistic neural networks , 1990, Neural Networks.

[4]  Hans Henrik Thodberg,et al.  A review of Bayesian neural networks with an application to near infrared spectroscopy , 1996, IEEE Trans. Neural Networks.

[5]  Andreas S. Weigend,et al.  Predictions with Confidence Intervals ( Local Error Bars ) , 1994 .

[6]  Christopher M. Bishop,et al.  Neural networks for pattern recognition , 1995 .

[7]  Kurt Hornik,et al.  Approximation capabilities of multilayer feedforward networks , 1991, Neural Networks.

[8]  Ashok N. Srivastava,et al.  Predicting conditional probability distributions: a connectionist approach , 1995, Int. J. Neural Syst..

[9]  Geoffrey E. Hinton,et al.  Bayesian Learning for Neural Networks , 1995 .

[10]  D. W. Allen,et al.  Learning Time Series by Neural Networks , 1994 .

[11]  Kurt Hornik,et al.  Multilayer feedforward networks are universal approximators , 1989, Neural Networks.

[12]  Dirk Husmeier,et al.  A universal approximator network for learning conditional probability densities , 1997 .

[13]  Athanasios Papoulis,et al.  Probability, Random Variables and Stochastic Processes , 1965 .

[14]  David J. C. MacKay,et al.  Bayesian Interpolation , 1992, Neural Computation.

[15]  C. Bishop Mixture density networks , 1994 .

[16]  Robert A. Jacobs,et al.  Increased rates of convergence through learning rate adaptation , 1987, Neural Networks.

[17]  Ralph Neuneier,et al.  Estimation of Conditional Densities: A Comparison of Neural Network Approaches , 1994 .