This paper points out the flaws in using the extended Kalman filter (EKE) and introduces an improvement, the unscented Kalman filter (UKF), proposed by Julier and Uhlman (1997). A central and vital operation performed in the Kalman filter is the propagation of a Gaussian random variable (GRV) through the system dynamics. In the EKF the state distribution is approximated by a GRV, which is then propagated analytically through the first-order linearization of the nonlinear system. This can introduce large errors in the true posterior mean and covariance of the transformed GRV, which may lead to sub-optimal performance and sometimes divergence of the filter. The UKF addresses this problem by using a deterministic sampling approach. The state distribution is again approximated by a GRV, but is now represented using a minimal set of carefully chosen sample points. These sample points completely capture the true mean and covariance of the GRV, and when propagated through the true nonlinear system, captures the posterior mean and covariance accurately to the 3rd order (Taylor series expansion) for any nonlinearity. The EKF in contrast, only achieves first-order accuracy. Remarkably, the computational complexity of the UKF is the same order as that of the EKF. Julier and Uhlman demonstrated the substantial performance gains of the UKF in the context of state-estimation for nonlinear control. Machine learning problems were not considered. We extend the use of the UKF to a broader class of nonlinear estimation problems, including nonlinear system identification, training of neural networks, and dual estimation problems. In this paper, the algorithms are further developed and illustrated with a number of additional examples.
[1]
D. Rubin,et al.
Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper
,
1977
.
[2]
Sharad Singhal,et al.
Training Multilayer Perceptrons with the Extende Kalman Algorithm
,
1988,
NIPS.
[3]
Lee A. Feldkamp,et al.
Decoupled extended Kalman filter training of feedforward layered networks
,
1991,
IJCNN-91-Seattle International Joint Conference on Neural Networks.
[4]
Steve Rogers,et al.
Adaptive Filter Theory
,
1996
.
[5]
Eric A. Wan,et al.
Neural dual extended Kalman filtering: applications in speech enhancement and monaural blind signal separation
,
1997,
Neural Networks for Signal Processing VII. Proceedings of the 1997 IEEE Signal Processing Society Workshop.
[6]
Jeffrey K. Uhlmann,et al.
New extension of the Kalman filter to nonlinear systems
,
1997,
Defense, Security, and Sensing.
[7]
Arnaud Doucet,et al.
Sequential Monte Carlo methods for optimisation of neural network models
,
1998
.
[8]
Rudolph van der Merwe,et al.
Dual Estimation and the Unscented Transformation
,
1999,
NIPS.
[9]
Nando de Freitas,et al.
The Unscented Particle Filter
,
2000,
NIPS.
[10]
Simon J. Julier,et al.
The scaled unscented transformation
,
2002,
Proceedings of the 2002 American Control Conference (IEEE Cat. No.CH37301).