COLLOCATION BASED APPROACH FOR TRAINING RECURRENT NEURAL NETWORKS

Modelling long range time dependencies in recurrent neural networks is a well known problem in the field. As previously established, the method of gradient descent fails due to the issue of exponentially decaying/exploding back propagated error. Moreover, gradient descent is not scale invariant in the parameters it seeks to optimize. Recently, a Hessian free approach for training recurrent neural nets was proposed to address these 2 issues. Although, successful to a large extent the problem of exploding/decaying gradients re-appears depending on the random initializations. We propose a collocation based scheme for training networks nets and report its performance for simple architectures. The training of recurrent neural networks based on the proposed scheme and comparison of its performance with the HF-based technique both on the rate of success and the use of computational resources is currently under study with no conclusive results yet.