Active Tuning

We introduce Active Tuning, a novel paradigm for optimizing the internal dynamics of recurrent neural networks (RNNs) on the fly. In contrast to the conventional sequence-to-sequence mapping scheme, Active Tuning decouples the RNN's recurrent neural activities from the input stream, using the unfolding temporal gradient signal to tune the internal dynamics into the data stream. As a consequence, the model output depends only on its internal hidden dynamics and the closed-loop feedback of its own predictions; its hidden state is continuously adapted by means of the temporal gradient resulting from backpropagating the discrepancy between the signal observations and the model outputs through time. In this way, Active Tuning infers the signal actively but indirectly based on the originally learned temporal patterns, fitting the most plausible hidden state sequence into the observations. We demonstrate the effectiveness of Active Tuning on several time series prediction benchmarks, including multiple super-imposed sine waves, a chaotic double pendulum, and spatiotemporal wave dynamics. Active Tuning consistently improves the robustness, accuracy, and generalization abilities of all evaluated models. Moreover, networks trained for signal prediction and denoising can be successfully applied to a much larger range of noise conditions with the help of Active Tuning. Thus, given a capable time series predictor, Active Tuning enhances its online signal filtering, denoising, and reconstruction abilities without the need for additional training.

[1]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[2]  Andreas Zell,et al.  Optimizing recurrent reservoirs with neuro-evolution , 2016, Neurocomputing.

[3]  Martin V. Butz,et al.  Simultaneously emerging Braitenberg codes and compositionality , 2011, Adapt. Behav..

[4]  Martin V. Butz,et al.  Inferring, Predicting, and Denoising Causal Wave Dynamics , 2020, ICANN.

[5]  Andreas Zell,et al.  Dynamic Cortex Memory: Enhancing Recurrent Neural Networks for Gradient-Based Sequence Learning , 2014 .

[6]  Yoshua Bengio,et al.  Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling , 2014, ArXiv.

[7]  Martin V. Butz,et al.  Balanced echo state networks , 2012, Neural Networks.

[8]  Martin V. Butz,et al.  Gradient-Based Learning of Compositional Dynamics with Modular RNNs , 2019, ICANN.

[9]  Timo Hartmann,et al.  Chaos: A Program Collection for the PC , 1994 .

[10]  Martin V. Butz,et al.  Learning, Planning, and Control in a Monolithic Neural Event Inference Architecture , 2018, Neural Networks.

[11]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[12]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[13]  Matthias Bethge,et al.  Generalisation in humans and deep neural networks , 2018, NeurIPS.

[14]  William H. Press,et al.  Numerical recipes in C. The art of scientific computing , 1987 .

[15]  Yu Tsao,et al.  Speech enhancement based on deep denoising autoencoder , 2013, INTERSPEECH.

[16]  Herbert Jaeger,et al.  The''echo state''approach to analysing and training recurrent neural networks , 2001 .

[17]  Andreas Zell,et al.  An analysis of Dynamic Cortex Memory networks , 2015, 2015 International Joint Conference on Neural Networks (IJCNN).

[18]  Michelle Girvan,et al.  Hybrid Forecasting of Chaotic Processes: Using Machine Learning in Conjunction with a Knowledge-Based Model , 2018, Chaos.

[19]  Alex Graves,et al.  Neural Machine Translation in Linear Time , 2016, ArXiv.

[20]  Vladlen Koltun,et al.  An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Modeling , 2018, ArXiv.

[21]  J. E. Glynn,et al.  Numerical Recipes: The Art of Scientific Computing , 1989 .

[22]  Paul J. Werbos,et al.  Backpropagation Through Time: What It Does and How to Do It , 1990, Proc. IEEE.