Causal Navigation by Continuous-time Neural Networks

Imitation learning enables high-fidelity, vision-based learning of policies within rich, photorealistic environments. However, such techniques often rely on traditional discrete-time neural models and face difficulties in generalizing to domain shifts by failing to account for the causal relationships between the agent and the environment. In this paper, we propose a theoretical and experimental framework for learning causal representations using continuous-time neural networks, specifically over their discrete-time counterparts. We evaluate our method in the context of visual-control learning of drones over a series of complex tasks, ranging from shortand long-term navigation, to chasing static and dynamic objects through photorealistic environments. Our results demonstrate that causal continuous-time deep models can perform robust navigation tasks, where advanced recurrent models fail. These models learn complex causal control representations directly from raw visual inputs and scale to solve a variety of tasks using imitation learning.

[1]  David Duvenaud,et al.  FFJORD: Free-form Continuous Dynamics for Scalable Reversible Generative Models , 2018, ICLR.

[2]  Stefano Ermon,et al.  Generative Adversarial Imitation Learning , 2016, NIPS.

[3]  Daphne Koller,et al.  Learning Continuous Time Bayesian Networks , 2002, UAI.

[4]  Geoffrey J. Gordon,et al.  A Reduction of Imitation Learning and Structured Prediction to No-Regret Online Learning , 2010, AISTATS.

[5]  Mariusz Bojarski,et al.  VisualBackProp: efficient visualization of CNNs , 2018 .

[6]  David Duvenaud,et al.  Neural Ordinary Differential Equations , 2018, NeurIPS.

[7]  Henry A. Kautz,et al.  Extending Continuous Time Bayesian Networks , 2005, AAAI.

[8]  Iain Murray,et al.  Neural Spline Flows , 2019, NeurIPS.

[9]  J. Pearl Causal inference in statistics: An overview , 2009 .

[10]  Xiaojuan Ma,et al.  Adversarial Imitation Learning from Incomplete Demonstrations , 2019, IJCAI.

[11]  Surya Ganguli,et al.  On the Expressive Power of Deep Neural Networks , 2016, ICML.

[12]  Karl J. Friston,et al.  Bilinear dynamical systems , 2005, Philosophical Transactions of the Royal Society B: Biological Sciences.

[13]  Sergey Levine,et al.  Causal Confusion in Imitation Learning , 2019, NeurIPS.

[14]  Jana Kosecka,et al.  Visual Representations for Semantic Target Driven Navigation , 2018, 2019 International Conference on Robotics and Automation (ICRA).

[15]  Stefan Schaal,et al.  Is imitation learning the route to humanoid robots? , 1999, Trends in Cognitive Sciences.

[16]  Omri Azencot,et al.  Lipschitz Recurrent Neural Networks , 2020, ICLR.

[17]  Radu Grosu,et al.  Model-based versus Model-free Deep Reinforcement Learning for Autonomous Racing Cars , 2021, ArXiv.

[18]  Andrew Y. Ng,et al.  Pharmacokinetics of a novel formulation of ivermectin after administration to goats , 2000, ICML.

[19]  Vladlen Koltun,et al.  Semi-parametric Topological Memory for Navigation , 2018, ICLR.

[20]  Bernhard Schölkopf,et al.  Algorithmic Recourse: from Counterfactual Explanations to Interventions , 2020, FAccT.

[21]  Vladlen Koltun,et al.  Learning to Control PDEs with Differentiable Physics , 2020, ICLR.

[22]  Radu Grosu,et al.  Gershgorin Loss Stabilizes the Recurrent Neural Network Compartment of an End-to-end Robot Learning Scheme , 2020, 2020 IEEE International Conference on Robotics and Automation (ICRA).

[23]  Shie Mannor,et al.  End-to-End Differentiable Adversarial Imitation Learning , 2017, ICML.

[24]  Hajime Asama,et al.  Dissecting Neural ODEs , 2020, NeurIPS.

[25]  Bernhard Scholkopf Causality for Machine Learning , 2019 .

[26]  Richard Gordon,et al.  OpenWorm: overview and recent advances in integrative biological simulation of Caenorhabditis elegans , 2018, Philosophical Transactions of the Royal Society B.

[27]  Siddhartha Mishra,et al.  Coupled Oscillatory Recurrent Neural Network (coRNN): An accurate and (gradient) stable architecture for learning long time dependencies , 2020, ICLR.

[28]  Masashi Sugiyama,et al.  Imitation Learning from Imperfect Demonstration , 2019, ICML.

[29]  M. Breakspear Dynamic models of large-scale brain activity , 2017, Nature Neuroscience.

[30]  Thomas A. Henzinger,et al.  GoTube: Scalable Stochastic Verification of Continuous-Depth Models , 2021, ArXiv.

[31]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[32]  M. L. Chambers The Mathematical Theory of Optimal Processes , 1965 .

[33]  Kurt Keutzer,et al.  ANODE: Unconditionally Accurate Memory-Efficient Gradients for Neural ODEs , 2019, IJCAI.

[34]  Mathias Lechner,et al.  Learning Long-Term Dependencies in Irregularly-Sampled Time Series , 2020, NeurIPS.

[35]  Marcin Andrychowicz,et al.  One-Shot Imitation Learning , 2017, NIPS.

[36]  Brandon M. Greenwell,et al.  Interpretable Machine Learning , 2019, Hands-On Machine Learning with R.

[37]  Vladlen Koltun,et al.  Learning by Cheating , 2019, CoRL.

[38]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[39]  Trevor Darrell,et al.  Fighting Copycat Agents in Behavioral Cloning from Observation Histories , 2020, NeurIPS.

[40]  Michael C. Mozer,et al.  Discrete Event, Continuous Time RNNs , 2017, ArXiv.

[41]  Louis-Philippe Morency,et al.  Diverse and Admissible Trajectory Forecasting through Multimodal Context Understanding , 2020, ECCV.

[42]  Vladlen Koltun,et al.  An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Modeling , 2018, ArXiv.

[43]  Rahul Sukthankar,et al.  Cognitive Mapping and Planning for Visual Navigation , 2017, International Journal of Computer Vision.

[44]  David Duvenaud,et al.  Latent Ordinary Differential Equations for Irregularly-Sampled Time Series , 2019, NeurIPS.

[45]  O. Nevanlinna Remarks on Picard-Lindelöf iteration , 1989 .

[46]  Sergey Levine,et al.  Can Autonomous Vehicles Identify, Recover From, and Adapt to Distribution Shifts? , 2020, ICML.

[47]  Yuichi Nakamura,et al.  Approximation of dynamical systems by continuous time recurrent neural networks , 1993, Neural Networks.

[48]  Judea Pearl,et al.  Probabilistic reasoning in intelligent systems - networks of plausible inference , 1991, Morgan Kaufmann series in representation and reasoning.

[49]  Mathias Lechner,et al.  Closed-form Continuous-Depth Models , 2021, ArXiv.

[50]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[51]  Radu Grosu,et al.  Neural circuit policies enabling auditable autonomy , 2020, Nature Machine Intelligence.

[52]  Yee Whye Teh,et al.  Augmented Neural ODEs , 2019, NeurIPS.

[53]  Jitendra Malik,et al.  Unifying Map and Landmark Based Representations for Visual Navigation , 2017, ArXiv.

[54]  Austin R. Benson,et al.  Neural Jump Stochastic Differential Equations , 2019, NeurIPS.

[55]  Bernhard Schölkopf,et al.  From Deterministic ODEs to Dynamic Structural Causal Models , 2016, UAI.

[56]  Radu Grosu,et al.  On The Verification of Neural ODEs with Stochastic Guarantees , 2020, AAAI.

[57]  Bernhard Schölkopf,et al.  Elements of Causal Inference: Foundations and Learning Algorithms , 2017 .

[58]  David Duvenaud,et al.  Latent ODEs for Irregularly-Sampled Time Series , 2019, ArXiv.

[59]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.

[60]  Danielle S. Bassett,et al.  Dynamic representations in networked neural systems , 2020, Nature Neuroscience.

[61]  Radu Grosu,et al.  A Natural Lottery Ticket Winner: Reinforcement Learning with Ordinary Neural Circuits , 2020, ICML.

[62]  Peter Norvig,et al.  Artificial Intelligence: A Modern Approach , 1995 .

[63]  Daphne Koller,et al.  Continuous Time Bayesian Networks , 2012, UAI.

[64]  Vladlen Koltun,et al.  Deep Drone Acrobatics , 2020, Robotics: Science and Systems.

[65]  Thomas A. Henzinger,et al.  Adversarial Training is Not Ready for Robot Learning , 2021, 2021 IEEE International Conference on Robotics and Automation (ICRA).

[66]  Bernhard Schölkopf,et al.  From Ordinary Differential Equations to Structural Causal Models: the deterministic case , 2013, UAI.

[67]  Ashish Kapoor,et al.  AirSim Drone Racing Lab , 2020, NeurIPS.

[68]  Yoshua Bengio,et al.  Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling , 2014, ArXiv.

[69]  Radu Grosu,et al.  Designing Worm-inspired Neural Networks for Interpretable Robotic Control , 2019, 2019 International Conference on Robotics and Automation (ICRA).

[70]  Sergey Levine,et al.  Deep Imitative Models for Flexible Inference, Planning, and Control , 2018, ICLR.

[71]  Ruslan Salakhutdinov,et al.  Learning to Explore using Active Neural SLAM , 2020, ICLR.

[72]  Lorenz Wellhausen,et al.  Learning quadrupedal locomotion over challenging terrain , 2020, Science Robotics.

[73]  Ruslan Salakhutdinov,et al.  Neural Topological SLAM for Visual Navigation , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[74]  Ramin Hasani,et al.  Sparse Flows: Pruning Continuous-depth Models , 2021, NeurIPS.

[75]  M. E. Jakobsen,et al.  Causal structure learning from time series: Large regression coefficients may predict causal links better in practice than small p-values , 2020, NeurIPS.

[76]  Joshua B. Tenenbaum,et al.  Learning to Learn Causal Models , 2010, Cogn. Sci..

[77]  Vincent Y. F. Tan,et al.  On Robustness of Neural Ordinary Differential Equations , 2020, ICLR.

[78]  Yannick Schroecker,et al.  State Aware Imitation Learning , 2017, NIPS.

[79]  Tom Burr,et al.  Causation, Prediction, and Search , 2003, Technometrics.

[80]  Andreas Geiger,et al.  Conditional Affordance Learning for Driving in Urban Environments , 2018, CoRL.

[81]  Josiah P. Hanna,et al.  An Imitation from Observation Approach to Transfer Learning with Dynamics Mismatch , 2020, NeurIPS.

[82]  Karl J. Friston,et al.  Dynamic causal modelling , 2003, NeuroImage.

[83]  Sergey Levine,et al.  One-Shot Hierarchical Imitation Learning of Compound Visuomotor Tasks , 2018, ArXiv.