GoTube: Scalable Stochastic Verification of Continuous-Depth Models

We introduce a new stochastic verification algorithm that formally quantifies the behavioral robustness of any timecontinuous process formulated as a continuous-depth model. Our algorithm solves a set of global optimization (Go) problems over a given time horizon to construct a tight enclosure (Tube) of the set of all process executions starting from a ball of initial states. We call our algorithm GoTube. Through its construction, GoTube ensures that the bounding tube is conservative up to a desired probability and up to a desired tightness. GoTube is implemented in JAX and optimized to scale to complex continuous-depth neural network models. Compared to advanced reachability analysis tools for timecontinuous neural networks, GoTube does not accumulate overapproximation errors between time steps and avoids the infamous wrapping effect inherent in symbolic techniques. We show that GoTube substantially outperforms state-of-theart verification tools in terms of the size of the initial ball, speed, time-horizon, task completion, and scalability on a large set of experiments. GoTube is stable and sets the stateof-the-art in terms of its ability to scale to time horizons well beyond what has been previously possible.

[1]  Guido Sanguinetti,et al.  A Statistical Approach for Computing Reachability of Non-linear and Stochastic Dynamical Systems , 2014, QEST.

[2]  Mathias Lechner,et al.  Closed-form Continuous-Depth Models , 2021, ArXiv.

[3]  Timothy A. Mann,et al.  On the Effectiveness of Interval Bound Propagation for Training Verifiably Robust Models , 2018, ArXiv.

[4]  Pushmeet Kohli,et al.  A Unified View of Piecewise Linear Neural Network Verification , 2017, NeurIPS.

[5]  Thomas A. Henzinger,et al.  Reachability Verification for Hybrid Automata , 1998, HSCC.

[6]  Joseph DelPreto,et al.  Plug-and-play supervisory control using muscle and brain signals for real-time gesture and error detection , 2018, Autonomous Robots.

[7]  Taylor T. Johnson,et al.  Improved Geometric Path Enumeration for Verifying ReLU Neural Networks , 2020, CAV.

[8]  Radu Grosu,et al.  Neural circuit policies enabling auditable autonomy , 2020, Nature Machine Intelligence.

[9]  Radu Grosu,et al.  On The Verification of Neural ODEs with Stochastic Guarantees , 2020, AAAI.

[10]  Martin Fränzle,et al.  Engineering constraint solvers for automatic analysis of probabilistic hybrid automata , 2010, J. Log. Algebraic Methods Program..

[11]  Matthew Mirman,et al.  Differentiable Abstract Interpretation for Provably Robust Neural Networks , 2018, ICML.

[12]  Rajeev Alur,et al.  Efficient Reachability Analysis of Hierarchical Reactive Machines , 2000, CAV.

[13]  Yarin Gal,et al.  Uncertainty in Deep Learning , 2016 .

[14]  Cho-Jui Hsieh,et al.  Efficient Neural Network Robustness Certification with General Activation Functions , 2018, NeurIPS.

[15]  Li Liu,et al.  A Review of Uncertainty Quantification in Deep Learning: Techniques, Applications and Challenges , 2020, Inf. Fusion.

[16]  Manfred Morari,et al.  Efficient and Accurate Estimation of Lipschitz Constants for Deep Neural Networks , 2019, NeurIPS.

[17]  G. William Walster,et al.  Global Optimization Using Interval Analysis: Revised and Expanded , 2007 .

[18]  Radu Grosu,et al.  Efficient modeling of complex Analog integrated circuits using neural networks , 2016, 2016 12th Conference on Ph.D. Research in Microelectronics and Electronics (PRIME).

[19]  Juan Pablo Vielma,et al.  The Convex Relaxation Barrier, Revisited: Tightened Single-Neuron Relaxations for Neural Network Verification , 2020, NeurIPS.

[20]  Jonathon Shlens,et al.  Explaining and Harnessing Adversarial Examples , 2014, ICLR.

[21]  Nicolas Vayatis,et al.  Global optimization of Lipschitz functions , 2017, ICML.

[22]  Alexandre Donzé,et al.  Breach, A Toolbox for Verification and Parameter Synthesis of Hybrid Systems , 2010, CAV.

[23]  David Duvenaud,et al.  Neural Ordinary Differential Equations , 2018, NeurIPS.

[24]  James Kapinski,et al.  Simulation-Driven Reachability Using Matrix Measures , 2017, ACM Trans. Embed. Comput. Syst..

[25]  Radu Grosu,et al.  Lagrangian Reachtubes: The Next Generation , 2020, 2020 59th IEEE Conference on Decision and Control (CDC).

[26]  J. Zico Kolter,et al.  Provable defenses against adversarial examples via the convex outer adversarial polytope , 2017, ICML.

[27]  Ramin M. Hasani Interpretable Recurrent Neural Networks in Continuous-time Control Environments , 2020 .

[28]  David A. Wagner,et al.  Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.

[29]  Edmund M. Clarke,et al.  Satisfiability modulo ODEs , 2013, 2013 Formal Methods in Computer-Aided Design.

[30]  Mark A. Stadtherr,et al.  Verified Solution and Propagation of Uncertainty in Physiological Models , 2011, Reliab. Comput..

[31]  J. Kiefer,et al.  Asymptotic Minimax Character of the Sample Distribution Function and of the Classical Multinomial Estimator , 1956 .

[32]  Pushmeet Kohli,et al.  Adversarial Risk and the Dangers of Evaluating Against Weak Attacks , 2018, ICML.

[33]  Ezio Bartocci,et al.  Parallel reachability analysis of hybrid systems in XSpeed , 2019, International Journal on Software Tools for Technology Transfer.

[34]  Radu Grosu,et al.  Compositional neural-network modeling of complex analog circuits , 2017, 2017 International Joint Conference on Neural Networks (IJCNN).

[35]  Meenakshi D'Souza,et al.  LipBaB: Computing exact Lipschitz constant of ReLU networks , 2021, ICANN.

[36]  P. Henriksen,et al.  Efficient Neural Network Verification via Adaptive Refinement and Adversarial Search , 2020, ECAI.

[37]  Oded Maler,et al.  Systematic Simulation Using Sensitivity Analysis , 2007, HSCC.

[38]  S. A. Piyavskii An algorithm for finding the absolute extremum of a function , 1972 .

[39]  Mathias Lechner,et al.  Causal Navigation by Continuous-time Neural Networks , 2021, NeurIPS.

[40]  Mathias Lechner,et al.  Learning Long-Term Dependencies in Irregularly-Sampled Time Series , 2020, NeurIPS.

[41]  Radu Grosu,et al.  Lagrangian Reachabililty , 2017, CAV.

[42]  Radu Grosu,et al.  Probabilistic reachability analysis of the tap withdrawal circuit in caenorhabditis elegans , 2016, 2016 IEEE International High Level Design Validation and Test Workshop (HLDVT).

[43]  Thomas A. Henzinger,et al.  Adversarial Training is Not Ready for Robot Learning , 2021, 2021 IEEE International Conference on Robotics and Automation (ICRA).

[44]  Daniel Wilczak,et al.  CAPD: : DynSys: a flexible C++ toolbox for rigorous numerical analysis of dynamical systems , 2020, Commun. Nonlinear Sci. Numer. Simul..

[45]  Xin Chen,et al.  Flow*: An Analyzer for Non-linear Hybrid Systems , 2013, CAV.

[46]  Junfeng Yang,et al.  Efficient Formal Safety Analysis of Neural Networks , 2018, NeurIPS.

[47]  Geoffrey E. Hinton,et al.  Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.

[48]  Cho-Jui Hsieh,et al.  A Convex Relaxation Barrier to Tight Robustness Verification of Neural Networks , 2019, NeurIPS.

[49]  Radu Grosu,et al.  Gershgorin Loss Stabilizes the Recurrent Neural Network Compartment of an End-to-end Robot Learning Scheme , 2020, 2020 IEEE International Conference on Robotics and Automation (ICRA).

[50]  Edmund M. Clarke,et al.  SReach: A Probabilistic Bounded Delta-Reachability Analyzer for Stochastic Hybrid Systems , 2015, CMSB.

[51]  Meeko M. K. Oishi,et al.  Stochastic reachability of a target tube: Theory and computation , 2018, Autom..

[52]  David Duvenaud,et al.  Infinitely Deep Bayesian Neural Networks with Stochastic Differential Equations , 2021, ArXiv.

[53]  Nikolaus Hansen,et al.  Completely Derandomized Self-Adaptation in Evolution Strategies , 2001, Evolutionary Computation.

[54]  M. L. Chambers The Mathematical Theory of Optimal Processes , 1965 .

[55]  Murat Arcak,et al.  TIRA: toolbox for interval reachability analysis , 2019, HSCC.

[56]  Radu Grosu,et al.  Response Characterization for Auditing Cell Dynamics in Long Short-term Memory Networks , 2018, 2019 International Joint Conference on Neural Networks (IJCNN).

[57]  Lijun Zhang,et al.  Measurability and safety verification for stochastic hybrid systems , 2011, HSCC '11.

[58]  Fabian Immler,et al.  Verified Reachability Analysis of Continuous Systems , 2015, TACAS.

[59]  Xin Chen,et al.  Probabilistic Safety Verification of Stochastic Hybrid Systems Using Barrier Certificates , 2017, ACM Trans. Embed. Comput. Syst..

[60]  Radu Grosu,et al.  On-Off Center-Surround Receptive Fields for Accurate and Robust Image Classification , 2021, ICML.

[61]  Thomas A. Henzinger,et al.  Scalable Verification of Quantized Neural Networks , 2021, AAAI.

[62]  Yaroslav D. Sergeyev,et al.  Lipschitz global optimization methods in control problems , 2013, Autom. Remote. Control..

[63]  Paolo Zuliani,et al.  ProbReach: verified probabilistic delta-reachability for stochastic hybrid systems , 2014, HSCC.

[64]  Dongxu Li,et al.  Reachability Analysis of Nonlinear Systems Using Hybridization and Dynamics Scaling , 2020, FORMATS.

[65]  Pushmeet Kohli,et al.  Branch and Bound for Piecewise Linear Neural Network Verification , 2020, J. Mach. Learn. Res..

[66]  Radu Grosu,et al.  A novel Bayesian network-based fault prognostic method for semiconductor manufacturing process , 2017, 2017 IEEE International Conference on Industrial Technology (ICIT).

[67]  Murat Arcak,et al.  PIRK: Scalable Interval Reachability Analysis for High-Dimensional Nonlinear Systems , 2020, CAV.

[68]  Russ Tedrake,et al.  Evaluating Robustness of Neural Networks with Mixed Integer Programming , 2017, ICLR.

[69]  Mykel J. Kochenderfer,et al.  Reluplex: An Efficient SMT Solver for Verifying Deep Neural Networks , 2017, CAV.

[70]  Stefan Roth,et al.  Covariance Matrix Adaptation for Multi-objective Optimization , 2007, Evolutionary Computation.

[71]  Paolo Zuliani,et al.  ProbReach: A Tool for Guaranteed Reachability Analysis of Stochastic Hybrid Systems , 2015, SNR@CAV.

[72]  P. Massart The Tight Constant in the Dvoretzky-Kiefer-Wolfowitz Inequality , 1990 .

[73]  A. Zhigljavsky Stochastic Global Optimization , 2008, International Encyclopedia of Statistical Science.

[74]  David Duvenaud,et al.  Scalable Gradients for Stochastic Differential Equations , 2020, AISTATS.

[75]  Radu Grosu,et al.  Designing Worm-inspired Neural Networks for Interpretable Robotic Control , 2019, 2019 International Conference on Robotics and Automation (ICRA).

[76]  Tiziana Margaria,et al.  Tools and algorithms for the construction and analysis of systems: a special issue for TACAS 2017 , 2001, International Journal on Software Tools for Technology Transfer.

[77]  Hajime Asama,et al.  Hypersolvers: Toward Fast Continuous-Depth Models , 2020, NeurIPS.

[78]  Radu Grosu,et al.  Liquid Time-constant Networks , 2020, AAAI.

[79]  B. Shubert A Sequential Method Seeking the Global Maximum of a Function , 1972 .

[80]  Pushmeet Kohli,et al.  Improved Branch and Bound for Neural Network Verification via Lagrangian Decomposition , 2021, ArXiv.

[81]  Radu Grosu,et al.  Model-based versus Model-free Deep Reinforcement Learning for Autonomous Racing Cars , 2021, ArXiv.

[82]  Pushmeet Kohli,et al.  Lagrangian Decomposition for Neural Network Verification , 2020, UAI.

[83]  Radu Grosu,et al.  A Natural Lottery Ticket Winner: Reinforcement Learning with Ordinary Neural Circuits , 2020, ICML.

[84]  Rüdiger Ehlers,et al.  Formal Verification of Piece-Wise Linear Feed-Forward Neural Networks , 2017, ATVA.

[85]  A. Neumaier Complete search in continuous global optimization and constraint satisfaction , 2004, Acta Numerica.

[86]  Mahesh Viswanathan,et al.  C2E2: A Verification Tool for Stateflow Models , 2015, TACAS.

[87]  Radu Grosu,et al.  A Machine Learning Suite for Machine Components' Health-Monitoring , 2019, AAAI.

[88]  Radu Grosu,et al.  A generative neural network model for the quality prediction of work in progress products , 2019, Appl. Soft Comput..