Probabilistic Guarantees for Safe Deep Reinforcement Learning
暂无分享,去创建一个
[1] Alessandro Abate,et al. FAUST 2 : Formal Abstractions of Uncountable-STate STochastic Processes , 2014, TACAS.
[2] Antonin Guttman,et al. R-trees: a dynamic index structure for spatial searching , 1984, SIGMOD '84.
[3] Swarat Chaudhuri,et al. AI2: Safety and Robustness Certification of Neural Networks with Abstract Interpretation , 2018, 2018 IEEE Symposium on Security and Privacy (SP).
[4] Ufuk Topcu,et al. Verifiable RNN-Based Policies for POMDPs Under Temporal Logic Constraints , 2020, IJCAI.
[5] Sebastian Junges,et al. Safety-Constrained Reinforcement Learning for MDPs , 2015, TACAS.
[6] Michael Schapira,et al. Verifying Deep-RL-Driven Systems , 2019, NetAI@SIGCOMM.
[7] Hans-Peter Kriegel,et al. The R*-tree: an efficient and robust access method for points and rectangles , 1990, SIGMOD '90.
[8] J. Kemeny,et al. Denumerable Markov chains , 1969 .
[9] Min Wu,et al. Safety Verification of Deep Neural Networks , 2016, CAV.
[10] Mykel J. Kochenderfer,et al. Reluplex: An Efficient SMT Solver for Verifying Deep Neural Networks , 2017, CAV.
[11] Sven Schewe,et al. Omega-Regular Objectives in Model-Free Reinforcement Learning , 2018, TACAS.
[12] Mohan M. Trivedi,et al. Looking at Humans in the Age of Self-Driving and Highly Automated Vehicles , 2016, IEEE Transactions on Intelligent Vehicles.
[13] Rüdiger Ehlers,et al. Formal Verification of Piece-Wise Linear Feed-Forward Neural Networks , 2017, ATVA.
[14] Marta Z. Kwiatkowska,et al. A game-based abstraction-refinement framework for Markov decision processes , 2010, Formal Methods Syst. Des..
[15] Pushmeet Kohli,et al. A Unified View of Piecewise Linear Neural Network Verification , 2017, NeurIPS.
[16] Luca Cardelli,et al. Statistical Guarantees for the Robustness of Bayesian Neural Networks , 2019, IJCAI.
[17] Krishnendu Chatterjee,et al. Verification of Markov Decision Processes Using Learning Algorithms , 2014, ATVA.
[18] Tom Schaul,et al. Prioritized Experience Replay , 2015, ICLR.
[19] Junfeng Yang,et al. Formal Security Analysis of Neural Networks using Symbolic Intervals , 2018, USENIX Security Symposium.
[20] Amnon Shashua,et al. Safe, Multi-Agent, Reinforcement Learning for Autonomous Driving , 2016, ArXiv.
[21] Nils Jansen,et al. Counterexample-Guided Strategy Improvement for POMDPs Using Recurrent Neural Networks , 2019, IJCAI.
[22] Marta Z. Kwiatkowska,et al. PRISM 4.0: Verification of Probabilistic Real-Time Systems , 2011, CAV.
[23] Wojciech Zaremba,et al. OpenAI Gym , 2016, ArXiv.
[24] Xiaowei Huang,et al. Reachability Analysis of Deep Neural Networks with Provable Guarantees , 2018, IJCAI.
[25] Giorgos B. Stamou,et al. Improving Fuel Economy with LSTM Networks and Reinforcement Learning , 2018, ICANN.
[26] Armando Solar-Lezama,et al. Verifiable Reinforcement Learning via Policy Extraction , 2018, NeurIPS.
[27] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[28] Marlos C. Machado,et al. Revisiting the Arcade Learning Environment: Evaluation Protocols and Open Problems for General Agents , 2017, J. Artif. Intell. Res..
[29] Javier García,et al. A comprehensive survey on safe reinforcement learning , 2015, J. Mach. Learn. Res..
[30] Isil Dillig,et al. Optimization and abstraction: a synergistic approach for analyzing neural network robustness , 2019, PLDI.
[31] Ufuk Topcu,et al. Probably Approximately Correct MDP Learning and Control With Temporal Logic Constraints , 2014, Robotics: Science and Systems.
[32] Calin Belta,et al. Formal Verification and Synthesis for Discrete-Time Stochastic Systems , 2015, IEEE Trans. Autom. Control..