暂无分享,去创建一个
Michael Oschwald | Kai Dresia | Jan Christian Deeken | Gunther Waxenegger-Wilfing | M. Oschwald | J. Deeken | Günther Waxenegger-Wilfing | Kai Dresia
[1] N. Girard,et al. PROMETHEUS , A LOX / LCH 4 REUSABLE ROCKET ENGINE , 2017 .
[2] Julien Marzat,et al. A survey of automatic control methods for liquid-propellant rocket engines , 2019, Progress in Aerospace Sciences.
[3] Javad Lavaei,et al. Stability-Certified Reinforcement Learning: A Control-Theoretic Perspective , 2018, IEEE Access.
[4] Julien Marzat,et al. Derivation and Analysis of a State-Space Model for Transient Control of Liquid-Propellant Rocket Engines , 2018, 2018 9th International Conference on Mechanical and Aerospace Engineering (ICMAE).
[5] Hussein A. Abbass,et al. Hierarchical Deep Reinforcement Learning for Continuous Action Control , 2018, IEEE Transactions on Neural Networks and Learning Systems.
[6] Julien Marzat,et al. An MPC Approach to Transient Control of Liquid-Propellant Rocket Engines , 2019, IFAC-PapersOnLine.
[7] Roberto Furfaro,et al. Deep reinforcement learning for six degree-of-freedom planetary landing , 2020 .
[8] Takuya Akiba,et al. Optuna: A Next-generation Hyperparameter Optimization Framework , 2019, KDD.
[9] Francois Lassoudiere,et al. Progress of the development of an all-electric control system of a rocket engine , 2012 .
[10] Jeffrey L. Musgrave,et al. Overview of rocket engine control , 2008 .
[11] Frank L. Lewis,et al. Optimal and Autonomous Control Using Reinforcement Learning: A Survey , 2018, IEEE Transactions on Neural Networks and Learning Systems.
[12] Ed Nemeth,et al. Reusable rocket engine intelligent control system framework design, phase 2 , 1991 .
[13] Sergey Levine,et al. Deep reinforcement learning for robotic manipulation with asynchronous off-policy updates , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).
[14] Sergey Levine,et al. Reinforcement Learning with Deep Energy-Based Policies , 2017, ICML.
[15] D. Bertsekas. Reinforcement Learning and Optimal ControlA Selective Overview , 2018 .
[16] Frank L. Lewis,et al. Dynamic Multiobjective Control for Continuous-Time Systems Using Reinforcement Learning , 2019, IEEE Transactions on Automatic Control.
[17] Yuval Tassa,et al. Continuous control with deep reinforcement learning , 2015, ICLR.
[18] Wojciech Zaremba,et al. Domain randomization for transferring deep neural networks from simulation to the real world , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[19] R. B. Gopaluni,et al. Deep reinforcement learning approaches for process control , 2017, 2017 6th International Symposium on Advanced Control of Industrial Processes (AdCONIP).
[20] Jan Deeken,et al. Numerically Efficient Fatigue Life Prediction of Rocket Combustion Chambers using Artificial Neural Networks , 2019 .
[21] Marc Parizeau,et al. DEAP: evolutionary algorithms made easy , 2012, J. Mach. Learn. Res..
[22] Amir Hussain,et al. Applications of Deep Learning and Reinforcement Learning to Biological Data , 2017, IEEE Transactions on Neural Networks and Learning Systems.
[23] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[24] Lucian Busoniu,et al. Reinforcement learning for control: Performance, stability, and deep approximators , 2018, Annu. Rev. Control..
[25] Erik-Jan Van Kampen,et al. Online Adaptive Incremental Reinforcement Learning Flight Control for a CS-25 Class Aircraft , 2020, AIAA Scitech 2020 Forum.
[26] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[27] L. Meland,et al. History of the Titan liquid rocket engines , 1989 .
[28] Antoine Iffly,et al. Performance model of the Vulcain Ariane 5 main engine , 1999 .
[29] Günther Waxenegger-Wilfing,et al. Heat Transfer Prediction for Methane in Regenerative Cooling Channels with Neural Networks , 2019, Journal of Thermophysics and Heat Transfer.
[30] Thomas Frohlich,et al. LOX/METHANE AND LOX/KEROSENE HIGH THRUST ENGINE TRADE-OFF , 2001 .
[31] Michael E. Hampson. Reusable Rocket Engine Turbopump Condition Monitoring , 1984 .
[32] Pieter Abbeel,et al. Constrained Policy Optimization , 2017, ICML.
[33] Xuesong Wang,et al. Approximate Policy-Based Accelerated Deep Reinforcement Learning , 2020, IEEE Transactions on Neural Networks and Learning Systems.
[34] R. Mazo. On the theory of brownian motion , 1973 .
[35] N. Munro,et al. PID controllers: recent tuning methods and design to specification , 2002 .
[36] Daniel E. Paxson,et al. A demonstration of an intelligent control system for a reusable rocket engine , 1992 .
[37] Herke van Hoof,et al. Addressing Function Approximation Error in Actor-Critic Methods , 2018, ICML.
[38] R. M. Ryan,et al. Effects of geometry and materials on low cycle fatigue life of turbine blades in LOX/hydrogen rocket engines , 1986 .
[39] Sergey Levine,et al. Trust Region Policy Optimization , 2015, ICML.
[40] R. J. Roelke. Miscellaneous losses. [tip clearance and disk friction] , 1973 .