Deep reinforcement learning achieves multifunctional morphing airfoil control

Smooth camber morphing aircraft offer increased control authority and improved aerodynamic efficiency. Smart material actuators have become a popular driving force for shape changes, capable of adhering to weight and size constraints and allowing for simplicity in mechanical design. As a step towards creating uncrewed aerial vehicles (UAVs) capable of autonomously responding to flow conditions, this work examines a multifunctional morphing airfoil’s ability to follow commands in various flows. We integrated an airfoil with a morphing trailing edge consisting of an antagonistic pair of macro fiber composites (MFCs), serving as both skin and actuator, and internal piezoelectric flex sensors to form a closed loop composite system. Closed loop feedback control is necessary to accurately follow deflection commands due to the hysteretic behavior of MFCs. Here we used a deep reinforcement learning algorithm, Proximal Policy Optimization, to control the morphing airfoil. Two neural controllers were trained in a simulation developed through time series modeling on long short-term memory recurrent neural networks. The learned controllers were then tested on the composite wing using two state inference methods in still air and in a wind tunnel at various flow speeds. We compared the performance of our neural controllers to one using traditional position-derivative feedback control methods. Our experimental results validate that the autonomous neural controllers were faster and more accurate than traditional methods. This research shows that deep learning methods can overcome common obstacles for achieving sufficient modeling and control when implementing smart composite actuators in an autonomous aerospace environment.

[1]  D. Inman,et al.  Si-based self-programming neuromorphic integrated circuits for intelligent morphing wings , 2022, Journal of Composite Materials.

[2]  D. Inman,et al.  Autonomous Learning in a Pseudo-Episodic Physical Environment , 2022, J. Intell. Robotic Syst..

[3]  Doina Precup,et al.  Reward is enough , 2021, Artif. Intell..

[4]  Zhongcai Pei,et al.  Design, Modeling and Control of a Novel Morphing Quadrotor , 2021, IEEE Robotics and Automation Letters.

[5]  Jakub W. Pachocki,et al.  Dota 2 with Large Scale Deep Reinforcement Learning , 2019, ArXiv.

[6]  Gang Chen,et al.  Morphing control of a new bionic morphing UAV with deep reinforcement learning , 2019, Aerospace Science and Technology.

[7]  Onur Avci,et al.  1D Convolutional Neural Networks and Applications: A Survey , 2019, Mechanical Systems and Signal Processing.

[8]  Gabriel Dulac-Arnold,et al.  Challenges of Real-World Reinforcement Learning , 2019, ArXiv.

[9]  Philippe Chatelain,et al.  Loads Alleviation on an Airfoil via Reinforcement Learning , 2019, AIAA Scitech 2019 Forum.

[10]  Demis Hassabis,et al.  A general reinforcement learning algorithm that masters chess, shogi, and Go through self-play , 2018, Science.

[11]  Jakub W. Pachocki,et al.  Learning dexterous in-hand manipulation , 2018, Int. J. Robotics Res..

[12]  Azer Bestavros,et al.  Reinforcement Learning for UAV Attitude Control , 2018, ACM Trans. Cyber Phys. Syst..

[13]  Daniel J Inman,et al.  A tale of two tails: developing an avian inspired morphing actuator for yaw control and stability , 2018, Bioinspiration & biomimetics.

[14]  Darren J. Hartl,et al.  Control of Morphing Wing Shapes with Deep Reinforcement Learning , 2018 .

[15]  Sergey Levine,et al.  Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor , 2018, ICML.

[16]  Alec Radford,et al.  Proximal Policy Optimization Algorithms , 2017, ArXiv.

[17]  Roland Siegwart,et al.  Control of a Quadrotor With Reinforcement Learning , 2017, IEEE Robotics and Automation Letters.

[18]  Daniel J. Inman,et al.  Stall Recovery of a Morphing Wing via Extended Nonlinear Lifting-Line Theory , 2017 .

[19]  Ming Liu,et al.  Virtual-to-real deep reinforcement learning: Continuous control of mobile robots for mapless navigation , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[20]  Yanju Liu,et al.  Morphing aircraft based on smart materials and structures: A state-of-the-art review , 2016 .

[21]  Demis Hassabis,et al.  Mastering the game of Go with deep neural networks and tree search , 2016, Nature.

[22]  Tianqi Chen,et al.  Empirical Evaluation of Rectified Activations in Convolutional Network , 2015, ArXiv.

[23]  Andres F. Arrieta,et al.  Planform, aero-structural, and flight control optimization for tailless morphing aircraft , 2015, Smart Structures.

[24]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[25]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[26]  Onur Bilgen,et al.  Wind tunnel testing of the fish bone active camber morphing concept , 2014 .

[27]  Alex Graves,et al.  Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.

[28]  Daniel J. Inman,et al.  Experimental testing of spanwise morphing trailing edge concept , 2013, Smart Structures.

[29]  Yuansheng Chen,et al.  Modeling hysteresis and creep behavior of macrofiber composite–based piezoelectric bimorph actuator , 2013 .

[30]  P. Goulart,et al.  Robust Gust Alleviation and Stabilization of Very Flexible Aircraft , 2013 .

[31]  Ephrahim Garcia,et al.  Morphing unmanned aerial vehicles , 2011 .

[32]  Daniel J. Inman,et al.  A Review of Morphing Aircraft , 2011 .

[33]  John Valasek,et al.  Reinforcement Learning of a Morphing Airfoil-Policy and Discrete Learning Analysis , 2008, J. Aerosp. Comput. Inf. Commun..

[34]  Monish D. Tandale,et al.  Improved Adaptive–Reinforcement Learning Control for Morphing Unmanned Air Vehicles , 2008, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[35]  Sridhar Kota,et al.  Flight testing of Mission Adaptive Compliant Wing , 2007 .

[36]  Monish D. Tandale,et al.  A Reinforcement Learning - Adaptive Control Architecture for Morphing , 2004, J. Aerosp. Comput. Inf. Commun..

[37]  Jayanth N. Kudva,et al.  Morphing aircraft concepts, classifications, and challenges , 2004, SPIE Smart Structures and Materials + Nondestructive Evaluation and Health Monitoring.

[38]  W. Keats Wilkie,et al.  Method of Fabricating NASA-Standard Macro-Fiber Composite Piezoelectric Actuators , 2003 .

[39]  Paul H. Mirick,et al.  Low-cost piezocomposite actuator for structural control applications , 2000, Smart Structures.

[40]  G. Haertling Ferroelectric ceramics : History and technology , 1999 .

[41]  Christopher Edwards,et al.  Anti-windup and bumpless-transfer schemes , 1998, Autom..

[42]  S. Hochreiter,et al.  Long Short-Term Memory , 1997, Neural Computation.

[43]  Daniel Svozil,et al.  Introduction to multi-layer feed-forward neural networks , 1997 .

[44]  William W. Gilbert,et al.  Mission Adaptive Wing System for Tactical Aircraft , 1981 .

[45]  J. G. Ziegler,et al.  Optimum Settings for Automatic Controllers , 1942, Journal of Fluids Engineering.

[46]  Lucian Busoniu,et al.  Reinforcement learning for control: Performance, stability, and deep approximators , 2018, Annu. Rev. Control..

[47]  Alexander M. Pankonien Smart Material Wing Morphing for Unmanned Aerial Vehicles. , 2015 .

[48]  Daniel Dewey,et al.  Reinforcement Learning and the Reward Engineering Principle , 2014, AAAI Spring Symposia.

[49]  Daniel J. Inman,et al.  Novel, Bidirectional, Variable-Camber Airfoil via Macro-Fiber Composite Actuators , 2010 .

[50]  Brian Sanders,et al.  Aerodynamic and Aeroelastic Characteristics of Wings with Conformal Control Surfaces for Morphing Aircraft , 2003 .