Priority Neuron: A Resource-Aware Neural Network for Cyber-Physical Systems

Advances in sensing, computation, storage, and actuation technologies have entered cyber-physical systems (CPSs) into the smart era where complex control applications requiring high performance are supported. Neural networks (NNs) models are proposed as a predictive model to be used in model predictive control (MPC) applications. However, the ability to efficiently exploit resource hungry NNs in embedded resource-bound settings is a major challenge. In this paper, we propose priority neuron network (PNN), a resource-aware NNs model that can be reconfigured into smaller subnetworks at runtime. This approach enables a tradeoff between the model’s computation time and accuracy based on available resources. The PNN model is memory efficient since it stores only one set of parameters to account for various subnetwork sizes. We propose a training algorithm that applies regularization techniques to constrain the activation value of neurons and assigns a priority to each one. We consider the neuron’s ordinal number as our priority criteria in that the priority of the neuron is inversely proportional to its ordinal number in the layer. This imposes a relatively sorted order on the activation values. We conduct experiments to employ our PNN as the predictive model of a vehicle in MPC for path tracking. To corroborate the effectiveness of our proposed methodology, we compare it with two state-of-the-art methods for resource-aware NN design. Compared to state-of-the-art work, our approach can cut down the training time by 87% and reduce the memory storage by 75% while achieving similar accuracy. Moreover, we decrease the computation overhead for the model reduction process that searches for neurons below a threshold, from <inline-formula> <tex-math notation="LaTeX">$ {O(n)}$ </tex-math></inline-formula> to <inline-formula> <tex-math notation="LaTeX">$ {O}$ </tex-math></inline-formula>(log<inline-formula> <tex-math notation="LaTeX">$ {n}$ </tex-math></inline-formula>).

[1]  Rüdiger W. Brause,et al.  The Performance of Approximating Ordinary Differential Equations by Neural Nets , 2008, 2008 20th IEEE International Conference on Tools with Artificial Intelligence.

[2]  Yiran Chen,et al.  Learning Structured Sparsity in Deep Neural Networks , 2016, NIPS.

[3]  C. Kelley Solving Nonlinear Equations with Newton's Method , 1987 .

[4]  Song Han,et al.  Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.

[5]  H. Unbehauen,et al.  Nonlinear predictive control with multirate optimisation step lengths , 2005 .

[6]  S. Karsoliya,et al.  Approximating Number of Hidden layer neurons in Multiple Hidden Layer BPNN Architecture , 2012 .

[7]  Yike Guo,et al.  DropNeuron: Simplifying the Structure of Deep Neural Networks , 2016, ArXiv.

[8]  Atsushi Sato,et al.  Layer-Wise Weight Decay for Deep Neural Networks , 2017, PSIVT.

[9]  Paul A. Fishwick,et al.  Feedforward Neural Nets as Models for Time Series Forecasting , 1993, INFORMS J. Comput..

[10]  Mohammad Abdullah Al Faruque,et al.  Eco-Friendly Automotive Climate Control and Navigation System for Electric Vehicles , 2016, 2016 ACM/IEEE 7th International Conference on Cyber-Physical Systems (ICCPS).

[11]  Xin Dong,et al.  Learning to Prune Deep Neural Networks via Layer-wise Optimal Brain Surgeon , 2017, NIPS.

[12]  J S Barlow,et al.  Data-based predictive control with multirate prediction step , 2010, Proceedings of the 2010 American Control Conference.

[13]  Uwe D. Hanebeck,et al.  A model-predictive switching approach to efficient intention recognition , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[14]  John Hauser,et al.  On the stability of receding horizon control with a general terminal cost , 2005, IEEE Transactions on Automatic Control.

[15]  Mohammad Abdullah Al Faruque,et al.  ACQUA: Adaptive and cooperative quality-aware control for automotive cyber-physical systems , 2017, 2017 IEEE/ACM International Conference on Computer-Aided Design (ICCAD).

[16]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[17]  Sungroh Yoon,et al.  Big/little deep neural network for ultra low power inference , 2015, 2015 International Conference on Hardware/Software Codesign and System Synthesis (CODES+ISSS).

[18]  Sherief Reda,et al.  Runtime configurable deep neural networks for energy-accuracy trade-off , 2016, 2016 International Conference on Hardware/Software Codesign and System Synthesis (CODES+ISSS).

[19]  Tao Zhang,et al.  Model Compression and Acceleration for Deep Neural Networks: The Principles, Progress, and Challenges , 2018, IEEE Signal Processing Magazine.

[20]  Rui Peng,et al.  Network Trimming: A Data-Driven Neuron Pruning Approach towards Efficient Deep Architectures , 2016, ArXiv.

[21]  Xiangyu Zhang,et al.  Channel Pruning for Accelerating Very Deep Neural Networks , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[22]  Wei He,et al.  Adaptive Neural Network Control of an Uncertain Robot With Full-State Constraints , 2016, IEEE Transactions on Cybernetics.

[23]  Mohammad Abdullah Al Faruque,et al.  Design and Analysis of Battery-Aware Automotive Climate Control for Electric Vehicles , 2018, ACM Trans. Embed. Comput. Syst..

[24]  Yong Wang,et al.  Improved mispronunciation detection with deep neural network trained acoustic models and transfer learning based logistic regression classifiers , 2015, Speech Commun..

[25]  Magnus Egerstedt,et al.  Adaptive time horizon optimization in model predictive control , 2011, Proceedings of the 2011 American Control Conference.

[26]  Song Han,et al.  Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.

[27]  Tao Zhang,et al.  A Survey of Model Compression and Acceleration for Deep Neural Networks , 2017, ArXiv.

[28]  Moritz Diehl,et al.  ACADO toolkit—An open‐source framework for automatic control and dynamic optimization , 2011 .

[29]  Tony Givargis,et al.  Hybrid state machine model for fast model predictive control: Application to path tracking , 2017, 2017 IEEE/ACM International Conference on Computer-Aided Design (ICCAD).

[30]  Franco Cicirelli,et al.  Model continuity in cyber-physical systems: A control-centered methodology based on agents , 2017, Simul. Model. Pract. Theory.

[31]  Marko Bacic,et al.  Model predictive control , 2003 .

[32]  Mohammad Abdullah Al Faruque,et al.  Battery lifetime-aware automotive climate control for Electric Vehicles , 2015, 2015 52nd ACM/EDAC/IEEE Design Automation Conference (DAC).

[33]  Hee-Jun Kang,et al.  Neural network-based adaptive tracking control of mobile robots in the presence of wheel slip and external disturbance force , 2016, Neurocomputing.

[34]  Song Han,et al.  EIE: Efficient Inference Engine on Compressed Deep Neural Network , 2016, 2016 ACM/IEEE 43rd Annual International Symposium on Computer Architecture (ISCA).

[35]  Dimitri P. Solomatine,et al.  Data-Driven Modelling: Concepts, Approaches and Experiences , 2009 .

[36]  Yixin Chen,et al.  Compressing Neural Networks with the Hashing Trick , 2015, ICML.

[37]  Yuan Yu,et al.  TensorFlow: A system for large-scale machine learning , 2016, OSDI.

[38]  Ricardo G. Sanfelice,et al.  Computationally aware control of autonomous vehicles: a hybrid model predictive control approach , 2015, Autonomous Robots.

[39]  Tony Givargis,et al.  HES machine: Harmonic equivalent state machine modeling for cyber-physical systems , 2017, 2017 IEEE International High Level Design Validation and Test Workshop (HLDVT).

[40]  Robert Jenssen,et al.  Recurrent Neural Networks for Short-Term Load Forecasting , 2017, SpringerBriefs in Computer Science.

[41]  Dirk Roose,et al.  Coping with complexity : model reduction and data analysis , 2011 .

[42]  Angelika Steger,et al.  Fast-Slow Recurrent Neural Networks , 2017, NIPS.

[43]  Vivienne Sze,et al.  Designing Energy-Efficient Convolutional Neural Networks Using Energy-Aware Pruning , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[44]  Bernard Girau,et al.  Fault and Error Tolerance in Neural Networks: A Review , 2017, IEEE Access.