Selecting accurate, robust, and minimal feedforward neural networks

Accuracy, robustness, and minimality are fundamental issues in system-level design. Such properties are generally associated with constraints limiting the feasible model space. The paper focuses on the optimal selection of feedforward neural networks under the accuracy, robustness, and minimality constraints. Model selection, with respect to accuracy, can be carried out within the theoretical framework delineated by the final prediction error (FPE), generalization error estimate (GEN), general prediction error (GPE) and network information criterion (NIC) or cross-validation-based techniques. Robustness is an appealing feature since a robust application provides a graceful degradation in performance once affected by perturbations in its structural parameters (e.g., associated with faults or finite precision representations). Minimality is related to model selection and attempts to reduce the computational load of the solution (with also silicon area and power consumption reduction in a digital implementation). A novel sensitivity analysis derived by the FPE selection criterion is suggested in the paper to quantify the relationship between performance loss and robustness; based on the definition of weak and acute perturbations, we introduce two criteria for estimating the robustness degree of a neural network. Finally, by ranking the features of the obtained models we identify the best constrained neural network.

[1]  Lennart Ljung,et al.  System Identification: Theory for the User , 1987 .

[2]  Jan Larsen,et al.  A generalization error estimate for nonlinear systems , 1992, Neural Networks for Signal Processing II Proceedings of the 1992 IEEE Workshop.

[3]  Yann LeCun,et al.  Optimal Brain Damage , 1989, NIPS.

[4]  Dhananjay S. Phatak,et al.  Complete and partial fault tolerance of feedforward neural nets , 1995, IEEE Trans. Neural Networks.

[5]  Sina Balkir,et al.  ANNSyS (an analog neural network synthesis system) , 1997, Proceedings of International Conference on Neural Networks (ICNN'97).

[6]  John E. Moody,et al.  Fast Pruning Using Principal Components , 1993, NIPS.

[7]  Petre Stoica,et al.  Decentralized Control , 2018, The Control Systems Handbook.

[8]  Rajesh Gupta,et al.  Hardware/software co-design , 1996, Proc. IEEE.

[9]  Cesare Alippi,et al.  FPE-based criteria to dimension feedforward neural topologies , 1999 .

[10]  G. Stewart,et al.  Matrix Perturbation Theory , 1990 .

[11]  Yoichi Koyanagi,et al.  Fault-Tolerant Design of Neural Networks for Solving Optimization Problems , 1996, IEEE Trans. Computers.

[12]  A. Poncet,et al.  Asymptotic probability density of the generalization error , 1996, Proceedings of International Workshop on Neural Networks for Identification, Control, Robotics and Signal/Image Processing.

[13]  D. Rumelhart,et al.  The effective dimension of the space of hidden units , 1991, [Proceedings] 1991 IEEE International Joint Conference on Neural Networks.

[14]  Cesare Alippi,et al.  Accuracy vs. Precision in Digital VLSI Architectures for Signal Processing , 1998, IEEE Trans. Computers.

[15]  Charles R. Johnson,et al.  Matrix analysis , 1985, Statistical Inference for Engineers and Data Scientists.

[16]  John Schewel Configurable Computing: Technology and Applications , 1998 .

[17]  M.H. Hassoun,et al.  Fundamentals of Artificial Neural Networks , 1996, Proceedings of the IEEE.

[18]  A. F. Murray,et al.  Fault tolerance via weight noise in analog VLSI implementations of MLPs-a case study with EPSILON , 1998 .

[19]  John E. Moody,et al.  The Effective Number of Parameters: An Analysis of Generalization and Regularization in Nonlinear Learning Systems , 1991, NIPS.

[20]  M. Buckingham Noise in electronic devices and systems , 1983 .

[21]  Anders Krogh,et al.  A Simple Weight Decay Can Improve Generalization , 1991, NIPS.

[22]  Heekuck Oh,et al.  Neural Networks for Pattern Recognition , 1993, Adv. Comput..

[23]  Tomaso A. Poggio,et al.  Regularization Theory and Neural Networks Architectures , 1995, Neural Computation.

[24]  Isabelle Guyon,et al.  Structural Risk Minimization for Character Recognition , 1991, NIPS.

[25]  Jenq-Neng Hwang,et al.  Finite precision error analysis of neural network electronic hardware implementations , 1991, IJCNN-91-Seattle International Joint Conference on Neural Networks.

[26]  Babak Hassibi,et al.  Second Order Derivatives for Network Pruning: Optimal Brain Surgeon , 1992, NIPS.

[27]  Günhan Dündar,et al.  The effects of quantization on multilayer neural networks , 1995, IEEE Trans. Neural Networks.

[28]  Lars Kai Hansen,et al.  Generalization performance of regularized neural network models , 1994, Proceedings of IEEE Workshop on Neural Networks for Signal Processing.

[29]  Ranga Vemuri,et al.  Configurable Computing: Technology and Applications - Guest Editors' Introduction , 2000, Computer.

[30]  Jenq-Neng Hwang,et al.  Finite Precision Error Analysis of Neural Network Hardware Implementations , 1993, IEEE Trans. Computers.

[31]  Donatella Sciuto Guest Editor's Introduction: Design Tools for Embedded Systems , 2000, IEEE Des. Test Comput..

[32]  Steve W. Piche,et al.  The selection of weight accuracies for Madalines , 1995, IEEE Trans. Neural Networks.

[33]  Alan F. Murray,et al.  Toward Optimally Distributed Computation , 1998, Neural Computation.

[34]  Shun-ichi Amari,et al.  Network information criterion-determining the number of hidden units for an artificial neural network model , 1994, IEEE Trans. Neural Networks.

[35]  Cesare Alippi Randomized Algorithms: A System-Level, Poly-Time Analysis of Robust Computation , 2002, IEEE Trans. Computers.