Generalized Entropy Loss Function in Neural Network: Variable's Importance and Sensitivity Analysis

Artificial neural networks are powerful tools for data analysis and are particularly suitable for modelling relationships between variables for best prediction of an outcome. A large number of error functions have been proposed in the literature to achieve a better predictive power of a neural network. Only a few works employ Tsallis statistics, although the method itself has been successfully applied in other machine learning techniques. This paper undertakes the effort to examine various characteristics of the \( q \)-generalized function based on Tsallis entropy as an alternative loss measure in neural networks. To achieve this goal, we will explore various methods that can be used to interpret supervised neural network models. These methods can be used to visualize the model using the neural network interpretation diagram, assess the importance of variables by disaggregating the model’s weights (Olden’s and Garson’s algorithms) and perform a sensitivity analysis of model responses to input variables (Lek’s profile).

[1]  Bin Huang,et al.  Opening the black box of neural networks: methods for interpreting neural network models in clinical applications. , 2018, Annals of translational medicine.

[2]  Marcus W Beck,et al.  NeuralNetTools: Visualization and Analysis Tools for Neural Networks. , 2018, Journal of statistical software.

[3]  A. T. C. Goh,et al.  Back-propagation neural networks for modeling complex systems , 1995, Artif. Intell. Eng..

[4]  Max Kuhn,et al.  Building Predictive Models in R Using the caret Package , 2008 .

[5]  C. Tsallis Introduction to Nonextensive Statistical Mechanics: Approaching a Complex World , 2009 .

[6]  Andries Petrus Engelbrecht,et al.  Determining the Significance of Input Parameters using Sensitivity Analysis , 1995, IWANN.

[7]  Russell G. Death,et al.  An accurate comparison of methods for quantifying variable importance in artificial neural networks using simulated data , 2004 .

[8]  Krzysztof Gajowniczek,et al.  Simulation Study on the Application of the Generalized Entropy Concept in Artificial Neural Networks , 2018, Entropy.

[9]  Krzysztof Gajowniczek,et al.  Do Customers Choose Proper Tariff? Empirical Analysis Based on Polish Data Using Unsupervised Techniques , 2018 .

[10]  Krzysztof Gajowniczek,et al.  Entropy Based Trees to Support Decision Making for Customer Churn Management , 2016 .

[11]  I. Dimopoulos,et al.  Application of neural networks to modelling nonlinear relationships in ecology , 1996 .

[12]  Yurong Liu,et al.  A survey of deep neural network architectures and their applications , 2017, Neurocomputing.

[13]  Krzysztof Gajowniczek,et al.  Q-Entropy Approach to Selecting High Income Households , 2015 .

[14]  Jacek M. Zurada,et al.  Sensitivity analysis for minimization of input data dimension for feedforward neural network , 1994, Proceedings of IEEE International Symposium on Circuits and Systems - ISCAS '94.

[15]  Yannis Dimopoulos,et al.  Use of some sensitivity criteria for choosing networks with good generalization ability , 1995, Neural Processing Letters.

[16]  Sang-Hyo Kim,et al.  Structural Monitoring System Based on Sensitivity Analysis and a Neural Network , 2000 .

[17]  Bo-Christer Björk Intelligent front-ends and product models , 1991, Artif. Intell. Eng..

[18]  Claude E. Shannon,et al.  The mathematical theory of communication , 1950 .

[19]  Hermann Ney,et al.  Cross-entropy vs. squared error training: a theoretical and experimental comparison , 2013, INTERSPEECH.

[20]  G. David Garson,et al.  Interpreting neural-network connection weights , 1991 .

[21]  Krzysztof Gajowniczek,et al.  Short term electricity forecasting based on user behavior from individual smart meter data , 2015, J. Intell. Fuzzy Syst..

[22]  C. E. SHANNON,et al.  A mathematical theory of communication , 1948, MOCO.

[23]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.