Machine learning facilitated business intelligence (Part I)

The purpose of this paper is to conduct a comprehensive review of the noteworthy contributions made in the area of the Feedforward neural network (FNN) to improve its generalization performance and convergence rate (learning speed); to identify new research directions that will help researchers to design new, simple and efficient algorithms and users to implement optimal designed FNNs for solving complex problems; and to explore the wide applications of the reviewed FNN algorithms in solving real-world management, engineering and health sciences problems and demonstrate the advantages of these algorithms in enhancing decision making for practical operations.,The FNN has gained much popularity during the last three decades. Therefore, the authors have focused on algorithms proposed during the last three decades. The selected databases were searched with popular keywords: “generalization performance,” “learning rate,” “overfitting” and “fixed and cascade architecture.” Combinations of the keywords were also used to get more relevant results. Duplicated articles in the databases, non-English language, and matched keywords but out of scope, were discarded.,The authors studied a total of 80 articles and classified them into six categories according to the nature of the algorithms proposed in these articles which aimed at improving the generalization performance and convergence rate of FNNs. To review and discuss all the six categories would result in the paper being too long. Therefore, the authors further divided the six categories into two parts (i.e. Part I and Part II). The current paper, Part I, investigates two categories that focus on learning algorithms (i.e. gradient learning algorithms for network training and gradient-free learning algorithms). Furthermore, the remaining four categories which mainly explore optimization techniques are reviewed in Part II (i.e. optimization algorithms for learning rate, bias and variance (underfitting and overfitting) minimization algorithms, constructive topology neural networks and metaheuristic search algorithms). For the sake of simplicity, the paper entitled “Machine learning facilitated business intelligence (Part II): Neural networks optimization techniques and applications” is referred to as Part II. This results in a division of 80 articles into 38 and 42 for Part I and Part II, respectively. After discussing the FNN algorithms with their technical merits and limitations, along with real-world management, engineering and health sciences applications for each individual category, the authors suggest seven (three in Part I and other four in Part II) new future directions which can contribute to strengthening the literature.,The FNN contributions are numerous and cannot be covered in a single study. The authors remain focused on learning algorithms and optimization techniques, along with their application to real-world problems, proposing to improve the generalization performance and convergence rate of FNNs with the characteristics of computing optimal hyperparameters, connection weights, hidden units, selecting an appropriate network architecture rather than trial and error approaches and avoiding overfitting.,This study will help researchers and practitioners to deeply understand the existing algorithms merits of FNNs with limitations, research gaps, application areas and changes in research studies in the last three decades. Moreover, the user, after having in-depth knowledge by understanding the applications of algorithms in the real world, may apply appropriate FNN algorithms to get optimal results in the shortest possible time, with less effort, for their specific application area problems.,The existing literature surveys are limited in scope due to comparative study of the algorithms, studying algorithms application areas and focusing on specific techniques. This implies that the existing surveys are focused on studying some specific algorithms or their applications (e.g. pruning algorithms, constructive algorithms, etc.). In this work, the authors propose a comprehensive review of different categories, along with their real-world applications, that may affect FNN generalization performance and convergence rate. This makes the classification scheme novel and significant.

[1]  Ali Dag,et al.  A comparative data analytic approach to construct a risk trade-off for cardiac patients' re-admissions , 2019, Ind. Manag. Data Syst..

[2]  T. V. Geetha,et al.  Indian Logic Ontology based Automatic Query Refinement , 2008 .

[3]  Guang-Bin Huang,et al.  Extreme Learning Machine for Multilayer Perceptron , 2016, IEEE Transactions on Neural Networks and Learning Systems.

[4]  Franco Scarselli,et al.  On the Complexity of Neural Network Classifiers: A Comparison Between Shallow and Deep Architectures , 2014, IEEE Transactions on Neural Networks and Learning Systems.

[5]  Akhil Kumar,et al.  An empirical comparison of neural network and logistic regression models , 1995 .

[6]  Bernard Widrow,et al.  Improving the learning speed of 2-layer neural networks by choosing initial values of the adaptive weights , 1990, 1990 IJCNN International Joint Conference on Neural Networks.

[7]  Chee Kheong Siew,et al.  Universal Approximation using Incremental Constructive Feedforward Networks with Random Hidden Nodes , 2006, IEEE Transactions on Neural Networks.

[8]  Xindong Wu,et al.  Support vector machines based on K-means clustering for real-time business intelligence systems , 2005, Int. J. Bus. Intell. Data Min..

[9]  Robert K. L. Gay,et al.  Error Minimized Extreme Learning Machine With Growth of Hidden Nodes and Incremental Learning , 2009, IEEE Transactions on Neural Networks.

[10]  Siriporn Supratid,et al.  Thailand tourism forecasting based on a hybrid of discrete wavelet decomposition and NARX neural network , 2016, Ind. Manag. Data Syst..

[11]  Hao Yu,et al.  Selection of Proper Neural Network Sizes and Architectures—A Comparative Study , 2012, IEEE Transactions on Industrial Informatics.

[12]  Guoqiang Peter Zhang,et al.  Neural networks for classification: a survey , 2000, IEEE Trans. Syst. Man Cybern. Part C.

[13]  Iasonas Kokkinos,et al.  DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[14]  Chee Kheong Siew,et al.  Extreme learning machine: Theory and applications , 2006, Neurocomputing.

[15]  J V Tu,et al.  Advantages and disadvantages of using artificial neural networks versus logistic regression for predicting medical outcomes. , 1996, Journal of clinical epidemiology.

[16]  Kurt Hornik,et al.  Multilayer feedforward networks are universal approximators , 1989, Neural Networks.

[17]  Guang-Bin Huang,et al.  Trends in extreme learning machines: A review , 2015, Neural Networks.

[18]  Bernard Widrow,et al.  The No-Prop algorithm: A new learning algorithm for multilayer neural networks , 2013, Neural Networks.

[19]  Yiqiang Chen,et al.  Weighted extreme learning machine for imbalance learning , 2013, Neurocomputing.

[20]  Chalavadi Krishna Mohan,et al.  Human action recognition using genetic algorithms and convolutional neural networks , 2016, Pattern Recognit..

[21]  Donald F. Specht,et al.  Probabilistic neural networks , 1990, Neural Networks.

[22]  Eleonora Bottani,et al.  Modelling wholesale distribution operations: an artificial intelligence framework , 2019, Ind. Manag. Data Syst..

[23]  Gaige Wang,et al.  Self-adaptive extreme learning machine , 2015, Neural Computing and Applications.

[24]  Robert F. Stengel,et al.  Smooth function approximation using neural networks , 2005, IEEE Transactions on Neural Networks.

[25]  Michal Tkác,et al.  Artificial neural networks in business: Two decades of research , 2016, Appl. Soft Comput..

[26]  Lei Chen,et al.  Enhanced random search based incremental extreme learning machine , 2008, Neurocomputing.

[27]  Hao Yu,et al.  Neural Network Learning Without Backpropagation , 2010, IEEE Transactions on Neural Networks.

[28]  Matthew D. Zeiler ADADELTA: An Adaptive Learning Rate Method , 2012, ArXiv.

[29]  Tjalling J. Ypma,et al.  Historical Development of the Newton-Raphson Method , 1995, SIAM Rev..

[30]  Donald F. Specht,et al.  A general regression neural network , 1991, IEEE Trans. Neural Networks.

[31]  Hecht-Nielsen Theory of the backpropagation neural network , 1989 .

[32]  Adrian S. Lewis,et al.  Nonsmooth optimization via quasi-Newton methods , 2012, Mathematical Programming.

[33]  Xiaoou Tang,et al.  Image Super-Resolution Using Deep Convolutional Networks , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[34]  T. Choi,et al.  Selling green first or not? A Bayesian analysis with service levels and environmental impact considerations in the Big Data Era , 2017, Technological Forecasting and Social Change.

[35]  Sang M. Lee,et al.  Text classification: neural networks vs support vector machines , 2009, Ind. Manag. Data Syst..

[36]  Ichiro Sakata,et al.  Machine learning approach for finding business partners and building reciprocal relationships , 2012, Expert Syst. Appl..

[37]  Dianhui Wang,et al.  An iterative learning algorithm for feedforward neural networks with random weights , 2016, Inf. Sci..

[38]  Yoichi Hayashi,et al.  Understanding consumer heterogeneity: A business intelligence application of neural networks , 2010, Knowl. Based Syst..

[39]  Yuanqing Xia,et al.  Fault Diagnosis of Tennessee-Eastman Process Using Orthogonal Incremental Extreme Learning Machine Based on Driving Amount , 2018, IEEE Transactions on Cybernetics.

[40]  Li Ying Orthogonal incremental extreme learning machine for regression and multiclass classification , 2014, Neural Computing and Applications.

[41]  Alain Yee-Loong Chong,et al.  Multi-class Twitter sentiment classification with emojis , 2018, Ind. Manag. Data Syst..

[42]  S. Minner,et al.  A review on supply chain contracting with information considerations: information updating and information asymmetry , 2019, Int. J. Prod. Res..

[43]  Hing Kai Chan,et al.  Cascading Delay Risk of Airline Workforce Deployments with Crew Pairing and Schedule Optimization. , 2017, Risk analysis : an official publication of the Society for Risk Analysis.

[44]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.

[45]  Geoffrey E. Hinton,et al.  Deep Learning , 2015, Nature.

[46]  Mohammad Bagher Menhaj,et al.  Training feedforward networks with the Marquardt algorithm , 1994, IEEE Trans. Neural Networks.

[47]  Tony R. Martinez,et al.  The general inefficiency of batch training for gradient descent learning , 2003, Neural Networks.

[48]  Gunasekaran Manogaran,et al.  Neural Network Based Brain Tumor Detection Using Wireless Infrared Imaging Sensor , 2019, IEEE Access.

[49]  Fei Han,et al.  An improved incremental constructive single-hidden-layer feedforward networks for extreme learning machine based on particle swarm optimization , 2017, Neurocomputing.

[50]  Chun-Ho Wu,et al.  Customer relationship mining system for effective strategies formulation , 2014, Ind. Manag. Data Syst..

[51]  Gerhard Rigoll,et al.  A deep convolutional neural network for video sequence background subtraction , 2018, Pattern Recognit..

[52]  Narasimhan Sundararajan,et al.  A Fast and Accurate Online Sequential Learning Algorithm for Feedforward Networks , 2006, IEEE Transactions on Neural Networks.

[53]  Tsan-Ming Choi,et al.  Sustainable Fashion Supply Chain Management: A System of Systems Analysis , 2019, IEEE Transactions on Engineering Management.

[54]  Xizhao Wang,et al.  A review on neural networks with random weights , 2018, Neurocomputing.

[55]  Daniel S. Yeung,et al.  Localized Generalization Error Model and Its Application to Architecture Selection for Radial Basis Function Neural Network , 2007, IEEE Transactions on Neural Networks.

[56]  T. C. Wong,et al.  Examining the key determinants towards online pro-brand and anti-brand community citizenship behaviours: A two-stage approach , 2018, Ind. Manag. Data Syst..

[57]  Xiaoming Liu,et al.  Multi-Task Convolutional Neural Network for Pose-Invariant Face Recognition , 2017, IEEE Transactions on Image Processing.

[58]  Keng-Boon Ooi,et al.  The effects of convenience and speed in m-payment , 2015, Ind. Manag. Data Syst..

[59]  Hae-Chang Rim,et al.  Business environmental analysis for textual data using data mining and sentence-level classification , 2019, Ind. Manag. Data Syst..

[60]  Okyay Kaynak,et al.  Computing Gradient Vector and Jacobian Matrix in Arbitrarily Connected Neural Networks , 2008, IEEE Transactions on Industrial Electronics.

[61]  Sule Yildirim Yayilgan,et al.  The impact of deep learning on document classification using semantically rich representations , 2019, Inf. Process. Manag..

[62]  Guang-Bin Huang,et al.  Convex incremental extreme learning machine , 2007, Neurocomputing.

[63]  Rudy Setiono,et al.  Use of a quasi-Newton method in a feedforward neural network construction algorithm , 1995, IEEE Trans. Neural Networks.

[64]  D. Shanno Conditioning of Quasi-Newton Methods for Function Minimization , 1970 .

[65]  Hongming Zhou,et al.  Extreme Learning Machine for Regression and Multiclass Classification , 2012, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[66]  Yaonan Wang,et al.  Bidirectional Extreme Learning Machine for Regression Problem and Its Learning Effectiveness , 2012, IEEE Transactions on Neural Networks and Learning Systems.