A Novel Technique for Optimizing the Hidden Layer Architecture in Artificial Neural Networks

The architecture of an artificial neural network has a great impact on the generalization power. More precisely, by changing the number of layers and neurons in each hidden layer generalization ability can be significantly changed. Therefore, the architecture is crucial in artificial neural network and hence, determining the hidden layer architecture has become a research challenge. In this paper a pruning technique has been presented to obtain an appropriate architecture based on the backpropagation training algorithm. Pruning is done by using the delta values of hidden layers. The proposed method has been tested with several benchmark problems in artificial neural networks and machine learning. The experimental results have been shown that the modified algorithm reduces the size of the network without degrading the performance. Also it tends to the desired error faster than the backpropagation algorithm.

[1]  Pravin Chandra,et al.  CONSTRUCTIVE NEURAL NETWORKS: A REVIEW , 2010 .

[2]  Suman Ahmmed,et al.  Architecture and Weight Optimization of ANN Using Sensitive Analysis and Adaptive Particle Swarm Optimization , 2010 .

[3]  Demetris Stathakis,et al.  How many hidden layers and nodes? , 2009 .

[4]  Yann LeCun,et al.  Optimal Brain Damage , 1989, NIPS.

[5]  Xin Yao,et al.  A New Adaptive Merging and Growing Algorithm for Designing Artificial Neural Networks , 2009, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[6]  Babak Hassibi,et al.  Second Order Derivatives for Network Pruning: Optimal Brain Surgeon , 1992, NIPS.

[7]  S. S. Sridhar,et al.  Improved Adaptive Learning Algorithm for Constructive Neural Networks , 2011 .

[8]  A. Asuncion,et al.  UCI Machine Learning Repository, University of California, Irvine, School of Information and Computer Sciences , 2007 .

[9]  N. M. Wagarachchi,et al.  Optimization of multi-layer artificial neural networks using delta values of hidden layers , 2013, 2013 IEEE Symposium on Computational Intelligence, Cognitive Algorithms, Mind, and Brain (CCMB).

[10]  D.R. Hush,et al.  Progress in supervised neural networks , 1993, IEEE Signal Processing Magazine.

[11]  Anthony N. Burkitt,et al.  Optimization of the Architecture of Feed-forward Neural Networks with Hidden Layers by Unit Elimination , 1991, Complex Syst..

[12]  Giovanna Castellano,et al.  An iterative pruning algorithm for feedforward neural networks , 1997, IEEE Trans. Neural Networks.

[13]  David Haussler,et al.  What Size Net Gives Valid Generalization? , 1989, Neural Computation.

[14]  Mahmood R. Azimi-Sadjadi,et al.  Recursive dynamic node creation in multilayer neural networks , 1993, IEEE Trans. Neural Networks.