Generalization Capability of Artificial Neural Network Incorporated with Pruning Method

In any real world application, the performance of Artificial Neural Networks (ANN) is mostly depends upon its generalization capability. Generalization of the ANN is ability to handle unseen data. The generalization capability of the network is mostly determined by system complexity and training of the network. Poor generalization is observed when the network is over-trained or system complexity (or degree of freedom) is relatively more than the training data. A smaller network which can fit the data will have the k good generalization ability. Network parameter pruning is one of the promising methods to reduce the degree of freedom of a network and hence improve its generalization. In recent years various pruning methods have been developed and found effective in real world applications. Next, it is important to estimate the improvement in generalization and rate of improvement as pruning being incorporated in the network. A method is developed in this research to evaluate generalization capability and rate of convergence towards the generalization. Using the proposed method, experiments have been conducted to evaluate Multi-Layer Perceptron neural network with pruning being incorporated for handwritten numeral recognition.

[1]  Giorgio Guariso,et al.  An application of pruning in the design of neural networks for real time flood forecasting , 2005, Neural Computing & Applications.

[2]  C. Lee Giles,et al.  Overfitting and neural networks: conjugate gradient and backpropagation , 2000, Proceedings of the IEEE-INNS-ENNS International Joint Conference on Neural Networks. IJCNN 2000. Neural Computing: New Challenges and Perspectives for the New Millennium.

[3]  David Haussler,et al.  Learnability and the Vapnik-Chervonenkis dimension , 1989, JACM.

[4]  Babak Hassibi,et al.  Second Order Derivatives for Network Pruning: Optimal Brain Surgeon , 1992, NIPS.

[5]  Kenji Suzuki,et al.  A Simple Neural Network Pruning Algorithm with Application to Filter Synthesis , 2001, Neural Processing Letters.

[6]  Yoshua Bengio,et al.  Neural networks for speech and sequence recognition , 1996 .

[7]  Paul M. Mather,et al.  ASSESSING ARTIFICIAL NEURAL NETWORK PRUNING ALGORITHMS , 2000 .

[8]  V. Vapnik Estimation of Dependences Based on Empirical Data , 2006 .

[9]  Ehud D. Karnin,et al.  A simple procedure for pruning back-propagation trained neural networks , 1990, IEEE Trans. Neural Networks.

[10]  Russell Reed,et al.  Pruning algorithms-a survey , 1993, IEEE Trans. Neural Networks.