Neurons vs Weights Pruning in Artificial Neural Networks

Artificial neural networks (ANN) are well known for their good classification abilities. Recent advances in deep learning imposed second ANN renaissance. But neural networks possesses some problems like choosing hyper parameters such as neuron layers count and sizes which can greatly influence classification rate. Thus pruning techniques were developed that can reduce network sizes, increase its generalization abilities and overcome overfitting. Pruning approaches, in contrast to growing neural networks approach, assume that sufficiently large ANN is already trained and can be simplified with acceptable classification accuracy loss. Current paper compares nodes vs weights pruning algorithms and gives experimental results for pruned networks accuracy rates versus their non-pruned counterparts. We conclude that nodes pruning is more preferable solution, with some sidenotes.

[1]  B. E. Segee,et al.  Fault tolerance of pruned multilayer networks , 1991, IJCNN-91-Seattle International Joint Conference on Neural Networks.

[2]  Yann LeCun,et al.  Optimal Brain Damage , 1989, NIPS.

[3]  Yves Chauvin,et al.  A Back-Propagation Algorithm with Optimal Use of Hidden Units , 1988, NIPS.

[4]  Hermann Ney,et al.  Cross-entropy vs. squared error training: a theoretical and experimental comparison , 2013, INTERSPEECH.

[5]  Lutz Prechelt,et al.  Adaptive parameter pruning in neural networks , 1997 .

[6]  Chuanyi Ji,et al.  Generalizing Smoothness Constraints from Discrete Samples , 1990, Neural Computation.

[7]  David E. Rumelhart,et al.  Generalization by Weight-Elimination with Application to Forecasting , 1990, NIPS.

[8]  Ehud D. Karnin,et al.  A simple procedure for pruning back-propagation trained neural networks , 1990, IEEE Trans. Neural Networks.

[9]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[10]  Russell Reed,et al.  Pruning algorithms-a survey , 1993, IEEE Trans. Neural Networks.

[11]  D. Rumelhart,et al.  Generalization by weight-elimination applied to currency exchange rate prediction , 1991, IJCNN-91-Seattle International Joint Conference on Neural Networks.

[12]  Gregory J. Wolff,et al.  Optimal Brain Surgeon and general network pruning , 1993, IEEE International Conference on Neural Networks.

[13]  Michael C. Mozer,et al.  Skeletonization: A Technique for Trimming the Fat from a Network via Relevance Assessment , 1988, NIPS.

[14]  A. Asuncion,et al.  UCI Machine Learning Repository, University of California, Irvine, School of Information and Computer Sciences , 2007 .

[15]  Geoffrey E. Hinton,et al.  Experiments on Learning by Back Propagation. , 1986 .

[16]  Andrey Bondarenko,et al.  Artificial Neural Network Generalization and Simplification via Pruning/ Mākslīgo neironu tīklu vispārināšana un vienkāršošana, izmantojot samazināšanu/ Обобщение и упрощение искусственных нейронных сетей через урезание , 2014 .

[17]  Huan Liu,et al.  Understanding Neural Networks via Rule Extraction , 1995, IJCAI.

[18]  Ferdinand Hergert,et al.  Improving model selection by nonconvergent methods , 1993, Neural Networks.

[19]  John K. Kruschke,et al.  Distributed bottlenecks for improved generalization in back-propagation networks , 1989 .

[20]  Arkady Borisov,et al.  Research of Artificial Neural Networks Abilities in Printed Words Recognition , 2010, Sci. J. Riga Tech. Univ. Ser. Comput. Sci..

[21]  Vikas Chaudhary,et al.  Growing Neural Networks using Soft Competitive Learning , 2011 .

[22]  David E. Rumelhart,et al.  BACK-PROPAGATION, WEIGHT-ELIMINATION AND TIME SERIES PREDICTION , 1991 .

[23]  Xinjian Qiang,et al.  An overview of some classical Growing Neural Networks and new developments , 2010, 2010 2nd International Conference on Education Technology and Computer.

[24]  Wee Kheng Leow,et al.  Pruned Neural Networks for Regression , 2000, PRICAI.