A new Neural Network pruning method based on the singular value decomposition and the weight initialisation

In this paper, we present an efficient procedure to determine the optimal hidden unit number of a feed-forward multi-layer Neural Network (NN) using the singular value decomposition (SVD) taking into account the function to be approximated by the NN and the initial values of the updating weights. The SVD is used to identify and eliminate redundant hidden nodes. Minimizing redundancy gives smaller networks, producing models that generalize better and thus eliminate the need of using cross-validation to avoid overfitting. Using this procedure we obtain a final model with fewer adjustable parameters and more accurate predictions than a network model with a fixed, a priori determined, size. We show these performances by applying this procedure to several problems such as function approximation and image recognition.

[1]  Rudy Setiono,et al.  Use of a quasi-Newton method in a feedforward neural network construction algorithm , 1995, IEEE Trans. Neural Networks.

[2]  Mohamed Najim,et al.  Evaluation of the feedforward neural network covariance matrix error , 2000, 2000 IEEE International Conference on Acoustics, Speech, and Signal Processing. Proceedings (Cat. No.00CH37100).

[3]  Alberto Del Bimbo,et al.  Block-structured recurrent neural networks , 1995, Neural Networks.

[4]  Mohamed Najim,et al.  A fast feedforward training algorithm using a modified form of the standard backpropagation algorithm , 2001, IEEE Trans. Neural Networks.

[5]  Lyle H. Ungar,et al.  SVD-NET: an algorithm that automatically selects network structure , 1994, IEEE Trans. Neural Networks.

[6]  Michael C. Mozer,et al.  Using Relevance to Reduce Network Size Automatically , 1989 .

[7]  Mohamed Najim,et al.  A new feedforward neural network hidden layer neuron pruning algorithm , 2001, 2001 IEEE International Conference on Acoustics, Speech, and Signal Processing. Proceedings (Cat. No.01CH37221).

[8]  Shun-ichi Amari,et al.  Network information criterion-determining the number of hidden units for an artificial neural network model , 1994, IEEE Trans. Neural Networks.

[9]  Babak Hassibi,et al.  Second Order Derivatives for Network Pruning: Optimal Brain Surgeon , 1992, NIPS.

[10]  Benjamin W. Wah,et al.  Automated learning for reducing the configuration of a feedforward neural network , 1996, IEEE Trans. Neural Networks.

[11]  Giovanna Castellano,et al.  An iterative pruning algorithm for feedforward neural networks , 1997, IEEE Trans. Neural Networks.