Determining the Efficient Structure of Feed-Forward Neural Network to Classify Breast Cancer Dataset

Classification is one of the most frequently encountered problems in data mining. A classification problem occurs when an object needs to be assigned in predefined classes based on a number of observed attributes related to that object. Neural networks have emerged as one of the tools that can handle the classification problem. Feed-forward Neural Networks (FNN's) have been widely applied in many different fields as a classification tool. Designing an efficient FNN structure with optimum number of hidden layers and minimum number of layer's neurons, given a specific application or dataset, is an open research problem. In this paper, experimental work is carried out to determine an efficient FNN structure, that is, a structure with the minimum number of hidden layer's neurons for classifying the Wisconsin Breast Cancer Dataset. We achieve this by measuring the classification performance using the Mean Square Error (MSE) and controlling the number of hidden layers, and the number of neurons in each layer. The experimental results show that the number of hidden layers has a significant effect on the classification performance and the best classification performance average is attained when the number of layers is 5, and number of hidden layer's neurons are small, typically 1 or 2.

[1]  Samuel Lukas,et al.  Backpropagation and Levenberg-Marquardt Algorithm for Training Finite Element Neural Network , 2012, 2012 Sixth UKSim/AMSS European Symposium on Computer Modeling and Simulation.

[2]  Martin T. Hagan,et al.  Neural network design , 1995 .

[3]  Wlodzislaw Duch,et al.  A new methodology of extraction, optimization and application of crisp and fuzzy logical rules , 2001, IEEE Trans. Neural Networks.

[4]  Mohammad Bagher Menhaj,et al.  MULTILAYER FEED FORWARD NEURAL NETWORKS FOR CONTROLLING DECENTRALIZED LARGE-SCALE NON-AFFINE NONLINEAR SYSTEMS WITH GUARANTEED STABILITY , 2010 .

[5]  M. Alexander,et al.  Principles of Neural Science , 1981 .

[6]  Huan Liu,et al.  Neural-network feature selector , 1997, IEEE Trans. Neural Networks.

[7]  Muhammad Ibn Ibrahimy,et al.  Design and Optimization of Levenberg-Marquardt based Neural Network Classifier for EMG Signals to Identify Hand Motions , 2013 .

[8]  Shuxiang Xu,et al.  A novel approach for determining the optimal number of hidden layer neurons for FNN’s and its application in data mining , 2008 .

[9]  G R Newell,et al.  Cancer and age. , 1989, Seminars in oncology.

[10]  F. Paulin,et al.  Classification of Breast cancer by comparing Back propagation training algorithms , 2011 .

[11]  Amol P. Pande,et al.  Neural Network Aided Breast Cancer Detection and Diagnosis Using Support Vector Machine , 2006 .

[12]  Bindu Garg,et al.  Optimizing Number of Inputs to Classify Breast Cancer Using Artificial Neural Network , 2009 .

[13]  A. Aminian,et al.  A multi-layer feed forward neural network model for accurate prediction of flue gas sulfuric acid dew points in process industries , 2010 .

[14]  Fernando Morgado-Dias,et al.  Performance comparison of ANN training algorithms for classification , 2013, 2013 IEEE 8th International Symposium on Intelligent Signal Processing.

[15]  Tulay Yildirim,et al.  BREAST CANCER DIAGNOSIS USING STATISTICAL NEURAL NETWORKS , 2004 .

[16]  Osamu Fujita,et al.  Statistical estimation of the number of hidden units for feedforward neural networks , 1998, Neural Networks.

[17]  Duc Truong Pham,et al.  Computational Intelligence: for Engineering and Manufacturing , 2007 .