Generalization and Selection of Examples in Feedforward Neural Networks

In this work, we study how the selection of examples affects the learning procedure in a boolean neural network and its relationship with the complexity of the function under study and its architecture. We analyze the generalization capacity for different target functions with particular architectures through an analytical calculation of the minimum number of examples needed to obtain full generalization (i.e., zero generalization error). The analysis of the training sets associated with such parameter leads us to propose a general architecture-independent criterion for selection of training examples. The criterion was checked through numerical simulations for various particular target functions with particular architectures, as well as for random target functions in a nonoverlapping receptive field perceptron. In all cases, the selection sampling criterion lead to an improvement in the generalization capacity compared with a pure random sampling. We also show that for the parity problem, one of the most used problems for testing learning algorithms, only the use of the whole set of examples ensures global learning in a depth two architecture. We show that this difficulty can be overcome by considering a tree-structured network of depth 2 log2(N) 1.

[1]  David A. Cohn,et al.  Neural Network Exploration Using Optimal Experiment Design , 1993, NIPS.

[2]  Simon Haykin,et al.  Neural Networks: A Comprehensive Foundation , 1998 .

[3]  N. Caticha,et al.  On-line learning in the committee machine , 1995 .

[4]  Leonardo Franco,et al.  Generalization properties of modular networks: implementing the parity function , 2001, IEEE Trans. Neural Networks.

[5]  Anders Krogh,et al.  Introduction to the theory of neural computation , 1994, The advanced book program.

[6]  Sergio A. Cannas Arithmetic Perceptrons , 1995, Neural Computation.

[7]  Gerald Tesauro,et al.  Scaling Relationships in Back-propagation Learning , 1988, Complex Syst..

[8]  Blatt,et al.  Computational capabilities of restricted two-layered perceptrons. , 1994, Physical review. E, Statistical physics, plasmas, fluids, and related interdisciplinary topics.

[9]  Wolfgang Kinzel,et al.  Improving a Network Generalization Ability by Selecting Examples , 1990 .

[10]  James L. McClelland,et al.  Parallel distributed processing: explorations in the microstructure of cognition, vol. 1: foundations , 1986 .

[11]  Leonardo Franco,et al.  Solving arithmetic problems using feed-forward neural networks , 1998, Neurocomputing.

[12]  Jungyand Manfred OpperzyPhysikalisches Selection of Examples for a Linear Classi erGeorg , 1995 .

[13]  RIJKSUNIVERSITEIT TE LEIDEN VAKGROEP INFORMATICA Designing Modular Artificial Neural Networks , 1993 .

[14]  Mark Plutowski,et al.  Selecting concise training sets from clean data , 1993, IEEE Trans. Neural Networks.

[15]  Eric B. Baum,et al.  Neural net algorithms that learn in polynomial time from examples and queries , 1991, IEEE Trans. Neural Networks.

[16]  David Haussler,et al.  What Size Net Gives Valid Generalization? , 1989, Neural Computation.