Learning with generalized-mean neuron model

Abstract The artificial neuron has come a long way in modeling the functional capabilities of various neuronal processes. The higher order neurons have shown improved computational power and generalization ability. However, these models are difficult to train because of a combinatorial explosion of higher order terms as the number of inputs to the neuron increases. This work presents an artificial neural network using a neuron architecture called generalized mean neuron (GMN) model. This neuron model consists of an aggregation function which is based on the generalized mean of the all the inputs applied to it. The proposed neuron model with same number of parameters as the McCulloch–Pitts model demonstrates better computational power. The performance of this model has been benchmarked on both classification and time series prediction problems.

[1]  Kaoru Hirota,et al.  A Solution for the N-bit Parity Problem Using a Single Translated Multiplicative Neuron , 2004, Neural Processing Letters.

[2]  David B. Fogel An information criterion for optimal neural network selection , 1991, IEEE Trans. Neural Networks.

[3]  Andrzej Piegat,et al.  Fuzzy Modeling and Control , 2001 .

[4]  Shun-ichi Amari,et al.  Network information criterion-determining the number of hidden units for an artificial neural network model , 1994, IEEE Trans. Neural Networks.

[5]  W. Pitts,et al.  A Logical Calculus of the Ideas Immanent in Nervous Activity (1943) , 2021, Ideas That Created the Future.

[6]  K. Schreiner Retail robots on the move , 2001 .

[7]  H. Akaike A new look at the statistical model identification , 1974 .

[8]  Richard Labib,et al.  New single neuron structure for solving nonlinear problems , 1999, IJCNN'99. International Joint Conference on Neural Networks. Proceedings (Cat. No.99CH36339).

[9]  Michael Schmitt,et al.  On the Complexity of Computing and Learning with Multiplicative Neural Networks , 2002, Neural Computation.

[10]  K. Schreiner Neuron function: the mystery persists , 2001 .

[11]  Brian D. Ripley,et al.  Pattern Recognition and Neural Networks , 1996 .

[12]  John G. Proakis,et al.  Digital Communications , 1983 .

[13]  Meng Wang,et al.  Logic operations based on single neuron rational model , 2000, IEEE Trans. Neural Networks Learn. Syst..

[14]  Tin Kam Ho,et al.  The learning behavior of single neuron classifiers on linearly separable or nonseparable input , 1999, IJCNN'99. International Joint Conference on Neural Networks. Proceedings (Cat. No.99CH36339).

[15]  T A Plate,et al.  Randomly connected sigma–pi neurons can form associator networks , 2000, Network.

[16]  Gwilym M. Jenkins,et al.  Time series analysis, forecasting and control , 1971 .

[17]  L. Glass,et al.  Oscillation and chaos in physiological control systems. , 1977, Science.