Generative learning structures and processes for generalized connectionist networks

Abstract Massively parallel networks of relatively simple computing elements offer an attractive and versatile framework for exploring a variety of learning structures and processes for intelligent systems. This paper briefly summarizes some popular learning structures and processes used in such networks. It outlines a range of potentially more powerful alternatives for pattern-directed inductive learning in such systems. It motivates and develops a class of new learning algorithms for massively parallel networks of simple computing elements. We call this class of learning processes generative for they offer a set of mechanisms for constructive and adaptive determination of the network architecture—the number of processing elements and the connectivity among them—as a function of experience. Generative learning algorithms attempt to overcome some of the limitations of some approaches to learning in networks that rely on modification of weights on the links within an otherwise fixed network topology, for example, rather slow learning and the need for an a priori choice of network architecture. Several alternative designs as well as a range of control structures and processes that can be used to regulate the form and content of internal representations learned by such networks are examined. Empirical results from the study of some generative learning algorithms are briefly summarized, and several extensions and refinements of such algorithms and directions for future research are outlined.

[1]  John R. Koza,et al.  Genetic programming: a paradigm for genetically breeding populations of computer programs to solve problems , 1990 .

[2]  David S. Johnson,et al.  Computers and Intractability: A Guide to the Theory of NP-Completeness , 1978 .

[3]  Claude E. Shannon,et al.  The mathematical theory of communication , 1950 .

[4]  Geoffrey E. Hinton Connectionist Learning Procedures , 1989, Artif. Intell..

[5]  Leslie G. Valiant,et al.  A theory of the learnable , 1984, STOC '84.

[6]  E. Mark Gold,et al.  Language Identification in the Limit , 1967, Inf. Control..

[7]  Stephen Grossberg,et al.  Competitive Learning: From Interactive Activation to Adaptive Resonance , 1987, Cogn. Sci..

[8]  C. H. Bailey,et al.  The anatomy of a memory: convergence of results across a diversity of tests , 1988, Trends in Neurosciences.

[9]  G. T. Ladd President's address before the New York Meeting of the American Psychological Association. , 1894 .

[10]  David W. Aha,et al.  Noise-Tolerant Instance-Based Learning Algorithms , 1989, IJCAI.

[11]  Allen Newell,et al.  SOAR: An Architecture for General Intelligence , 1987, Artif. Intell..

[12]  Geoffrey E. Hinton,et al.  A general framework for parallel distributed processing , 1986 .

[13]  Vasant Honavar,et al.  Coordination and control structures and processes: possibilities for connectionist networks (CN) , 1990, J. Exp. Theor. Artif. Intell..

[14]  R. Michalski,et al.  Learning from Observation: Conceptual Clustering , 1983 .

[15]  Steven E. Hampson,et al.  Connectionistic Problem Solving , 1990, Birkhäuser Boston.

[16]  S. Grossberg How does a brain build a cognitive code , 1980 .

[17]  Yann LeCun,et al.  Optimal Brain Damage , 1989, NIPS.

[18]  John K. Tsotsos Analyzing vision at the complexity level , 1990, Behavioral and Brain Sciences.

[19]  Vasant Honavar,et al.  Experiments with the cascade-correlation algorithm , 1991, [Proceedings] 1991 IEEE International Joint Conference on Neural Networks.

[20]  Vasant Honavar,et al.  Faster Learning in Multi-Layer Networks by Handling Output-Layer Flat-Spots , 1992 .

[21]  Stephen Jose Hanson,et al.  Meiosis Networks , 1989, NIPS.

[22]  Jean-Pierre Nadal,et al.  Study of a Growth Algorithm for a Feedforward Network , 1989, Int. J. Neural Syst..

[23]  John H. Holland,et al.  Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence , 1992 .

[24]  Ryszard S. Michalski,et al.  Research in machine learning: recent progress, classification of methods, and future directions , 1990 .

[25]  Vasant Honavar,et al.  Generation, Local Receptive Fields and Global Convergence Improve Perceptual Learning in Connectionist Networks , 1989, IJCAI.

[26]  Raoul Tawel Does the Neuron "Learn" Like the Synapse? , 1988, NIPS.

[27]  Vasant G Honavar Perceptual Development and Learning: From Behavioral, Neurophysiological, and Morphological Evidence To Computational Models , 1989 .

[28]  David H. Wolpert,et al.  A Mathematical Theory of Generalization: Part II , 1990, Complex Syst..

[29]  Christian Lebiere,et al.  The Cascade-Correlation Learning Architecture , 1989, NIPS.

[30]  Nils J. Nilsson,et al.  The Mathematical Foundations of Learning Machines , 1990 .

[31]  Thomas G. Dietterich,et al.  A Comparative Review of Selected Methods for Learning from Examples , 1983 .

[32]  Alan S. Perelson,et al.  The immune system, adaptation, and machine learning , 1986 .

[33]  Stephen I. Gallant,et al.  Perceptron-based learning algorithms , 1990, IEEE Trans. Neural Networks.

[34]  F ROSENBLATT,et al.  The perceptron: a probabilistic model for information storage and organization in the brain. , 1958, Psychological review.

[35]  Translator-IEEE Expert staff Machine Learning: A Theoretical Approach , 1992, IEEE Expert.

[36]  E. Capaldi,et al.  The organization of behavior. , 1992, Journal of applied behavior analysis.

[37]  P. Werbos,et al.  Beyond Regression : "New Tools for Prediction and Analysis in the Behavioral Sciences , 1974 .

[38]  Marvin Minsky,et al.  Perceptrons: An Introduction to Computational Geometry , 1969 .

[39]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .

[40]  S. Grossberg,et al.  How does a brain build a cognitive code? , 1980, Psychological review.

[41]  Lawrence J. Fogel,et al.  Artificial Intelligence through Simulated Evolution , 1966 .

[42]  Bruce Porter,et al.  Protos: a unified approach to concept representation, classification, and learning , 1988 .

[43]  Vasant Honavar,et al.  Brain-structured Connectionist Networks that Perceive and Learn , 1989 .

[44]  Peter Anthony Sandon Learning object-centered representations , 1987 .

[45]  David H. Wolpert,et al.  A Mathematical Theory of Generalization: Part I , 1990, Complex Syst..

[46]  John R. Anderson,et al.  MACHINE LEARNING An Artificial Intelligence Approach , 2009 .

[47]  Vasant Honavar Learning Parsimonious Representations of Three-Dimensional Shapes , 1992 .

[48]  P. Anandan,et al.  Pattern-recognizing stochastic learning automata , 1985, IEEE Transactions on Systems, Man, and Cybernetics.

[49]  T. Ash,et al.  Dynamic node creation in backpropagation networks , 1989, International 1989 Joint Conference on Neural Networks.

[50]  David Haussler,et al.  Two Algorithms That Learn DNF by Discovering Relevant Features , 1989, ML.

[51]  C. E. SHANNON,et al.  A mathematical theory of communication , 1948, MOCO.

[52]  Jerome A. Feldman,et al.  Connectionist Models and Their Properties , 1982, Cogn. Sci..

[53]  Eric B. Baum,et al.  A Proposal for More Powerful Learning Algorithms , 1989, Neural Computation.

[54]  Allen Newell,et al.  Physical Symbol Systems , 1980, Cogn. Sci..

[55]  Michael C. Mozer,et al.  Skeletonization: A Technique for Trimming the Fat from a Network via Relevance Assessment , 1988, NIPS.