Task Decompostiion Through Competition in a Modular Connectionist Architecture: The What and Where Vision Tasks

A novel modular connectionist architecture is presented in which the networks composing the architecture compete to learn the training potterns. An outcome of the competition is that different networks learn different training patterns and, thus, learn to compute different functions. The architecture performs task decomposition in the sense that it learns to partition a task into two or more functionally independent tasks and allocates distinct networks to learn each task. In addition, the architecture tends to allocate to each task the network whose topology is most appropriate to that task. The architecture’s performance on “what” and “where” vision tasks is presented and compared with the performcmce of two multilayer networks. Finally, it is noted that function decomposition is an underconstrained problem, and, thus, different modular architectures may decompose a function in different ways. A desirable decomposition con be achieved if the architecture is suitably restricted in the types of functions that it can compute. Finding appropriate restrictions is possible through the application of domain knowledge. A strength of the modular architecture is that its structure is well suited for incorporating domain knowledge.

[1]  Richard O. Duda,et al.  Pattern classification and scene analysis , 1974, A Wiley-Interscience publication.

[2]  P. Werbos,et al.  Beyond Regression : "New Tools for Prediction and Analysis in the Behavioral Sciences , 1974 .

[3]  James S. Albus,et al.  New Approach to Manipulator Control: The Cerebellar Model Articulation Controller (CMAC)1 , 1975 .

[4]  Geoffrey E. Hinton Shape Representation in Parallel Systems , 1981, IJCAI.

[5]  Geoffrey E. Hinton A Parallel Computation that Assigns Canonical Object-Based Frames of Reference , 1981, IJCAI.

[6]  J. Fodor The Modularity of mind. An essay on faculty psychology , 1986 .

[7]  Leslie G. Ungerleider,et al.  Object vision and spatial vision: two cortical pathways , 1983, Trends in Neurosciences.

[8]  Dana H. Ballard,et al.  Parameter Nets , 1984, Artif. Intell..

[9]  P. Anandan,et al.  Pattern-recognizing stochastic learning automata , 1985, IEEE Transactions on Systems, Man, and Cybernetics.

[10]  David Zipser,et al.  Feature Discovery by Competive Learning , 1986, Cogn. Sci..

[11]  Yann LeCun,et al.  Une procedure d'apprentissage pour reseau a seuil asymmetrique (A learning scheme for asymmetric threshold networks) , 1985 .

[12]  K. Forster,et al.  The psychological status of overgenerated sentences , 1985, Cognition.

[13]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .

[14]  M. Minsky The Society of Mind , 1986 .

[15]  James L. McClelland The programmable blackboard model of reading , 1986 .

[16]  Dana H. Ballard,et al.  Cortical connections and parallel processing: Structure and function , 1986, Behavioral and Brain Sciences.

[17]  H. Barlow Why have multiple cortical areas? , 1986, Vision Research.

[18]  James L. McClelland,et al.  Parallel distributed processing: explorations in the microstructure of cognition, vol. 1: foundations , 1986 .

[19]  Stephen Grossberg,et al.  Competitive Learning: From Interactive Activation to Adaptive Resonance , 1987, Cogn. Sci..

[20]  Lawrence D. Jackel,et al.  Large Automatic Learning, Rule Extraction, and Generalization , 1987, Complex Syst..

[21]  Paul J. Werbos,et al.  Building and Understanding Adaptive Systems: A Statistical/Numerical Approach to Factory Automation and Brain Research , 1987, IEEE Transactions on Systems, Man, and Cybernetics.

[22]  Geoffrey E. Hinton,et al.  Learning sets of filters using back-propagation , 1987 .

[23]  J. Beaumont Cerebral Lateralization: Biological Mechanisms, Associations, and Pathology , 1987 .

[24]  S. Kosslyn Seeing and imagining in the cerebral hemispheres: a computational approach. , 1987, Psychological review.

[25]  C. Lee Giles,et al.  Nonlinear dynamics of artificial neural systems , 1987 .

[26]  Dean A. Pomerleau,et al.  The meta-generalized delta rule : a new algorithm for learning in connectionist networks , 1987 .

[27]  Richard Durbin,et al.  An analogue approach to the travelling salesman problem using an elastic net method , 1987, Nature.

[28]  Dana H. Ballard,et al.  Modular Learning in Neural Networks , 1987, AAAI.

[29]  Jerome A. Feldman,et al.  Connectionist Models and Their Properties , 1982, Cogn. Sci..

[30]  Michael I. Jordan Supervised learning and systems with excess degrees of freedom , 1988 .

[31]  Yann LeCun,et al.  Generalization and network design strategies , 1989 .

[32]  John Scott Bridle,et al.  Probabilistic Interpretation of Feedforward Classification Network Outputs, with Relationships to Statistical Pattern Recognition , 1989, NATO Neurocomputing.

[33]  Kumpati S. Narendra,et al.  Learning automata - an introduction , 1989 .

[34]  Steven J. Nowlan,et al.  Maximum Likelihood Competitive Learning , 1989, NIPS.

[35]  Dit-Yan Yeung,et al.  Using a context-sensitive learning for robot arm control , 1989, Proceedings, 1989 International Conference on Robotics and Automation.

[36]  Peter Földiák,et al.  Adaptation and decorrelation in the cortex , 1989 .

[37]  George A. Bekey,et al.  Using a context-sensitive learning network for robot arm control , 1989 .

[38]  Alan L. Yuille,et al.  A Winner-Take-All Mechanism Based on Presynaptic Inhibition Feedback , 1989, Neural Computation.

[39]  Geoffrey E. Hinton Connectionist Learning Procedures , 1989, Artif. Intell..

[40]  S. Kosslyn,et al.  Why are What and Where Processed by Separate Cortical Visual Systems? A Computational Investigation , 1989, Journal of Cognitive Neuroscience.

[41]  Alexander H. Waibel,et al.  Modular Construction of Time-Delay Neural Networks for Speech Recognition , 1989, Neural Computation.

[42]  Kurt Hornik,et al.  Multilayer feedforward networks are universal approximators , 1989, Neural Networks.

[43]  Tomaso A. Poggio,et al.  Extensions of a Theory of Networks for Approximation and Learning , 1990, NIPS.

[44]  Michael I. Jordan Attractor dynamics and parallelism in a connectionist sequential machine , 1990 .

[45]  Richard Durbin,et al.  A dimension reduction framework for understanding cortical maps , 1990, Nature.

[46]  Michael I. Jordan,et al.  Task Decomposition Through Competition in a Modular Connectionist Architecture: The What and Where Vision Tasks , 1990, Cogn. Sci..

[47]  Geoffrey E. Hinton,et al.  Adaptive Mixtures of Local Experts , 1991, Neural Computation.