A neuroidal architecture for cognitive computation

An architecture is described for designing systems that acquire and manipulate large amounts of unsystematized, or so-called commonsense, knowledge. Its aim is to exploit to the full those aspects of computational learning that are known to offer powerful solutions in the acquisition and maintenance of robust knowledge bases. The architecture makes explicit the requirements on the basic computational tasks that are to be performed and is designed to make these computationally tractable even for very large databases. The main claims are that (i) the basic learning tasks are tractable and (ii) tractable learning offers viable approaches to a range of issues that have been previously identified as problematic for artificial intelligence systems that are entirely programmed. In particular, attribute efficiency holds a central place in the definition of the learning tasks, as does also the capability to handle relational information efficiently. Among the issues that learning offers to resolve are robustness to inconsistencies, robustness to incomplete information and resolving among alternatives.

[1]  David Haussler,et al.  Quantifying Inductive Bias: AI Learning Algorithms and Valiant's Learning Framework , 1988, Artif. Intell..

[2]  Leslie G. Valiant,et al.  Learning Disjunction of Conjunctions , 1985, IJCAI.

[3]  Drew McDermott,et al.  Nonmonotonic Logic II: Nonmonotonic Modal Theories , 1982, JACM.

[4]  David Haussler,et al.  How to use expert advice , 1993, STOC.

[5]  Drew McDermott,et al.  Non-Monotonic Logic I , 1987, Artif. Intell..

[6]  Edith Cohen,et al.  Learning noisy perceptrons by a perceptron in polynomial time , 1997, Proceedings 38th Annual Symposium on Foundations of Computer Science.

[7]  Michael Kearns,et al.  Efficient noise-tolerant learning from statistical queries , 1993, STOC.

[8]  A. M. Turing,et al.  Computing Machinery and Intelligence , 1950, The Philosophy of Artificial Intelligence.

[9]  Ramanathan V. Guha,et al.  Cyc: toward programs with common sense , 1990, CACM.

[10]  Ming Li,et al.  Learning in the presence of malicious errors , 1993, STOC '88.

[11]  Nick Littlestone,et al.  From on-line to batch learning , 1989, COLT '89.

[12]  Joseph Y. Halpern An Analysis of First-Order Logics of Probability , 1989, IJCAI.

[13]  A. Newell Unified Theories of Cognition , 1990 .

[14]  G. A. Miller THE PSYCHOLOGICAL REVIEW THE MAGICAL NUMBER SEVEN, PLUS OR MINUS TWO: SOME LIMITS ON OUR CAPACITY FOR PROCESSING INFORMATION 1 , 1956 .

[15]  Stephen Muggleton,et al.  Inductive logic programming: derivations, successes and shortcomings , 1993, SGAR.

[16]  A. M. Turing,et al.  Computing Machinery and Intelligence , 1950, The Philosophy of Artificial Intelligence.

[17]  Leslie G. Valiant,et al.  A neuroidal architecture for cognitive computation , 2000, JACM.

[18]  Robert E. Schapire,et al.  Efficient distribution-free learning of probabilistic concepts , 1990, Proceedings [1990] 31st Annual Symposium on Foundations of Computer Science.

[19]  M. D. MacLaren The Art of Computer Programming—Volume 1: Fundamental Algorithms (Donald E. Knuth) , 1969 .

[20]  Steven Minton,et al.  Quantitative Results Concerning the Utility of Explanation-based Learning , 1988, Artif. Intell..

[21]  John McCarthy,et al.  SOME PHILOSOPHICAL PROBLEMS FROM THE STANDPOINT OF ARTI CIAL INTELLIGENCE , 1987 .

[22]  Dan Roth,et al.  Learning to Reason: The Non-Monotonic Case , 1995, IJCAI.

[23]  Alan M. Frieze,et al.  A Polynomial-Time Algorithm for Learning Noisy Linear Threshold Functions , 1996, Algorithmica.

[24]  John R. Anderson,et al.  Rules of the Mind , 1993 .

[25]  Dan Roth,et al.  Applying Winnow to Context-Sensitive Spelling Correction , 1996, ICML.

[26]  Leslie G. Valiant Projection learning , 1998, COLT' 98.

[27]  Dan Roth,et al.  Learning to reason , 1994, JACM.

[28]  Leslie G. Valiant,et al.  Circuits of the mind , 1994 .

[29]  N. Littlestone Learning Quickly When Irrelevant Attributes Abound: A New Linear-Threshold Algorithm , 1987, 28th Annual Symposium on Foundations of Computer Science (sfcs 1987).

[30]  Umesh V. Vazirani,et al.  An Introduction to Computational Learning Theory , 1994 .

[31]  Leslie G. Valiant,et al.  A general lower bound on the number of examples needed for learning , 1988, COLT '88.

[32]  Matthew L. Ginsberg,et al.  Readings in Nonmonotonic Reasoning , 1987, AAAI 1987.

[33]  Marvin Minsky,et al.  A framework for representing knowledge" in the psychology of computer vision , 1975 .

[34]  A. Tversky,et al.  Causal Schemata in Judgments under Uncertainty , 1982 .

[35]  Tom Bylander,et al.  Learning linear threshold functions in the presence of classification noise , 1994, COLT '94.

[36]  P. Langley,et al.  Production system models of learning and development , 1987 .

[37]  A. M. Turing,et al.  Solvable and Unsolvable Problems , 1954 .

[38]  Judea Pearl,et al.  Probabilistic reasoning in intelligent systems - networks of plausible inference , 1991, Morgan Kaufmann series in representation and reasoning.

[39]  Allen Newell,et al.  Human Problem Solving. , 1973 .

[40]  Marvin Minsky,et al.  A framework for representing knowledge , 1974 .

[41]  Leslie G. Valiant,et al.  A theory of the learnable , 1984, STOC '84.

[42]  A. A. Mullin,et al.  Principles of neurodynamics , 1962 .

[43]  Ronald L. Rivest,et al.  Learning decision lists , 2004, Machine Learning.

[44]  Leslie G. Valiant,et al.  Computational limitations on learning from examples , 1988, JACM.

[45]  Dan Roth A Connectionist Framework for Reasoning: Reasoning with Examples , 1996, AAAI/IAAI, Vol. 2.

[46]  Manfred K. Warmuth,et al.  The perceptron algorithm vs. Winnow: linear vs. logarithmic mistake bounds when few input variables are relevant , 1995, COLT '95.

[47]  John McCarthy,et al.  Circumscription - A Form of Non-Monotonic Reasoning , 1980, Artif. Intell..