The State Space of Artificial Intelligence

The goal of the paper is to develop and propose a general model of the state space of AI. Given the breathtaking progress in AI research and technologies in recent years, such conceptual work is of substantial theoretical interest. The present AI hype is mainly driven by the triumph of deep learning neural networks. As the distinguishing feature of such networks is the ability to self-learn, self-learning is identified as one important dimension of the AI state space. Another dimension is recognized as generalization, the possibility to go over from specific to more general types of problems. A third dimension is semantic grounding. Our overall analysis connects to a number of known foundational issues in the philosophy of mind and cognition: the blockhead objection, the Turing test, the symbol grounding problem, the Chinese room argument, and use theories of meaning. It shall finally be argued that the dimension of grounding decomposes into three sub-dimensions. And the dimension of self-learning turns out as only one of a whole range of “self-x-capacities” (based on ideas of organic computing) that span the self-x-subspace of the full AI state space.

[1]  Arno Schubbach,et al.  Judging machines: philosophical aspects of deep learning , 2019, Synthese.

[2]  Yoshua Bengio,et al.  Towards Biologically Plausible Deep Learning , 2015, ArXiv.

[3]  Rolf P. Wrtz Organic Computing - Understanding Complex Systems , 2008 .

[4]  Terrence J. Sejnowski,et al.  The Deep Learning Revolution , 2018 .

[5]  A. M. Turing,et al.  Computing Machinery and Intelligence , 1950, The Philosophy of Artificial Intelligence.

[6]  J. Weizenbaum Computer Power And Human Reason: From Judgement To Calculation , 1978 .

[7]  Cameron Buckner,et al.  Empiricism without magic: transformational abstraction in deep convolutional neural networks , 2018, Synthese.

[8]  H. Lyre Active Content Externalism , 2016 .

[9]  Ned Block,et al.  Semantics, Conceptual Role , 1997 .

[10]  Demis Hassabis,et al.  Mastering the game of Go without human knowledge , 2017, Nature.

[11]  Nicholas Shea,et al.  Representation in Cognitive Science , 2018, Oxford Scholarship Online.

[12]  Jürgen Schmidhuber,et al.  Deep learning in neural networks: An overview , 2014, Neural Networks.

[13]  Carlos Zednik,et al.  Solving the Black Box Problem: A Normative Framework for Explainable Artificial Intelligence , 2019, Philosophy & Technology.

[14]  R. Goodstein,et al.  Remarks on the Foundations of Mathematics , 1957, The Mathematical Gazette.

[15]  Noam Chomsky,et al.  Rules and representations , 1980, Behavioral and Brain Sciences.

[16]  P. Robbins,et al.  The Cambridge Handbook of Situated Cognition , 2001 .

[17]  John R. Searle,et al.  Minds, brains, and programs , 1980, Behavioral and Brain Sciences.

[18]  Hava T. Siegelmann,et al.  On the Computational Power of Neural Nets , 1995, J. Comput. Syst. Sci..

[19]  Shimon Ullman,et al.  Using neuroscience to develop artificial intelligence , 2019, Science.

[20]  Stevan Harnad,et al.  What's Wrong and Right About Searle's Chinese Room Argument? , 2001 .

[21]  Cameron Buckner Deep learning: A philosophical introduction , 2019, Philosophy Compass.

[22]  Jane X. Wang,et al.  Reinforcement Learning, Fast and Slow , 2019, Trends in Cognitive Sciences.

[23]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[24]  N. Block Psychologism and Behaviorism , 1981 .

[25]  Andrés Páez,et al.  The Pragmatic Turn in Explainable Artificial Intelligence (XAI) , 2019, Minds and Machines.

[26]  Mariarosaria Taddeo,et al.  Solving the symbol grounding problem: a critical review of fifteen years of research , 2005, J. Exp. Theor. Artif. Intell..

[27]  Max Tegmark Life 3.0: Being Human in the Age of Artificial Intelligence , 2017 .

[28]  Geoffrey E. Hinton,et al.  Deep Learning , 2015, Nature.

[29]  Ezequiel López-Rubio,et al.  Computational Functionalism for the Deep Learning Era , 2018, Minds and Machines.

[30]  Alex Graves,et al.  Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.

[31]  J. Searle,et al.  Is the brain's mind a computer program? , 1990, Scientific American.

[32]  D. Hassabis,et al.  Neuroscience-Inspired Artificial Intelligence , 2017, Neuron.

[33]  R. Cummins Representations, targets, and attitudes , 1996 .

[34]  Christian Müller-Schloer,et al.  Organic Computing – Technical Systems for Survival in the Real World , 2017, Autonomic Systems.

[35]  Stevan Harnad,et al.  Symbol grounding problem , 1990, Scholarpedia.

[36]  H. Lyre Humean Perspectives on Structural Realism , 2010 .

[37]  Feng-Hsiung Hsu,et al.  Behind Deep Blue: Building the Computer that Defeated the World Chess Champion , 2002 .

[38]  Stevan Harnad,et al.  Minds, machines and Searle , 1989, J. Exp. Theor. Artif. Intell..

[39]  R. Kirk Language, Thought, and Other Biological Categories , 1985 .

[40]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[41]  Demis Hassabis,et al.  Mastering the game of Go with deep neural networks and tree search , 2016, Nature.

[42]  Saul A. Kripke,et al.  Wittgenstein on Rules and Private Language. , 1985 .

[43]  SchmidhuberJürgen Deep learning in neural networks , 2015 .