On the Learnability of Causal Domains: Inferring Temporal Reality from Appearances

We examine the feasibility of learning causal domains by observing transitions between states as a result of taking certain actions. We take the approach that the observed transitions are only a macro-level manifestation of the underlying microlevel dynamics of the environment, which an agent does not directly observe. In this setting, we ask that domains learned through macro-level state transitions are accompanied by formal guarantees on their predictive power on future instances. We show that even if the underlying dynamics of the environment are significantly restricted, and even if the learnability requirements are severely relaxed, it is still intractable for an agent to learn a model of its environment. Our negative results are universal in that they apply independently of the syntax and semantics of the framework the agent utilizes as its modelling tool. We close with a discussion of what a complete theory for domain learning should take into account, and how existing work can be utilized to this effect.

[1]  Loizos Michael,et al.  Learning from Partial Observations , 2007, IJCAI.

[2]  Ramón P. Otero,et al.  Induction of the Indirect Effects of Actions by Monotonic Methods , 2005, ILP.

[3]  Enrico Giunchiglia,et al.  Nonmonotonic causal theories , 2004, Artif. Intell..

[4]  Ming Li,et al.  Learning in the presence of malicious errors , 1993, STOC '88.

[5]  Nick Littlestone,et al.  From on-line to batch learning , 1989, COLT '89.

[6]  Patrick Doherty,et al.  TAL: Temporal Action Logics Language Specification and Tutorial , 1998, Electron. Trans. Artif. Intell..

[7]  Leslie G. Valiant,et al.  A theory of the learnable , 1984, STOC '84.

[8]  Antonis C. Kakas,et al.  Modular-epsilon: An Elaboration Tolerant Approach to the Ramification and Qualification Problems , 2005, LPNMR.

[9]  Michael Kearns,et al.  Efficient noise-tolerant learning from statistical queries , 1993, STOC.

[10]  Murray Shanahan,et al.  Some Alternative Formulations of the Event Calculus , 2002, Computational Logic: Logic Programming and Beyond.

[11]  Sally A. Goldman,et al.  Teaching a Smarter Learner , 1996, J. Comput. Syst. Sci..

[12]  Umesh V. Vazirani,et al.  A Markovian extension of Valiant's learning model , 1990, Proceedings [1990] 31st Annual Symposium on Foundations of Computer Science.

[13]  Michael Thielscher,et al.  Introduction to the Fluent Calculus , 1998, Electron. Trans. Artif. Intell..

[14]  J. Petley Appearance and Reality , 2006 .

[15]  Katsumi Inoue,et al.  Inducing Causal Laws by Regular Inference , 2005, ILP.

[16]  John McCarthy,et al.  SOME PHILOSOPHICAL PROBLEMS FROM THE STANDPOINT OF ARTI CIAL INTELLIGENCE , 1987 .

[17]  Dana Angluin,et al.  Queries and concept learning , 1988, Machine Learning.

[18]  Adi Shamir,et al.  A method for obtaining digital signatures and public-key cryptosystems , 1978, CACM.

[19]  A. Kakas,et al.  Modular-E: an Elaboration Tolerant Approach to the Ramification and Qualification Problems - Preliminary Report , 2005 .

[20]  Michael Kharitonov,et al.  Cryptographic hardness of distribution-specific learning , 1993, STOC.

[21]  F. H. Bradley,et al.  Appearance and Reality , 1893 .

[22]  N. Littlestone Learning Quickly When Irrelevant Attributes Abound: A New Linear-Threshold Algorithm , 1987, 28th Annual Symposium on Foundations of Computer Science (sfcs 1987).

[23]  Umesh V. Vazirani,et al.  An Introduction to Computational Learning Theory , 1994 .

[24]  Michael Gelfond,et al.  Representing Actions in Extended Logic Programming , 1992, JICSLP.

[25]  Leslie G. Valiant,et al.  Cryptographic Limitations on Learning Boolean Formulae and Finite Automata , 1993, Machine Learning: From Theory to Applications.

[26]  Jerzy Tiuryn,et al.  Dynamic logic , 2001, SIGA.

[27]  D. Gabbay,et al.  Handbook of Philosophical Logic, Volume II. Extensions of Classical Logic , 1986 .