Defaults and Relevance in Model-Based Reasoning

Reasoning with model-based representations is an intuitive paradigm, which has been shown to be theoretically sound and to possess some computational advantages over reasoning with formula-based representations of knowledge. In this paper we present more evidence to the value of such representations. Our results hinge on the notion of relevance, and model based representations are shown to be useful in capturing relevant information, and in allowing to ignore irrelevant information. In particular, we consider situations where context-specific information is used in the process of reasoning. We show that reasoning with model-based representations can be done efficiently in the presence of varying context information. We then consider the task of default reasoning. We show that default reasoning is a generalization of reasoning within context, in which the reasoner has many ``context" rules, which may be conflicting. We develop model-based algorithms that handle efficiently fragments of Reiter''s default logic. Our intuition about relevance is best captured in the model for reasoning within context, where model-based representations enable us to filter out irrelevant information. Interestingly, default logic is somewhat in contrast with our intuition about relevance. Default rules do not tell us explicitly what the context information is. Instead, we have to figure out what are the possible ``extensions" and then use those as possible contexts. As we show, model-based representations capture all possible extensions in an accessible form, thereby supporting efficient default reasoning whenever possible. Lastly, we argue that these results support an incremental view of reasoning in a natural way. We discuss the Learning to Reason framework, which emphasizes this view, and the notion of relevance as manifested in it. In particular, we discuss results on Learning to Reason in which model-based representations are used to represent relevant knowledge.

[1]  Raymond Reiter,et al.  A Theory of Diagnosis from First Principles , 1986, Artif. Intell..

[2]  David S. Touretzky,et al.  A Clash of Intuitions: The Current State of Nonmonotonic Multiple Inheritance Systems , 1987, IJCAI.

[3]  BeeriCatriel,et al.  On the Structure of Armstrong Relations for Functional Dependencies , 1984 .

[4]  Marvin Minsky,et al.  A framework for representing knowledge" in the psychology of computer vision , 1975 .

[5]  Bart Selman,et al.  Knowledge compilation and theory approximation , 1996, JACM.

[6]  Bart Selman,et al.  Knowledge Compilation using Horn Approximations , 1991, AAAI.

[7]  Hector J. Levesque,et al.  Hard problems for simple default logics , 1992 .

[8]  B. Selman Tractable default reasoning , 1991 .

[9]  Grigoris Antoniou,et al.  Nonmonotonic reasoning , 1997 .

[10]  S. Kosslyn Image and mind , 1982 .

[11]  Hector J. Levesque,et al.  Abductive and Default Reasoning: A Computational Core , 1990, AAAI.

[12]  Heikki Mannila,et al.  Design by Example: An Application of Armstrong Relations , 1986, J. Comput. Syst. Sci..

[13]  Dan Roth,et al.  Reasoning with Models , 1994, Artif. Intell..

[14]  Raymond Reiter,et al.  A Logic for Default Reasoning , 1987, Artif. Intell..

[15]  Dale Schuurmans,et al.  Learning Default Concepts , 1994 .

[16]  Noga Alon,et al.  Construction Of Asymptotically Good Low-rate Error-correcting Codes Through Pseudo-random Graphs , 1991, Proceedings. 1991 IEEE International Symposium on Information Theory.

[17]  Marvin Minsky,et al.  A framework for representing knowledge , 1974 .

[18]  Leslie G. Valiant,et al.  A theory of the learnable , 1984, STOC '84.

[19]  Nader H. Bshouty,et al.  Exact learning via the Monotone theory , 1993, Proceedings of 1993 IEEE 34th Annual Foundations of Computer Science.

[20]  C.H. Papadimitriou,et al.  On selecting a satisfying truth assignment , 1991, [1991] Proceedings 32nd Annual Symposium of Foundations of Computer Science.

[21]  Noga Alon,et al.  Construction of asymptotically good low-rate error-correcting codes through pseudo-random graphs , 1992, IEEE Trans. Inf. Theory.

[22]  John McCarthy,et al.  SOME PHILOSOPHICAL PROBLEMS FROM THE STANDPOINT OF ARTI CIAL INTELLIGENCE , 1987 .

[23]  Dan Roth,et al.  Learning to Reason: The Non-Monotonic Case , 1995, IJCAI.

[24]  Bart Selman,et al.  A General Framework for Knowledge Compilation , 1991, PDK.

[25]  Patrick Henry Winston,et al.  The psychology of computer vision , 1976, Pattern Recognit..

[26]  Bart Selman,et al.  Horn Approximations of Empirical Data , 1995, Artif. Intell..

[27]  Heikki Mannila,et al.  Reasoning with examples: propositional formulae and database dependencies , 1999, Acta Informatica.

[28]  Dan Roth,et al.  Learning to reason , 1994, JACM.

[29]  Dan Roth,et al.  On the Hardness of Approximate Reasoning , 1993, IJCAI.

[30]  Moni Naor,et al.  Small-Bias Probability Spaces: Efficient Constructions and Applications , 1993, SIAM J. Comput..

[31]  Hector J. Levesque,et al.  Making Believers out of Computers , 1986, Artif. Intell..

[32]  Judea Pearl,et al.  Probabilistic reasoning in intelligent systems - networks of plausible inference , 1991, Morgan Kaufmann series in representation and reasoning.

[33]  Nader H. Bshouty Exact Learning Boolean Function via the Monotone Theory , 1995, Inf. Comput..

[34]  Bart Selman,et al.  Model-Preference Default Theories , 1990, Artif. Intell..