When Prior Knowledge Hinders Learning

There are two general approaches to learning classification rules. Empirical learning programs operate by finding regularities among a group of training examples. Analytic learning systems use prior knowledge to explain the classification of examples and form a general description of the class of examples with the same explanation. 1989) have combined empirical and analytic learning methods. In such systems, the prior knowledge is used to bias the learner to prefer concepts that are consistent with the prior knowledge. This bias is intended to make a learner create concept descriptions that are more accurate classifiers than both the original prior knowledge (that serves as input to the analytic learning component) and the rules that would arise if only the empirical learning component were used. In previous work (Pazzani, Brunk, & Silverstein, 1991; Pazzani & Kibler, in press), we have described FOCL, a system that extends Quinlan's (1990) FOIL program in a number of ways, most significantly by adding a compatible explanation-based learning (EBL) component. Here, we describe a simple example deliberately constructed to show that prior knowledge can also hinder learning in FOCL as well as other combined learned systems. We analyze why this example presents problems for these systems and suggest an approach that allows FOCL to exploit this type of prior knowledge to facilitate learning. 1.1 FOIL FOIL learns classification rules by constructing a set of Horn Clauses in terms of a set of known operational predicates. Each clause body consists of a conjunction of literals that cover some positive and no negative examples. FOIL starts to learn a clause body by finding the literal with the maximum information gain 1 , and continues to add literals to the clause body until the clause does not cover any negative examples. After learning each clause, FOIL removes the positive examples covered by that clause from further consideration. The learning process terminates when all positive examples have been covered by a clause. 1.2 FOCL FOCL extends FOIL by including a compatible EBL component. This allows FOCL to take advantage of a given domain theory. When constructing a clause body, there are two ways that FOCL can add literals. First, it can create literals via the same empirical method used by FOIL. Second, it can create literals by operationalizing a target concept, i.e., a non-operational definition of the concept to be learned (Mitchell, Keller, & Kedar-Cabelli, 1986). FOCL uses the information-based evaluation …

[1]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .

[2]  Michael Lebowitz,et al.  Integrated Learning: Controlling Explanation , 1986, Cogn. Sci..

[3]  Ryszard S. Michalski,et al.  On the Nature of Explanation or Why Did the Wine Bottle Shatter , 1988 .

[4]  Francesco Bergadano,et al.  A Knowledge Intensive Approach to Concept Induction , 1988, ML Workshop.

[5]  Jude W. Shavlik,et al.  Combining Explanation-Based Learning and Artificial Neural Networks , 1989, ML.

[6]  David C. Wilkins,et al.  Knowledge Base Refinement as Improving an Incorrect, Inconsistent and Incomplete Domain Theory , 1989, ML.

[7]  Haym Hirsh Combining Empirical and Analytical Learning with Version Spaces , 1989, ML.

[8]  Raymond J. Mooney,et al.  Changing the Rules: A Comprehensive Approach to Theory Refinement , 1990, AAAI.

[9]  Michael J. Pazzani,et al.  Creating a memory of causal relationships - an integration of empirical and explanation-based learning methods , 1990 .

[10]  Jude W. Shavlik,et al.  Interpretation of Artificial Neural Networks: Mapping Knowledge-Based Neural Networks into Rules , 1991, NIPS.

[11]  Michael J. Pazzani,et al.  A Knowledge-intensive Approach to Learning Relational Concepts , 1991, ML.

[12]  William W. Cohen Compiling prior knowledge into an explicit basis , 1992, ICML 1992.

[13]  Geoffrey G. Towell,et al.  Symbolic knowledge and neural networks: insertion, refinement and extraction , 1992 .

[14]  William W. Cohen Abductive explanation-based learning: A solution to the multiple inconsistent explanation problem , 1992, Machine Learning.

[15]  Thomas G. Dietterich,et al.  A Study of Explanation-Based Methods for Inductive Learning , 1989, Machine Learning.

[16]  J. Ross Quinlan,et al.  Induction of Decision Trees , 1986, Machine Learning.

[17]  J. Ross Quinlan,et al.  Learning logical definitions from relations , 1990, Machine Learning.

[18]  D. Haussler,et al.  Boolean Feature Discovery in Empirical Learning , 1990, Machine Learning.