Incremental Concept Learning for Bounded Data Mining

Important refinements of concept learning in the limit from positive data considerably restricting the accessibility of input data are studied. Let c be any concept; every infinite sequence of elements exhausting c is called positive presentation of c. In all learning models considered the learning machine computes a sequence of hypotheses about the target concept from a positive presentation of it. With iterative learning, the learning machine, in making a conjecture, has access to its previous conjecture and the latest data items coming in. In k-bounded example-memory inference (k is a priori fixed) the learner is allowed to access, in making a conjecture, its previous hypothesis, its memory of up to k data items it has already seen, and the next element coming in. In the case of k-feedback identification, the learning machine, in making a conjecture, has access to its previous conjecture, the latest data item coming in, and, on the basis of this information, it can compute k items and query the database of previous data to find out, for each of the k items, whether or not it is in the database (k is again a priori fixed). In all cases, the sequence of conjectures has to converge to a hypothesis correctly describing the target concept. Our results are manyfold. An infinite hierarchy of more and more powerful feedback learners in dependence on the number k of queries allowed to be asked is established. However, the hierarchy collapses to 1-feedback inference if only indexed families of infinite concepts are considered, and moreover, its learning power is then equal to learning in the limit. But it remains infinite for concept classes of only infinite r.e. concepts. Both k-feedback inference and k-bounded example-memory identification are more powerful than iterative learning but incomparable to one another. Furthermore, there are cases where redundancy in the hypothesis space is shown to be a resource increasing the learning power of iterative learners. Finally, the union of at most k pattern languages is shown to be iteratively inferable.

[1]  Keith Wright Identification of unions of languages drawn from an identifiable class , 1989, COLT '89.

[2]  Patrick Brézillon,et al.  Lecture Notes in Artificial Intelligence , 1999 .

[3]  Setsuo Arikawa,et al.  Pattern Inference , 1995, GOSLER Final Report.

[4]  James B. Morris Formal Languages and their Relation to Automata , 1970 .

[5]  Dana Angluin,et al.  Finding Patterns Common to a Set of Strings , 1980, J. Comput. Syst. Sci..

[6]  M. R. K. Krishna Rao A Class of Prolog Programs Inferable from Positive Data , 1996, ALT.

[7]  Léa Meyer Probabilistic Language Learning Under Monotonicity Constraints , 1995, ALT.

[8]  Thomas Zeugmann,et al.  Monotonic Versus Nonmonotonic Language Learning , 1991, Nonmonotonic and Inductive Logic.

[9]  Usama M. Fayyad,et al.  Automating the Analysis and Cataloging of Sky Surveys , 1996, Advances in Knowledge Discovery and Data Mining.

[10]  Gregory Piatetsky-Shapiro,et al.  Selecting and reporting What Is Interesting , 1996, Advances in Knowledge Discovery and Data Mining.

[11]  Ronald J. Brachman,et al.  The Process of Knowledge Discovery in Databases , 1996, Advances in Knowledge Discovery and Data Mining.

[12]  Arun Sharma,et al.  On Monotonic Strategies for Learning r.e. Languages , 1994, AII/ALT.

[13]  Hiroki Arimura,et al.  Inductive Inference of Unbounded Unions of Pattern Languages from Positive Data , 1996, ALT.

[14]  Thomas Zeugmann,et al.  Characterizations of Monotonic and Dual Monotonic Language Learning , 1995, Inf. Comput..

[15]  Robert Nix,et al.  Editing by example , 1985, POPL '84.

[16]  Thomas Zeugmann,et al.  Learning k-Variable Pattern Languages Efficiently Stochastically Finite on Average from Positive Data , 1998, ICGI.

[17]  Ayumi Shinohara,et al.  Knowledge Acquisition from Amino Acid Sequences by Machine Learning System BONSAI , 1992 .

[18]  Padhraic Smyth,et al.  From Data Mining to Knowledge Discovery: An Overview , 1996, Advances in Knowledge Discovery and Data Mining.

[19]  Hiroki Arimura,et al.  Inductive Inference of Prolog Programs with Linear Data Dependency from Positive Data , 1993 .

[20]  E. Mark Gold,et al.  Language Identification in the Limit , 1967, Inf. Control..

[21]  Arun Sharma,et al.  On Identification by Teams and Probabilistic Machines , 1995, GOSLER Final Report.

[22]  Akihiro Yamamoto,et al.  Learning Elementary Formal Systems , 1992, Theor. Comput. Sci..

[23]  Arto Salomaa,et al.  Return to Patterns , 1995, Bull. EATCS.

[24]  John Case,et al.  Infinitary self-reference in learning theory , 1994, J. Exp. Theor. Artif. Intell..

[25]  John Case The power of vacillation , 1988, COLT '88.

[26]  Thomas Zeugmann,et al.  A Guided Tour Across the Boundaries of Learning Recursive Languages , 1995, GOSLER Final Report.

[27]  Rolf Wiehagen,et al.  Ignoring data may be the only way to learn efficiently , 1994, J. Exp. Theor. Artif. Intell..

[28]  Dana Angluin,et al.  Inductive Inference of Formal Languages from Positive Data , 1980, Inf. Control..

[29]  Jr. Hartley Rogers Theory of Recursive Functions and Effective Computability , 1969 .

[30]  John Case,et al.  Comparison of Identification Criteria for Machine Inductive Inference , 1983, Theor. Comput. Sci..

[31]  Rolf Wiehagen Limes-Erkennung rekursiver Funktionen durch spezielle Strategien , 1975, J. Inf. Process. Cybern..

[32]  Hiroki Arimura,et al.  Inductive inference of unbounded unions of pattern languages from positive data , 2000, Theor. Comput. Sci..

[33]  Setsuo Arikawa,et al.  Proceedings of the 7th International Workshop on Algorithmic Learning Theory , 1996 .

[34]  Manuel Blum,et al.  Toward a Mathematical Theory of Inductive Inference , 1975, Inf. Control..

[35]  Jeffrey D. Ullman,et al.  Formal languages and their relation to automata , 1969, Addison-Wesley series in computer science and information processing.

[36]  Carl H. Smith,et al.  On the Role of Procrastination in Machine Learning , 1993, Inf. Comput..

[37]  Tao Jiang,et al.  Inclusion is Undecidable for Pattern Languages , 1993, ICALP.

[38]  Thomas Zeugmann,et al.  Learning Recursive Languages with Bounded Mind Changes , 1993, Int. J. Found. Comput. Sci..

[39]  T. Shinohara INFERRING UNIONS OF TWO PATTERN LANGUAGES , 1983 .

[40]  Thomas Zeugmann,et al.  Language learning in dependence on the space of hypotheses , 1993, COLT '93.

[41]  Manuel Blum,et al.  A Machine-Independent Theory of the Complexity of Recursive Functions , 1967, JACM.

[42]  Léa Meyer Monotonic and Dual-Monotonic Probabilistic Language Learning of Indexed Families with High Probability , 1997, EuroCOLT.

[43]  Robert Cori,et al.  Proceedings of the 6th Annual Symposium on Theoretical Aspects of Computer Science on STACS 89 , 1988 .

[44]  Léa Meyer,et al.  Probabilistic Language Learning Under Monotonicity Constraints , 1995, Theor. Comput. Sci..

[45]  Sanjay Jain,et al.  Open Problems in "Systems That Learn" , 1994, J. Comput. Syst. Sci..

[46]  John Case,et al.  The Power of Vacillation in Language Learning , 1999, SIAM J. Comput..

[47]  Daniel N. Osherson,et al.  Systems That Learn: An Introduction to Learning Theory for Cognitive and Computer Scientists , 1990 .

[48]  Carl H. Smith,et al.  On the role of procrastination for machine learning , 1992, COLT '92.

[49]  R. Smullyan Theory of formal systems , 1962 .

[50]  Gilberto Filé The Relation of Two Patterns with Comparable Languages , 1988, STACS.

[51]  Thomas Zeugmann,et al.  Incremental Learning from Positive Data , 1996, J. Comput. Syst. Sci..