Probabilistic Hill-climbing: Theory and Applications

Many learning systems search through a space of possible performance elements, seeking an element with high expected utility. As the task of nding the globally optimal element is usually intractable, many practical learning systems use hill-climbing to nd a local optimum. Unfortunately, even this is diicult, as it depends on the distribution of problems, which is typically unknown. This paper addresses the task of approximating this hill-climbing search when the utility function can only be estimated by sampling. We present an algorithm that returns an element that is, with provably high probability, essentially a local optimum. We then demonstrate the generality of this algorithm by sketching three meaningful applications, that respectively nd an element whose eeciency, accuracy or completeness is nearly optimal. These results suggest approaches to solving the utility problem from explanation-based learning, the multiple extension problem from nonmonotonic reasoning and the tractability/completeness tradeoo problem from knowledge representation.

[1]  Herbert A. Simon,et al.  Optimal Problem-Solving Search: All-Oor-None Solutions , 1975, Artif. Intell..

[2]  William F. Clocksin,et al.  Programming in Prolog , 1987, Springer Berlin Heidelberg.

[3]  Leslie G. Valiant,et al.  A theory of the learnable , 1984, STOC '84.

[4]  Jean H. Gallier,et al.  Linear-Time Algorithms for Testing the Satisfiability of Propositional Horn Formulae , 1984, J. Log. Program..

[5]  Paul E. Utgoff,et al.  Shift of bias for inductive concept learning , 1984 .

[6]  Tom M. Mitchell,et al.  LEAP: A Learning Apprentice for VLSI Design , 1985, IJCAI.

[7]  Drew McDermott,et al.  Default Reasoning, Nonmonotonic Logics, and the Frame Problem , 1986, AAAI.

[8]  Michael R. Genesereth,et al.  Logical foundations of artificial intelligence , 1987 .

[9]  Benjamin N. Grosof,et al.  A Declarative Approach to Bias in Concept Learning , 1987, AAAI.

[10]  Paul Morris,et al.  Curing Anomalous Extensions , 1987, AAAI.

[11]  Allen Newell,et al.  SOAR: An Architecture for General Intelligence , 1987, Artif. Intell..

[12]  Randy Goebel,et al.  Theorist: A Logical Reasoning System for Defaults and Diagnosis , 1987 .

[13]  Mark S. Boddy,et al.  An Analysis of Time-Dependent Planning , 1988, AAAI.

[14]  David Haussler,et al.  Quantifying Inductive Bias: AI Learning Algorithms and Valiant's Learning Framework , 1988, Artif. Intell..

[15]  Richard M. Keller,et al.  Defining Operationality for Explanation-Based Learning , 1987, Artificial Intelligence.

[16]  Gerhard Brewka,et al.  Preferred Subtheories: An Extended Logical Framework for Default Reasoning , 1989, IJCAI.

[17]  Geoffrey E. Hinton Connectionist Learning Procedures , 1989, Artif. Intell..

[18]  Lokendra Shastri,et al.  Default Reasoning in Semantic Networks: A Formalization of Recognition and Inheritance , 1989, Artificial Intelligence.

[19]  Mark S. Boddy,et al.  Solving Time-Dependent Planning Problems , 1989, IJCAI.

[20]  Pekka Orponen,et al.  On the sample complexity of finding good search strategies , 1990, COLT '90.

[21]  Paul van Arragon Nested default reasoning with priority levels , 1990 .

[22]  Gerald DeJong,et al.  A Hybrid Approach to Guaranteed Effective Control Strategies , 1991, ML.

[23]  Bart Selman,et al.  Knowledge Compilation using Horn Approximations , 1991, AAAI.

[24]  Benjamin N. Grosof,et al.  Generalizing Prioritization , 1991, KR.

[25]  Russell Greiner,et al.  Measuring and Improving the Effectiveness of Representations , 1991, IJCAI.

[26]  Pekka Orponen,et al.  Probably Approximately Optimal Derivation Strategies , 1991, KR.

[27]  Russell Greiner,et al.  A Statistical Approach to Solving the EBL Utility Problem , 1992, AAAI.

[28]  David Haussler,et al.  Decision Theoretic Generalizations of the PAC Model for Neural Net and Other Learning Applications , 1992, Inf. Comput..