Exact Identi cation of Read-once Formulas Using Fixed Points ofAmpli cation

In this paper we describe a new technique for exactly identifying certain classes of read-once Boolean formulas. The method is based on sampling the input-output behavior of the target formula on a probability distribution that is determined by the xed point of the formula's ampli cation function (de ned as the probability that a 1 is output by the formula when each input bit is 1 independently with probability p). By performing various statistical tests on easily sampled variants of the xedpoint distribution, we are able to e ciently infer all structural information about any logarithmic-depth formula (with high probability). We apply our results to prove the existence of short universal identi cation sequences for large classes of formulas. We also describe extensions of our algorithms to handle high rates of noise, and to learn formulas of unbounded depth in Valiant's model with respect to speci c distributions. Most of this research was carried out while all three authors were at MIT Laboratory for Computer Science with support provided by ARO Grant DAAL03-86-K-0171, DARPA Contract N00014-89-J-1988, NSF Grant CCR-88914428, and a grant from the Siemens Corporation. R. Schapire received additional support from AFOSR Grant 89-0506 while at Harvard University. S. Goldman is currently supported in part by a G.E. Foundation Junior Faculty Grant and NSF Grant CCR-9110108.

[1]  Ming Li,et al.  Learning in the presence of malicious errors , 1993, STOC '88.

[2]  Marek Karpinski,et al.  Learning read-once formulas with queries , 1993, JACM.

[3]  Marek Karpinski,et al.  Read-Once Threshold Formulas, Justifying Assignments, and Generic Transformations , 1991 .

[4]  Michael Kearns,et al.  On the complexity of teaching , 1991, COLT '91.

[5]  Yishay Mansour,et al.  Learning monotone ku DNF formulas on product distributions , 1991, COLT '91.

[6]  Lisa Hellerstein,et al.  Learning read-once formulas over fields and extended bases , 1991, COLT '91.

[7]  Leonard Pitt,et al.  Prediction-Preserving Reducibility , 1990, J. Comput. Syst. Sci..

[8]  Michael Kearns,et al.  Computational complexity of machine learning , 1990, ACM distinguished dissertations.

[9]  Karsten A. Verbeurgt Learning DNF under the uniform distribution in quasi-polynomial time , 1990, COLT '90.

[10]  Thomas R. Hancock,et al.  Identifying μ-formula decision trees with queries , 1990, COLT '90.

[11]  Noam Nisan,et al.  Constant depth circuits, Fourier transform, and learnability , 1989, 30th Annual Symposium on Foundations of Computer Science.

[12]  M. Kearns,et al.  Crytographic limitations on learning Boolean formulae and finite automata , 1989, STOC '89.

[13]  Robert H. Sloan,et al.  Corrigendum to types of noise in data for concept learning , 1988, COLT '92.

[14]  Alon Itai,et al.  Learnability by fixed distributions , 1988, COLT '88.

[15]  Leslie G. Valiant,et al.  On the learnability of Boolean formulae , 1987, STOC.

[16]  Ravi B. Boppana,et al.  Amplification of probabilistic boolean formulas , 1985, 26th Annual Symposium on Foundations of Computer Science (sfcs 1985).

[17]  L. Valiant,et al.  A theory of the learnable , 1984, CACM.

[18]  W. Hoeffding Probability inequalities for sum of bounded random variables , 1963 .

[19]  Robert E. Schapire,et al.  Learning probabilistic read-once formulas on product distributions , 1991, COLT '91.

[20]  B. Ravi Lower bounds for monotone circuits and formulas , 1986 .