On agnostic boosting and parity learning

The motivating problem is agnostically learning parity functions, i.e., parity with arbitrary or adversarial noise. Specifically, given random labeled examples from an *arbitrary* distribution, we would like to produce an hypothesis whose accuracy nearly matches the accuracy of the best parity function. Our algorithm runs in time 2O(n/log n), which matches the best known for the easier cases of learning parities with random classification noise (Blum et al, 2003) and for agnostically learning parities over the uniform distribution on inputs (Feldman et al, 2006). Our approach is as follows. We give an agnostic boosting theorem that is capable of nearly achieving optimal accuracy, improving upon earlier studies (starting with Ben David et al, 2001). To achieve this, we circumvent previous lower bounds by altering the boosting model. We then show that the (random noise) parity learning algorithm of Blum et al (2000) fits our new model of agnostic weak learner. Our agnostic boosting framework is completely general and may be applied to other agnostic learning problems. Hence, it also sheds light on the actual difficulty of agnostic learning by showing that full agnostic boosting is indeed possible.

[1]  Ari Juels,et al.  Authenticating Pervasive Devices with Human Protocols , 2005, CRYPTO.

[2]  Dmitry Gavinsky Optimally-smooth adaptive boosting and application to agnostic learning , 2003 .

[3]  Yishay Mansour,et al.  Boosting Using Branching Programs , 2000, J. Comput. Syst. Sci..

[4]  Vitaly Feldman,et al.  New Results for Learning Noisy Parities and Halfspaces , 2006, 2006 47th Annual IEEE Symposium on Foundations of Computer Science (FOCS'06).

[5]  Ravi Kumar,et al.  A sieve algorithm for the shortest lattice vector problem , 2001, STOC '01.

[6]  B. E. Eckbo,et al.  Appendix , 1826, Epilepsy Research.

[7]  Rocco A. Servedio,et al.  Agnostically learning halfspaces , 2005, 46th Annual IEEE Symposium on Foundations of Computer Science (FOCS'05).

[8]  Michael Kearns,et al.  Efficient noise-tolerant learning from statistical queries , 1993, STOC.

[9]  Elchanan Mossel,et al.  Learning nonsingular phylogenies and hidden Markov models , 2005, STOC '05.

[10]  Leslie G. Valiant,et al.  A theory of the learnable , 1984, STOC '84.

[11]  Richard J. Lipton,et al.  Cryptographic Primitives Based on Hard Learning Problems , 1993, CRYPTO.

[12]  Shai Ben-David,et al.  Agnostic Boosting , 2001, COLT/EuroCOLT.

[13]  Vadim Lyubashevsky,et al.  The Parity Problem in the Presence of Noise, Decoding Random Linear Codes, and the Subset Sum Problem , 2005, APPROX-RANDOM.

[14]  Rocco A. Servedio,et al.  Boosting in the presence of noise , 2003, STOC '03.

[15]  R. Schapire The Strength of Weak Learnability , 1990, Machine Learning.

[16]  David Haussler,et al.  Equivalence of models for polynomial learnability , 1988, COLT '88.

[17]  Manuel Blum,et al.  Secure Human Identification Protocols , 2001, ASIACRYPT.

[18]  Ronitt Rubinfeld,et al.  Learning polynomials with queries: The highly noisy case , 1995, Proceedings of IEEE 36th Annual Foundations of Computer Science.

[19]  R. Schapire,et al.  Toward efficient agnostic learning , 1992, COLT '92.

[20]  Oded Regev,et al.  On lattices, learning with errors, random linear codes, and cryptography , 2005, STOC '05.

[21]  Yishay Mansour,et al.  On the boosting ability of top-down decision tree learning algorithms , 1996, STOC '96.