Honest Compressions and Their Application to Compression Schemes

The existence of a compression scheme for every concept class with bounded VC-dimension is one of the oldest open problems in statistical learning theory. Here we demonstrate the existence of such compression schemes under stronger assumptions than nite VCdimension. Specically, for each concept class we associate a family of concept classes that we call the alternating concept classes. Under the assumption that these concept classes have bounded VC-dimension, we prove existence of a compression scheme. This result is motivated by recent progress in the eld of model theory with respect to an analogues problem. In fact, our proof can be considered as a constructive proof of these advancements. This means that we describe the reconstruction function explicitly. Not less important, the theorems and proofs we present are in purely combinatorial terms and are available to the reader who is unfamiliar with model theory. Also, using tools from model theory, we apply our results and prove existence of compression schemes in interesting cases such as concept classes dened by hyperplanes, polynomials, exponentials, restricted analytic functions and compositions, additions and multiplications of all of the above.

[1]  Chris Miller,et al.  Lecture notes on o-minimal structures and real analytic geometry , 2012 .

[2]  A. Chernikov,et al.  Externally definable sets and dependent pairs , 2010, Israel Journal of Mathematics.

[3]  Benjamin I. P. Rubinstein,et al.  A Geometric Approach to Sample Compression , 2009, J. Mach. Learn. Res..

[4]  Michael C. Laskowski,et al.  Compression Schemes, Stable Definable Families, and o-Minimal Structures , 2010, Discret. Comput. Geom..

[5]  Manfred K. Warmuth,et al.  Sample compression, learnability, and the Vapnik-Chervonenkis dimension , 1995, Machine Learning.

[6]  Vincent Guingona On uniform definability of types over finite sets , 2012, J. Symb. Log..

[7]  Jirí Matousek,et al.  Bounded VC-Dimension Implies a Fractional Helly Theorem , 2004, Discret. Comput. Geom..

[8]  Pierre Simon Lecture notes on NIP theories , 2012 .

[9]  Marek Karpinski,et al.  Polynomial bounds for VC dimension of sigmoidal neural networks , 1995, STOC '95.

[10]  Hans Adler,et al.  Introduction to theories without the independence property , 2008 .

[11]  S. Shelah Stability, the f.c.p., and superstability; model theoretic properties of formulas in first order theory , 1971 .

[12]  Patrick Speissegger Pfaffian Sets and O-minimality , 2012 .

[13]  Yoav Freund,et al.  Boosting a weak learning algorithm by majority , 1990, COLT '90.

[14]  Shai Ben-David,et al.  Combinatorial Variability of Vapnik-chervonenkis Classes with Applications to Sample Compression Schemes , 1998, Discret. Appl. Math..

[15]  Vladimir Vapnik,et al.  Chervonenkis: On the uniform convergence of relative frequencies of events to their probabilities , 1971 .

[16]  Peter L. Bartlett,et al.  Shifting: One-inclusion mistake bounds and sample compression , 2009, J. Comput. Syst. Sci..

[17]  Michael C. Laskowski,et al.  Vapnik-Chervonenkis classes of definable sets , 1992 .

[18]  Manfred K. Warmuth,et al.  Unlabeled Compression Schemes for Maximum Classes, , 2007, COLT.

[19]  Manfred K. Warmuth,et al.  Relating Data Compression and Learnability , 2003 .

[20]  S. Ben-David,et al.  Combinatorial Variability of Vapnik-chervonenkis Classes with Applications to Sample Compression Schemes , 1998, Discrete Applied Mathematics.

[21]  Pierre Simon,et al.  Externally definable sets and dependent pairs II , 2012, 1202.2650.

[22]  N. Alon,et al.  Piercing convex sets and the hadwiger-debrunner (p , 1992 .

[23]  Eduardo D. Sontag,et al.  Finiteness results for sigmoidal “neural” networks , 1993, STOC.

[24]  Sally Floyd,et al.  Space-bounded learning and the Vapnik-Chervonenkis dimension , 1989, COLT '89.