Closure Properties for Private Classification and Online Prediction

Let~$\cH$ be a class of boolean functions and consider a {\it composed class} $\cH'$ that is derived from~$\cH$ using some arbitrary aggregation rule (for example, $\cH'$ may be the class of all 3-wise majority-votes of functions in $\cH$). We upper bound the Littlestone dimension of~$\cH'$ in terms of that of~$\cH$. As a corollary, we derive closure properties for online learning and private PAC learning. The derived bounds on the Littlestone dimension exhibit an undesirable exponential dependence. For private learning, we prove close to optimal bounds that circumvents this suboptimal dependency. The improved bounds on the sample complexity of private learning are derived algorithmically via transforming a private learner for the original class $\cH$ to a private learner for the composed class~$\cH'$. Using the same ideas we show that any ({\em proper or improper}) private algorithm that learns a class of functions $\cH$ in the realizable case (i.e., when the examples are labeled by some function in the class) can be transformed to a private algorithm that learns the class $\cH$ in the agnostic case.

[1]  J. Freitag,et al.  MODEL THEORY AND COMBINATORICS OF BANNED SEQUENCES , 2018, The Journal of Symbolic Logic.

[2]  Peter L. Bartlett,et al.  Neural Network Learning - Theoretical Foundations , 1999 .

[3]  Frank Plumpton Ramsey,et al.  On a Problem of Formal Logic , 1930 .

[4]  Haim Kaplan,et al.  Private Learning of Halfspaces: Simplifying the Construction and Reducing the Sample Complexity , 2020, NeurIPS.

[5]  Shai Ben-David,et al.  Agnostic Online Learning , 2009, COLT.

[6]  R. Dudley Central Limit Theorems for Empirical Measures , 1978 .

[7]  Wilfrid Hodges,et al.  A Shorter Model Theory , 1997 .

[8]  Vladimir Vapnik,et al.  Chervonenkis: On the uniform convergence of relative frequencies of events to their probabilities , 1971 .

[9]  Roi Livni,et al.  An Equivalence Between Private Classification and Online Prediction , 2020, 2020 IEEE 61st Annual Symposium on Foundations of Computer Science (FOCS).

[10]  Cynthia Dwork,et al.  Calibrating Noise to Sensitivity in Private Data Analysis , 2006, TCC.

[11]  Ambuj Tewari,et al.  Online Learning via Differential Privacy , 2017, ArXiv.

[12]  Toniann Pitassi,et al.  The reusable holdout: Preserving validity in adaptive data analysis , 2015, Science.

[13]  Moni Naor,et al.  Our Data, Ourselves: Privacy Via Distributed Noise Generation , 2006, EUROCRYPT.

[14]  David Haussler,et al.  Learnability and the Vapnik-Chervonenkis dimension , 1989, JACM.

[15]  Amos Beimel,et al.  Bounds on the sample complexity for private learning and private data release , 2010, Machine Learning.

[16]  Noga Alon,et al.  Private PAC learning implies finite Littlestone dimension , 2018, STOC.

[17]  Raef Bassily,et al.  Algorithmic stability for adaptive data analysis , 2015, STOC.

[18]  Shay Moran,et al.  Online Agnostic Boosting via Regret Minimization , 2020, NeurIPS.

[19]  Sofya Raskhodnikova,et al.  What Can We Learn Privately? , 2008, 2008 49th Annual IEEE Symposium on Foundations of Computer Science.

[20]  Kunal Talwar,et al.  Mechanism Design via Differential Privacy , 2007, 48th Annual IEEE Symposium on Foundations of Computer Science (FOCS'07).

[21]  R. Greenwood,et al.  Combinatorial Relations and Chromatic Graphs , 1955, Canadian Journal of Mathematics.

[22]  Kobbi Nissim,et al.  Differentially Private Release and Learning of Threshold Functions , 2015, 2015 IEEE 56th Annual Symposium on Foundations of Computer Science.

[23]  Shay Moran,et al.  Private Learning Implies Online Learning: An Efficient Reduction , 2019, NeurIPS.

[24]  Siddharth Bhaskar THICKET DENSITY , 2021, The Journal of Symbolic Logic.

[25]  Manfred K. Warmuth,et al.  The weighted majority algorithm , 1989, 30th Annual Symposium on Foundations of Computer Science.

[26]  Guy N. Rothblum,et al.  Boosting and Differential Privacy , 2010, 2010 IEEE 51st Annual Symposium on Foundations of Computer Science.

[27]  Thomas Steinke,et al.  Generalization for Adaptively-chosen Estimators via Stable Median , 2017, COLT.

[28]  W. Hodges CLASSIFICATION THEORY AND THE NUMBER OF NON‐ISOMORPHIC MODELS , 1980 .

[29]  Seth Neel,et al.  The Role of Interactivity in Local Differential Privacy , 2019, 2019 IEEE 60th Annual Symposium on Foundations of Computer Science (FOCS).

[30]  N. Littlestone Learning Quickly When Irrelevant Attributes Abound: A New Linear-Threshold Algorithm , 1987, 28th Annual Symposium on Foundations of Computer Science (sfcs 1987).

[31]  Amos Beimel,et al.  Learning Privately with Labeled and Unlabeled Examples , 2014, Algorithmica.

[32]  John Shawe-Taylor,et al.  A Result of Vapnik with Applications , 1993, Discret. Appl. Math..

[33]  Hunter Chase,et al.  Model Theory and Machine Learning , 2019, Bull. Symb. Log..

[34]  Christopher Jung,et al.  A new analysis of differential privacy’s generalization guarantees (invited paper) , 2019, ITCS.

[35]  Haim Kaplan,et al.  Privately Learning Thresholds: Closing the Exponential Gap , 2019, COLT.

[36]  Aaron Roth,et al.  Max-Information, Differential Privacy, and Post-selection Hypothesis Testing , 2016, 2016 IEEE 57th Annual Symposium on Foundations of Computer Science (FOCS).

[37]  Haipeng Luo,et al.  Optimal and Adaptive Algorithms for Online Boosting , 2015, ICML.

[38]  F ROSENBLATT,et al.  The perceptron: a probabilistic model for information storage and organization in the brain. , 1958, Psychological review.

[39]  Shay Moran,et al.  Private Center Points and Learning of Halfspaces , 2019, COLT.

[40]  Hsuan-Tien Lin,et al.  An Online Boosting Algorithm with Theoretical Justifications , 2012, ICML.

[41]  Amos Beimel,et al.  Characterizing the Sample Complexity of Pure Private Learners , 2019, J. Mach. Learn. Res..

[42]  Leslie G. Valiant,et al.  A theory of the learnable , 1984, STOC '84.

[43]  Norbert Sauer,et al.  On the Density of Families of Sets , 1972, J. Comb. Theory A.