Incentive-Aware PAC Learning

We study PAC learning in the presence of strategic manipulation, where data points may modify their features in certain predefined ways in order to receive a better outcome. We show that the vanilla ERM principle fails to achieve any nontrivial guarantee in this context. Instead, we propose an incentive-aware version of the ERM principle which has asymptotically optimal sample complexity. We then focus our attention on incentive-compatible classifiers, which provably prevent any kind of strategic manipulation. We give a sample complexity bound that is, curiously, independent of the hypothesis class, for the ERM principle restricted to incentivecompatible classifiers. This suggests that incentive compatibility alone can act as an effective means of regularization. We further show that it is without loss of generality to consider only incentive-compatible classifiers when opportunities for strategic manipulation satisfy a transitivity condition. As a consequence, in such cases, our hypothesis-classindependent sample complexity bound applies even without incentive compatibility. Our results set the foundations of incentive-aware PAC learning.

[1]  Vincent Conitzer,et al.  The Revelation Principle for Mechanism Design with Reporting Costs , 2016, EC.

[2]  Ariel D. Procaccia,et al.  Incentive compatible regression learning , 2008, SODA '08.

[3]  David M. Pennock,et al.  No-Regret and Incentive-Compatible Prediction with Expert Advice , 2019 .

[4]  Haoming Li,et al.  Classification with Strategically Withheld Data , 2020, AAAI.

[5]  Hanrui Zhang,et al.  Automated Mechanism Design for Classification with Partial Verification , 2021, AAAI.

[6]  Vladimir Vapnik,et al.  Chervonenkis: On the uniform convergence of relative frequencies of events to their probabilities , 1971 .

[7]  Christos H. Papadimitriou,et al.  Strategic Classification , 2015, ITCS.

[8]  Lan Yu Mechanism design with partial verification and revelation principle , 2010, Autonomous Agents and Multi-Agent Systems.

[9]  Jon M. Kleinberg,et al.  How Do Classifiers Induce Agents to Invest Effort Strategically? , 2018, EC.

[10]  J. Lamperti ON CONVERGENCE OF STOCHASTIC PROCESSES , 1962 .

[11]  Shai Ben-David,et al.  Understanding Machine Learning: From Theory to Algorithms , 2014 .

[12]  Jerry R. Green,et al.  Partially Verifiable Information and Mechanism Design , 1986 .

[13]  Prateek Mittal,et al.  PAC-learning in the presence of adversaries , 2018, NeurIPS.

[14]  Peter L. Bartlett,et al.  Rademacher and Gaussian Complexities: Risk Bounds and Structural Results , 2003, J. Mach. Learn. Res..

[15]  Shai Ben-David,et al.  Multiclass Learnability and the ERM principle , 2011, COLT.

[16]  Nicole Immorlica,et al.  Maximizing Welfare with Incentive-Aware Evaluation Mechanisms , 2020, IJCAI.

[17]  Ariel D. Procaccia,et al.  Strategyproof Linear Regression in High Dimensions , 2018, EC.

[18]  Vincent Conitzer,et al.  Distinguishing Distributions When Samples Are Strategically Transformed , 2019, NeurIPS.

[19]  László Györfi,et al.  A Probabilistic Theory of Pattern Recognition , 1996, Stochastic Modelling and Applied Probability.

[20]  M. Talagrand Sharper Bounds for Gaussian and Empirical Processes , 1994 .

[21]  Haifeng Xu,et al.  The Intrinsic Robustness of Stochastic Bandits to Strategic Manipulation , 2020, ICML.

[22]  Javier Perote,et al.  Strategy-proof estimators for simple regression , 2004, Math. Soc. Sci..

[23]  Tim Roughgarden,et al.  Online Prediction with Selfish Experts , 2017, NIPS.

[24]  S. Matthew Weinberg,et al.  Multi-armed Bandit Problems with Strategic Arms , 2017, COLT.

[25]  Vincent Conitzer,et al.  Complexity of Mechanism Design with Signaling Costs , 2015, AAMAS.

[26]  HannekeSteve The optimal sample complexity OF PAC learning , 2016 .

[27]  Vincent Conitzer,et al.  When Samples Are Strategically Selected , 2019, ICML.

[28]  Noga Alon,et al.  Scale-sensitive dimensions, uniform convergence, and learnability , 1997, JACM.

[29]  Leslie G. Valiant,et al.  A theory of the learnable , 1984, STOC '84.

[30]  Steve Hanneke,et al.  The Optimal Sample Complexity of PAC Learning , 2015, J. Mach. Learn. Res..

[31]  Hanrui Zhang,et al.  Classification with Few Tests through Self-Selection , 2021, AAAI.