Best Response Regression

In a regression task, a predictor is given a set of instances, along with a real value for each point. Subsequently, she has to identify the value of a new instance as accurately as possible. In this work, we initiate the study of strategic predictions in machine learning. We consider a regression task tackled by two players, where the payoff of each player is the proportion of the points she predicts more accurately than the other player. We first revise the probably approximately correct learning framework to deal with the case of a duel between two predictors. We then devise an algorithm which finds a linear regression predictor that is a best response to any (not necessarily linear) regression algorithm. We show that it has linearithmic sample complexity, and polynomial time complexity when the dimension of the instances domain is fixed. We also test our approach in a high-dimensional setting, and show it significantly defeats classical regression algorithms in the prediction duel. Together, our work introduces a novel machine learning task that lends itself well to current competitive online settings, provides its theoretical foundations, and illustrates its applicability.

[1]  Vladimir Vapnik,et al.  A new learning paradigm: Learning using privileged information , 2009, Neural Networks.

[2]  D. Rubinfeld,et al.  Hedonic housing prices and the demand for clean air , 1978 .

[3]  Edoardo Amaldi,et al.  The Complexity and Approximability of Finding Maximum Feasible Subsystems of Linear Relations , 1995, Theor. Comput. Sci..

[4]  Shai Ben-David,et al.  Understanding Machine Learning: From Theory to Algorithms , 2014 .

[5]  Leslie G. Valiant,et al.  A theory of the learnable , 1984, CACM.

[6]  Vladimir Vapnik,et al.  Chervonenkis: On the uniform convergence of relative frequencies of events to their probabilities , 1971 .

[7]  Vladimir Vapnik,et al.  Learning using hidden information: Master-class learning , 2007, NATO ASI Mining Massive Data Sets for Security.

[8]  Richard Cole,et al.  The sample complexity of revenue maximization , 2014, STOC.

[9]  Vladimir Vapnik,et al.  On the Theory of Learnining with Privileged Information , 2010, NIPS.

[10]  Ariel D. Procaccia,et al.  Algorithms for strategyproof classification , 2012, Artif. Intell..

[11]  Ariel D. Procaccia,et al.  Incentive compatible regression learning , 2008, SODA '08.

[12]  Noam Nisan,et al.  Algorithmic mechanism design (extended abstract) , 1999, STOC '99.

[13]  V. Vapnik,et al.  On the theory of learning with Privileged Information , 2010, NIPS 2010.

[14]  Noam Nisan,et al.  Algorithmic Mechanism Design , 2001, Games Econ. Behav..

[15]  Adam Tauman Kalai,et al.  Dueling algorithms , 2011, STOC '11.

[16]  Norbert Sauer,et al.  On the Density of Families of Sets , 1972, J. Comb. Theory A.