Online A-Optimal Design and Active Linear Regression

We consider in this paper the problem of optimal experiment design where a decision maker can choose which points to sample to obtain an estimate β̂ of the hidden parameter β of an underlying linear model. The key challenge of this work lies in the heteroscedasticity assumption that we make, meaning that each covariate has a different and unknown variance. The goal of the decision maker is then to figure out on the fly the optimal way to allocate the total budget of T samples between covariates, as sampling several times a specific one will reduce the variance of the estimated model around it (but at the cost of a possible higher variance elsewhere). By trying to minimize the `2-loss E[‖β̂ − β?‖2] the decision maker is actually minimizing the trace of the covariance matrix of the problem, which corresponds then to online A-optimal design. Combining techniques from bandit and convex optimization we propose a new active sampling algorithm and we compare it with existing ones. We provide theoretical guarantees of this algorithm in different settings, including a O(T−2) regret bound in the case where the covariates form a basis of the feature space, generalizing and improving existing results. Numerical experiments validate our theoretical findings.

[1]  Eric Price,et al.  Active Regression via Linear-Sample Sparsification , 2017, COLT.

[2]  Alessandro Lazaric,et al.  Active Learning for Accurate Estimation of Linear Models , 2017, ICML.

[3]  Vianney Perchet,et al.  Fast Rates for Bandit Optimization with Upper-Confidence Frank-Wolfe , 2017, NIPS.

[4]  Martin J. Wainwright,et al.  High-Dimensional Statistics , 2019 .

[5]  P. Whittle A MULTIVARIATE GENERALIZATION OF TCHEBICHEV'S INEQUALITY , 1958 .

[6]  Guillaume Sagnol,et al.  Optimal design of experiments with application to the inference of traffic matrices in large networks: second order cone programming and submodularity , 2010 .

[7]  David A. Cohn,et al.  Active Learning with Statistical Models , 1996, NIPS.

[8]  Shie Mannor,et al.  Approachability in unknown games: Online learning meets multi-objective optimization , 2014, COLT.

[9]  Marta Soare,et al.  Sequential Resource Allocation in Linear Stochastic Bandits , 2015 .

[10]  Hon Keung Tony Ng,et al.  Efficient computational algorithm for optimal allocation in regression models , 2013, J. Comput. Appl. Math..

[11]  Masashi Sugiyama,et al.  Active Learning with Model Selection in Linear Regression , 2008, SDM.

[12]  Stefanie Biedermann,et al.  On Optimal Designs for Nonlinear Models: A General and Efficient Algorithm , 2013 .

[13]  O. Papaspiliopoulos High-Dimensional Probability: An Introduction with Applications in Data Science , 2020 .

[14]  Andreas Krause,et al.  Information Directed Sampling and Bandits with Heteroscedastic Noise , 2018, COLT.

[15]  Manfred K. Warmuth,et al.  Unbiased estimators for random design regression , 2019, ArXiv.

[16]  Michael N. Katehakis,et al.  Normal Bandits of Unknown Means and Variances: Asymptotic Optimality, Finite Horizon Regret Bounds, and a Solution to an Open Problem , 2015, ArXiv.

[17]  Elad Hazan,et al.  Hard-Margin Active Linear Regression , 2014, ICML.

[18]  Alessandro Lazaric,et al.  Trading off Rewards and Errors in Multi-Armed Bandits , 2017, AISTATS.

[19]  Djalil Chafaï,et al.  Interactions between compressed sensing, random matrices, and high dimensional geometry , 2012 .

[20]  W. R. Thompson ON THE LIKELIHOOD THAT ONE UNKNOWN PROBABILITY EXCEEDS ANOTHER IN VIEW OF THE EVIDENCE OF TWO SAMPLES , 1933 .

[21]  Massimiliano Pontil,et al.  Empirical Bernstein Bounds and Sample-Variance Penalization , 2009, COLT.

[22]  Sham M. Kakade,et al.  An Analysis of Random Design Linear Regression , 2011, ArXiv.

[23]  Carlos Riquelme,et al.  Online Active Linear Regression via Thresholding , 2016, AAAI.

[24]  Varun Grover,et al.  Active learning in heteroscedastic noise , 2010, Theor. Comput. Sci..

[25]  Wei Chen,et al.  Improving Regret Bounds for Combinatorial Semi-Bandits with Probabilistically Triggered Arms and Its Applications , 2017, NIPS.

[26]  Rémi Munos,et al.  Active Regression by Stratification , 2014, NIPS.

[27]  Yuanzhi Li,et al.  Near-optimal discrete optimization for experimental design: a regret minimization approach , 2017, Mathematical Programming.

[28]  Vianney Perchet,et al.  Regularized Contextual Bandits , 2019, AISTATS.

[29]  Peter Auer,et al.  Finite-time Analysis of the Multiarmed Bandit Problem , 2002, Machine Learning.

[30]  Michael Jackson,et al.  Optimal Design of Experiments , 1994 .

[31]  H. Sebastian Seung,et al.  Selective Sampling Using the Query by Committee Algorithm , 1997, Machine Learning.

[32]  Alessandro Lazaric,et al.  Upper-Confidence-Bound Algorithms for Active Learning in Multi-Armed Bandits , 2011, ALT.

[33]  Peter Goos,et al.  Optimal Design of Experiments: A Case Study Approach , 2011 .

[34]  Nikhil R. Devanur,et al.  Bandits with concave rewards and convex knapsacks , 2014, EC.

[35]  Roman Vershynin,et al.  High-Dimensional Probability , 2018 .