Some label efficient learning results

We investigate the value of labels in a simple version of the standard on-line prediction model (the “experts” setting). We present algorithms and adversary arguments defining tradeoffs between the number of mistakes made and the number of labels that the learner requests. One version of this question can be viewed as a family of games whose value is given by a complicated recurrence. Although our attempts to tind a closed form for this recurrence have been unsuccessful, we show how an algorithm can efficiently compute its value, enabling it to perform optimally.

[1]  David Haussler,et al.  How to use expert advice , 1993, STOC.

[2]  Manfred K. Warmuth,et al.  On Weak Learning , 1995, J. Comput. Syst. Sci..

[3]  Philip M. Long,et al.  Apple tasting and nearly one-sided learning , 1992, Proceedings., 33rd Annual Symposium on Foundations of Computer Science.

[4]  Vladimir Vovk,et al.  Aggregating strategies , 1990, COLT '90.

[5]  Nick Littlestone,et al.  From on-line to batch learning , 1989, COLT '89.

[6]  Leslie G. Valiant,et al.  A theory of the learnable , 1984, STOC '84.

[7]  Thomas G. Dietterich What is machine learning? , 2020, Archives of Disease in Childhood.

[8]  Manfred K. Warmuth,et al.  The Weighted Majority Algorithm , 1994, Inf. Comput..

[9]  H. Sebastian Seung,et al.  Query by committee , 1992, COLT '92.

[10]  David A. Cohn,et al.  Training Connectionist Networks with Queries and Selective Sampling , 1989, NIPS.