A Closer Look at Adaptive Regret

For the prediction with expert advice setting, we consider methods to construct algorithms that have low adaptive regret. The adaptive regret of an algorithm on a time interval [t1, t2] is the loss of the algorithm minus the loss of the best expert over that interval. Adaptive regret measures how well the algorithm approximates the best expert locally, and so is different from, although closely related to, both the classical regret, measured over an initial time interval [1, t], and the tracking regret, where the algorithm is compared to a good sequence of experts over [1, t]. We investigate two existing intuitive methods for deriving algorithms with low adaptive regret, one based on specialist experts and the other based on restarts. Quite surprisingly, we show that both methods lead to the same algorithm, namely Fixed Share, which is known for its tracking regret. We provide a thorough analysis of the adaptive regret of Fixed Share. We obtain the exact worst-case adaptive regret for Fixed Share, from which the classical tracking bounds follow. We prove that Fixed Share is optimal for adaptive regret: the worst-case adaptive regret of any algorithm is at least that of an instance of Fixed Share.

[1]  Wouter M. Koolen,et al.  Combining Expert Advice Efficiently , 2008, COLT.

[2]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.

[3]  Wouter M. Koolen Combining strategies efficiently: high-quality decisions from conflicting advice , 2011 .

[4]  Vladimir Vovk,et al.  Prediction with Expert Evaluators' Advice , 2009, ALT.

[5]  Steven de Rooij,et al.  Learning the Switching Rate by Discretising Bernoulli Sources Online , 2009, AISTATS.

[6]  Martin Zinkevich,et al.  Online Convex Programming and Generalized Infinitesimal Gradient Ascent , 2003, ICML.

[7]  Tommi S. Jaakkola,et al.  Online Learning of Non-stationary Sequences , 2003, NIPS.

[8]  Seshadhri Comandur,et al.  Efficient learning algorithms for changing environments , 2009, ICML '09.

[9]  Tamás Linder,et al.  Efficient Tracking of Large Classes of Experts , 2012, IEEE Trans. Inf. Theory.

[10]  Manfred K. Warmuth,et al.  The Weighted Majority Algorithm , 1994, Inf. Comput..

[11]  Vladimir Vovk,et al.  Derandomizing Stochastic Prediction Strategies , 1997, COLT '97.

[12]  W. M. Koolen-Wijkstra,et al.  Combining expert advice efficiently , 2008 .

[13]  Nicolò Cesa-Bianchi,et al.  Mirror Descent Meets Fixed Share (and feels no regret) , 2012, NIPS.

[14]  Mark Herbster,et al.  Tracking the Best Expert , 1995, Machine Learning.

[15]  Manfred K. Warmuth,et al.  Tracking a Small Set of Experts by Mixing Past Posteriors , 2003, J. Mach. Learn. Res..

[16]  Wouter M. Koolen,et al.  Universal Codes From Switching Strategies , 2013, IEEE Transactions on Information Theory.

[17]  Vladimir Vovk,et al.  Aggregating strategies , 1990, COLT '90.

[18]  Vladimir Vovk,et al.  A game of prediction with expert advice , 1995, COLT '95.

[19]  Elad Hazan,et al.  Adaptive Algorithms for Online Optimization , 2007 .

[20]  Gábor Lugosi,et al.  Prediction, learning, and games , 2006 .

[21]  Neri Merhav,et al.  Low-complexity sequential lossless coding for piecewise-stationary memoryless sources , 1998, IEEE Trans. Inf. Theory.

[22]  N. Merhav,et al.  Low complexity sequential lossless coding for piecewise stationary memoryless sources , 1998, Proceedings. 1998 IEEE International Symposium on Information Theory (Cat. No.98CH36252).

[23]  Yoram Singer,et al.  Using and combining predictors that specialize , 1997, STOC '97.

[24]  V. Vovk Competitive On‐line Statistics , 2001 .

[25]  Nicolò Cesa-Bianchi,et al.  A new look at shifting regret , 2012, ArXiv.