Open Problem: Parameter-Free and Scale-Free Online Algorithms

Existing vanilla algorithms for online linear optimization have O((ηR(u) + 1/η) √ T ) regret with respect to any competitor u, whereR(u) is a 1-strongly convex regularizer and η > 0 is a tuning parameter of the algorithm. For certain decision sets and regularizers, the so-called parameter-free algorithms have Õ( √ R(u)T ) regret with respect to any competitor u. Vanilla algorithm can achieve the same bound only for a fixed competitor u known ahead of time by setting η = 1/ √ R(u). A drawback of both vanilla and parameter-free algorithms is that they assume that the norm of the loss vectors is bounded by a constant known to the algorithm. There exist scale-free algorithms that haveO((ηR(u)+1/η) √ T max1≤t≤T ‖`t‖) regret with respect to any competitor u and for any sequence of loss vector `1, . . . , `T . Parameter-free analogue of scale-free algorithms have never been designed. Is is possible to design algorithms that are simultaneously parameter-free and scale-free?

[1]  Karthik Sridharan,et al.  Adaptive Online Learning , 2015, NIPS.

[2]  Haipeng Luo,et al.  Achieving All with No Parameters: AdaNormalHedge , 2015, COLT.

[3]  Haipeng Luo,et al.  A Drifting-Games Analysis for Online Learning and Applications to Boosting , 2014, NIPS.

[4]  Yoav Freund,et al.  A Parameter-free Hedging Algorithm , 2009, NIPS.

[5]  Gábor Lugosi,et al.  Prediction, learning, and games , 2006 .

[6]  Francesco Orabona,et al.  Unconstrained Online Linear Learning in Hilbert Spaces: Minimax Algorithms and Normal Approximations , 2014, COLT.

[7]  Vladimir Vovk,et al.  Prediction with Advice of Unknown Number of Experts , 2010, UAI.

[8]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.

[9]  Francesco Orabona,et al.  Simultaneous Model Selection and Optimization through Parameter-free Stochastic Learning , 2014, NIPS.

[10]  H. Brendan McMahan,et al.  Minimax Optimal Algorithms for Unconstrained Linear Optimization , 2013, NIPS.

[11]  Francesco Orabona,et al.  Scale-free online learning , 2016, Theor. Comput. Sci..

[12]  Francesco Orabona,et al.  Dimension-Free Exponentiated Gradient , 2013, NIPS.

[13]  Shai Shalev-Shwartz,et al.  Online Learning and Online Convex Optimization , 2012, Found. Trends Mach. Learn..

[14]  Wouter M. Koolen,et al.  Follow the leader if you can, hedge if you must , 2013, J. Mach. Learn. Res..

[15]  Francesco Orabona,et al.  From Coin Betting to Parameter-Free Online Learning , 2016, ArXiv.

[16]  Wouter M. Koolen,et al.  Second-order Quantile Methods for Experts and Combinatorial Games , 2015, COLT.

[17]  Francesco Orabona,et al.  Scale-Free Algorithms for Online Linear Optimization , 2015, ALT.

[18]  Matthew J. Streeter,et al.  No-Regret Algorithms for Unconstrained Online Convex Optimization , 2012, NIPS.