A closer look at temporal variability in dynamic online learning

This work focuses on the setting of dynamic regret in the context of online learning with full information. In particular, we analyze regret bounds with respect to the temporal variability of the loss functions. By assuming that the sequence of loss functions does not vary much with time, we show that it is possible to incur improved regret bounds compared to existing results. The key to our approach is to use the loss function (and not its gradient) during the optimization process. Building on recent advances in the analysis of Implicit algorithms, we propose an adaptation of the Implicit version of Online Mirror Descent to the dynamic setting. Our proposed algorithm is adaptive not only to the temporal variability of the loss functions, but also to the path length of the sequence of comparators when an upper bound is known. Furthermore, our analysis reveals that our results are tight and cannot be improved without further assumptions. Next, we show how our algorithm can be applied to the setting of learning with expert advice or to settings with composite loss functions. Finally, when an upper bound to the path-length is not fixed beforehand we show how to combine a greedy strategy with existing strongly-adaptive algorithms to compete optimally against different sequences of comparators simultaneously.

[1]  Francesco Orabona A Modern Introduction to Online Learning , 2019, ArXiv.

[2]  Jinfeng Yi,et al.  Tracking Slowly Moving Clairvoyant: Optimal Dynamic Regret of Online Learning with True and Noisy Gradient , 2016, ICML.

[3]  Martin Zinkevich,et al.  Online Convex Programming and Generalized Infinitesimal Gradient Ascent , 2003, ICML.

[4]  Francesco Orabona,et al.  Temporal Variability in Implicit Online Learning , 2020, NeurIPS.

[5]  Mark Herbster,et al.  Tracking the Best Linear Predictor , 2001, J. Mach. Learn. Res..

[6]  Mark Herbster,et al.  Tracking the Best Expert , 1995, Machine Learning.

[7]  Amit Daniely,et al.  Strongly Adaptive Online Learning , 2015, ICML.

[8]  Koby Crammer,et al.  Online Passive-Aggressive Algorithms , 2003, J. Mach. Learn. Res..

[9]  Shai Shalev-Shwartz,et al.  Online Learning and Online Convex Optimization , 2012, Found. Trends Mach. Learn..

[10]  Lijun Zhang,et al.  Adaptive Online Learning in Dynamic Environments , 2018, NeurIPS.

[11]  J. Moreau Proximité et dualité dans un espace hilbertien , 1965 .

[12]  Emiliano Dall'Anese,et al.  Inexact Online Proximal-gradient Method for Time-varying Convex Optimization , 2020, 2020 American Control Conference (ACC).

[13]  Shahin Shahrampour,et al.  Online Optimization : Competing with Dynamic Comparators , 2015, AISTATS.

[14]  Nicolò Cesa-Bianchi,et al.  Mirror Descent Meets Fixed Share (and feels no regret) , 2012, NIPS.

[15]  Francesco Orabona,et al.  Scale-free online learning , 2016, Theor. Comput. Sci..

[16]  Tong Zhang,et al.  Fully Implicit Online Learning , 2018, ArXiv.

[17]  Ambuj Tewari,et al.  Composite objective mirror descent , 2010, COLT 2010.

[18]  Rebecca Willett,et al.  Dynamical Models and tracking regret in online convex programming , 2013, ICML.

[19]  Seshadhri Comandur,et al.  Electronic Colloquium on Computational Complexity, Report No. 88 (2007) Adaptive Algorithms for Online Decision Problems , 2022 .

[20]  Claudio Gentile,et al.  On the generalization ability of on-line learning algorithms , 2001, IEEE Transactions on Information Theory.

[21]  Rong Jin,et al.  Dynamic Regret of Strongly Adaptive Methods , 2017, ICML.

[22]  Ketan Rajawat,et al.  Online Learning With Inexact Proximal Online Gradient Descent Algorithms , 2018, IEEE Transactions on Signal Processing.

[23]  Yurii Nesterov,et al.  Introductory Lectures on Convex Optimization - A Basic Course , 2014, Applied Optimization.

[24]  Ashok Cutkosky,et al.  Parameter-free, Dynamic, and Strongly-Adaptive Online Learning , 2020, ICML.

[25]  Gilles Stoltz,et al.  A second-order bound with excess losses , 2014, COLT.

[26]  Alessandro Lazaric,et al.  Exploiting easy data in online optimization , 2014, NIPS.