Adapting to Online Label Shift with Provable Guarantees

The standard supervised learning paradigm works effectively when training data shares the same distribution as the upcoming testing samples. However, this stationary assumption is often violated in real-world applications, especially when testing data appear in an online fashion. In this paper, we formulate and investigate the problem of \emph{online label shift} (OLaS): the learner trains an initial model from the labeled offline data and then deploys it to an unlabeled online environment where the underlying label distribution changes over time but the label-conditional density does not. The non-stationarity nature and the lack of supervision make the problem challenging to be tackled. To address the difficulty, we construct a new unbiased risk estimator that utilizes the unlabeled data, which exhibits many benign properties albeit with potential non-convexity. Building upon that, we propose novel online ensemble algorithms to deal with the non-stationarity of the environments. Our approach enjoys optimal \emph{dynamic regret}, indicating that the performance is competitive with a clairvoyant who knows the online environments in hindsight and then chooses the best decision for each round. The obtained dynamic regret bound scales with the intensity and pattern of label distribution shift, hence exhibiting the adaptivity in the OLaS problem. Extensive experiments are conducted to validate the effectiveness and support our theoretical findings.

[1]  Zhi-Hua Zhou,et al.  Model Reuse With Reduced Kernel Mean Embedding Specification , 2020, IEEE Transactions on Knowledge and Data Engineering.

[2]  Peng Zhao,et al.  Dynamic Regret of Online Markov Decision Processes , 2022, ICML.

[3]  Zhi Zhou Open-environment machine learning , 2022, National science review.

[4]  Ashok Cutkosky,et al.  Parameter-free Mirror Descent , 2022, COLT.

[5]  Haipeng Luo,et al.  No-Regret Learning in Time-Varying Zero-Sum Games , 2022, ICML.

[6]  Wenwu Zhu,et al.  Online Continual Adaptation with Active Self-Training , 2021, AISTATS.

[7]  Yu-Xiang Wang,et al.  Non-stationary Online Learning with Memory and Non-stochastic Control , 2021, AISTATS.

[8]  Zhi-Hua Zhou,et al.  Adaptivity and Non-stationarity: Problem-dependent Dynamic Regret for Online Convex Optimization , 2021, ArXiv.

[9]  Kilian Q. Weinberger,et al.  Online Adaptation to Label Distribution Shift , 2021, NeurIPS.

[10]  Geoffrey E. Hinton,et al.  Deep learning for AI , 2021, Commun. ACM.

[11]  Zhi-Hua Zhou,et al.  Towards Enabling Learnware to Handle Unseen Jobs , 2021, AAAI.

[12]  Yu-Xiang Wang,et al.  Optimal Dynamic Regret in Exp-Concave Online Learning , 2021, COLT.

[13]  Tianbao Yang,et al.  Revisiting Smoothed Online Learning , 2021, NeurIPS.

[14]  Peng Zhao,et al.  Improved Analysis for Dynamic Regret of Strongly Convex and Smooth Functions , 2020, L4DC.

[15]  Kane,et al.  Beyond the Worst-Case Analysis of Algorithms , 2020 .

[16]  Zhi-Hua Zhou,et al.  Skewness Ranking Optimization for Personalized Recommendation , 2020, UAI.

[17]  Ashok Cutkosky,et al.  Parameter-free, Dynamic, and Strongly-Adaptive Online Learning , 2020, ICML.

[18]  Zhi-Hua Zhou,et al.  Dynamic Regret of Convex and Smooth Functions , 2020, NeurIPS.

[19]  Francesco Orabona,et al.  Temporal Variability in Implicit Online Learning , 2020, NeurIPS.

[20]  Peng Zhao,et al.  Optimal Margin Distribution Learning in Dynamic Environments , 2020, AAAI.

[21]  Sivaraman Balakrishnan,et al.  A Unified View of Label Shift Estimation , 2020, NeurIPS.

[22]  Tianbao Yang,et al.  Minimizing Dynamic Regret and Adaptive Regret Simultaneously , 2020, AISTATS.

[23]  Zhi-Hua Zhou,et al.  Bandit Convex Optimization in Non-stationary Environments , 2019, AISTATS.

[24]  Baharan Mirzasoleiman,et al.  Coresets for Data-efficient Training of Machine Learning Models , 2019, ICML.

[25]  Thomas Wolf,et al.  DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter , 2019, ArXiv.

[26]  Yu-Xiang Wang,et al.  Online Forecasting of Total-Variation-bounded Sequences , 2019, NeurIPS.

[27]  Kamyar Azizzadenesheli,et al.  Regularized Learning for Domain Adaptation under Label Shifts , 2019, ICLR.

[28]  Zhi-Hua Zhou,et al.  Handling concept drift via model reuse , 2018, Machine Learning.

[29]  Lijun Zhang,et al.  Adaptive Online Learning in Dynamic Environments , 2018, NeurIPS.

[30]  Amos J. Storkey,et al.  School of Informatics, University of Edinburgh , 2022 .

[31]  Lin Wang,et al.  The University of Sussex-Huawei Locomotion and Transportation Dataset for Multimodal Analytics With Mobile Devices , 2018, IEEE Access.

[32]  Andreas Dengel,et al.  Introducing Eurosat: A Novel Dataset and Deep Learning Benchmark for Land Use and Land Cover Classification , 2018, IGARSS 2018 - 2018 IEEE International Geoscience and Remote Sensing Symposium.

[33]  Alexander J. Smola,et al.  Detecting and Correcting for Label Shift with Black Box Predictors , 2018, ICML.

[34]  Margaret Kosmala,et al.  Automatically identifying, counting, and describing wild animals in camera-trap images with deep learning , 2017, Proceedings of the National Academy of Sciences.

[35]  Rong Jin,et al.  Dynamic Regret of Strongly Adaptive Methods , 2017, ICML.

[36]  Thomas G. Dietterich Steps Toward Robust Artificial Intelligence , 2017, AI Mag..

[37]  Roland Vollgraf,et al.  Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.

[38]  Yu-Xiang Wang,et al.  Non-stationary Stochastic Optimization under L p,q -Variation Measures , 2018 .

[39]  Andreas Krause,et al.  Practical Coreset Constructions for Machine Learning , 2017, 1703.06476.

[40]  Bernhard Schölkopf,et al.  Kernel Mean Embedding of Distributions: A Review and Beyonds , 2016, Found. Trends Mach. Learn..

[41]  Jesús Francisco Vargas-Bonilla,et al.  Towards automatic wild animal monitoring: Identification of animal species in camera-trap images using very deep convolutional neural networks , 2016, Ecol. Informatics.

[42]  Gang Niu,et al.  Class-prior estimation for learning from positive and unlabeled data , 2016, Machine Learning.

[43]  Elad Hazan,et al.  Introduction to Online Convex Optimization , 2016, Found. Trends Optim..

[44]  Jinfeng Yi,et al.  Tracking Slowly Moving Clairvoyant: Optimal Dynamic Regret of Online Learning with True and Noisy Gradient , 2016, ICML.

[45]  Aryan Mokhtari,et al.  Optimization in Dynamic Environments : Improved Regret Rates for Strongly Convex Problems , 2016 .

[46]  Ambuj Tewari,et al.  Mixture Proportion Estimation via Kernel Embeddings of Distributions , 2016, ICML.

[47]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[48]  Gang Niu,et al.  Convex Formulation for Learning from Positive and Unlabeled Data , 2015, ICML.

[49]  Haipeng Luo,et al.  Fast Convergence of Regularized Learning in Games , 2015, NIPS.

[50]  Clayton Scott,et al.  A Rate of Convergence for Mixture Proportion Estimation, with Application to Learning from Noisy Labels , 2015, AISTATS.

[51]  Shahin Shahrampour,et al.  Online Optimization : Competing with Dynamic Comparators , 2015, AISTATS.

[52]  Omar Besbes,et al.  Non-Stationary Stochastic Optimization , 2013, Oper. Res..

[53]  Masashi Sugiyama,et al.  Continuous Target Shift Adaptation in Supervised Learning , 2015, ACML.

[54]  Bernhard Schölkopf,et al.  Domain Adaptation under Target and Conditional Shift , 2013, ICML.

[55]  Karthik Sridharan,et al.  Online Learning with Predictable Sequences , 2012, COLT.

[56]  Rong Jin,et al.  25th Annual Conference on Learning Theory Online Optimization with Gradual Variations , 2022 .

[57]  Motoaki Kawanabe,et al.  Machine Learning in Non-Stationary Environments - Introduction to Covariate Shift Adaptation , 2012, Adaptive computation and machine learning.

[58]  Shai Shalev-Shwartz,et al.  Online Learning and Online Convex Optimization , 2012, Found. Trends Mach. Learn..

[59]  Masashi Sugiyama,et al.  Semi-Supervised Learning of Class Balance under Class-Prior Change by Distribution Matching , 2012, ICML.

[60]  Joel A. Tropp,et al.  User-Friendly Tail Bounds for Sums of Random Matrices , 2010, Found. Comput. Math..

[61]  Peter L. Bartlett,et al.  Implicit Online Learning , 2010, ICML.

[62]  Gilles Blanchard,et al.  Semi-Supervised Novelty Detection , 2010, J. Mach. Learn. Res..

[63]  Elad Hazan,et al.  Extracting certainty from uncertainty: regret bounded by variation in costs , 2008, Machine Learning.

[64]  Fei-Fei Li,et al.  ImageNet: A large-scale hierarchical image database , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.

[65]  Neil D. Lawrence,et al.  Dataset Shift in Machine Learning , 2009 .

[66]  Sham M. Kakade,et al.  A spectral algorithm for learning Hidden Markov Models , 2008, J. Comput. Syst. Sci..

[67]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[68]  Colin Fyfe,et al.  Online Clustering Algorithms , 2008, Int. J. Neural Syst..

[69]  Ambuj Tewari,et al.  Optimal Stragies and Minimax Lower Bounds for Online Convex Games , 2008, COLT.

[70]  Elad Hazan,et al.  Logarithmic regret algorithms for online convex optimization , 2006, Machine Learning.

[71]  Gábor Lugosi,et al.  Prediction, learning, and games , 2006 .

[72]  Martin Zinkevich,et al.  Online Convex Programming and Generalized Infinitesimal Gradient Ascent , 2003, ICML.

[73]  Claudio Gentile,et al.  Adaptive and Self-Confident On-Line Learning Algorithms , 2000, J. Comput. Syst. Sci..

[74]  Marco Saerens,et al.  Adjusting the Outputs of a Classifier to New a Priori Probabilities: A Simple Procedure , 2002, Neural Computation.

[75]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[76]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1997, EuroCOLT.

[77]  David Haussler,et al.  How to use expert advice , 1993, STOC.