Mitigating Concept Drift via Rejection

Learning in non-stationary environments is challenging, because under such conditions the common assumption of independent and identically distributed data does not hold; when concept drift is present it necessitates continuous system updates. In recent years, several powerful approaches have been proposed. However, these models typically classify any input, regardless of their confidence in the classification – a strategy, which is not optimal, particularly in safety-critical environments where alternatives to a (possibly unclear) decision exist, such as additional tests or a short delay of the decision. Formally speaking, this alternative corresponds to classification with rejection, a strategy which seems particularly promising in the context of concept drift, i.e. the occurrence of situations where the current model is wrong due to a concept change. In this contribution, we propose to extend learning under concept drift with rejection. Specifically, we extend two recent learning architectures for drift, the self-adjusting memory architecture (SAM-kNN) and adaptive random forests (ARF), to incorporate a reject option, resulting in highly competitive state-of-the-art technologies. We evaluate their performance in learning scenarios with different types of drift.

[1]  T. Watkin,et al.  THE STATISTICAL-MECHANICS OF LEARNING A RULE , 1993 .

[2]  Siddhartha S. Srinivasa,et al.  Perceived robot capability , 2015, 2015 24th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN).

[3]  Senén Barro,et al.  Do we need hundreds of classifiers to solve real world classification problems? , 2014, J. Mach. Learn. Res..

[4]  Peter L. Bartlett,et al.  Classification with a Reject Option using a Hinge Loss , 2008, J. Mach. Learn. Res..

[5]  Christophe Marsala,et al.  Droplet Ensemble Learning on Drifting Data Streams , 2017, IDA.

[6]  G. Shafer,et al.  Algorithmic Learning in a Random World , 2005 .

[7]  Radu Herbei,et al.  Classification with reject option , 2006 .

[8]  Gregory Ditzler,et al.  Learning in Nonstationary Environments: A Survey , 2015, IEEE Computational Intelligence Magazine.

[9]  Geoff Holmes,et al.  MOA: Massive Online Analysis , 2010, J. Mach. Learn. Res..

[10]  W. Gasarch,et al.  The Book Review Column 1 Coverage Untyped Systems Simple Types Recursive Types Higher-order Systems General Impression 3 Organization, and Contents of the Book , 2022 .

[11]  Heiko Wersing,et al.  KNN Classifier with Self Adjusting Memory for Heterogeneous Concept Drift , 2016, 2016 IEEE 16th International Conference on Data Mining (ICDM).

[12]  Heiko Wersing,et al.  Tackling heterogeneous concept drift with the Self-Adjusting Memory (SAM) , 2017, Knowledge and Information Systems.

[13]  Heiko Wersing,et al.  Optimal local rejection for classifiers , 2016, Neurocomputing.

[14]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[15]  Anca D. Dragan,et al.  Expressing Robot Incapability , 2018, 2018 13th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[16]  John Platt,et al.  Probabilistic Outputs for Support vector Machines and Comparisons to Regularized Likelihood Methods , 1999 .

[17]  Martin E. Hellman,et al.  The Nearest Neighbor Classification Rule with a Reject Option , 1970, IEEE Trans. Syst. Sci. Cybern..

[18]  Holly A. Yanco,et al.  Impact of robot failures and feedback on real-time trust , 2013, 2013 8th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[19]  Blaise Hanczar,et al.  Accuracy-Rejection Curves (ARCs) for Comparing Classification Methods with a Reject Option , 2009, MLSB.

[20]  C. K. Chow,et al.  On optimum recognition error and reject tradeoff , 1970, IEEE Trans. Inf. Theory.

[21]  Padraig Cunningham,et al.  Generating Estimates of Classification Confidence for a Case-Based Spam Filter , 2005, ICCBR.

[22]  Rich Caruana,et al.  Predicting good probabilities with supervised learning , 2005, ICML.

[23]  Thierry Denoeux A k -Nearest Neighbor Classification Rule Based on Dempster-Shafer Theory , 2008, Classic Works of the Dempster-Shafer Theory of Belief Functions.

[24]  Talel Abdessalem,et al.  Adaptive random forests for evolving data stream classification , 2017, Machine Learning.

[25]  Jean Paul Barddal,et al.  A Survey on Ensemble Learning for Data Stream Classification , 2017, ACM Comput. Surv..

[26]  Christophe Marsala,et al.  Classification with a reject option under Concept Drift: The Droplets algorithm , 2015, 2015 IEEE International Conference on Data Science and Advanced Analytics (DSAA).

[27]  Thomas Villmann,et al.  Self-Adjusting Reject Options in Prototype Based Classification , 2016, WSOM.