暂无分享,去创建一个
Jayadeva | Prashant Gupta | Aashi Jindal | Debarka Sengupta | Debarka Sengupta | Prashant Gupta | Aashi Jindal
[1] Philip S. Yu,et al. Mining concept-drifting data streams using ensemble classifiers , 2003, KDD '03.
[2] Geoff Holmes,et al. Leveraging Bagging for Evolving Data Streams , 2010, ECML/PKDD.
[3] KlinkenbergRalf. Learning drifting concepts: Example selection vs. example weighting , 2004 .
[4] Chee Peng Lim,et al. Online pattern classification with multiple neural network systems: an experimental study , 2003, IEEE Trans. Syst. Man Cybern. Part C.
[5] Heiko Wersing,et al. KNN Classifier with Self Adjusting Memory for Heterogeneous Concept Drift , 2016, 2016 IEEE 16th International Conference on Data Mining (ICDM).
[6] Albert Bifet,et al. Efficient Online Evaluation of Big Data Stream Classifiers , 2015, KDD.
[7] Lida Xu,et al. The internet of things: a survey , 2014, Information Systems Frontiers.
[8] Piotr Indyk,et al. Approximate nearest neighbors: towards removing the curse of dimensionality , 1998, STOC '98.
[9] Yoav Freund,et al. A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.
[10] Ricard Gavaldà,et al. Learning from Time-Changing Data with Adaptive Windowing , 2007, SDM.
[11] João Gama,et al. A survey on concept drift adaptation , 2014, ACM Comput. Surv..
[12] Roberto Souto Maior de Barros,et al. An overview and comprehensive comparison of ensembles for concept drift , 2019, Inf. Fusion.
[13] Marcus A. Maloof,et al. Using additive expert ensembles to cope with concept drift , 2005, ICML.
[14] Marcus A. Maloof,et al. Dynamic Weighted Majority: An Ensemble Method for Drifting Concepts , 2007, J. Mach. Learn. Res..
[15] Joelle Pineau,et al. Online Bagging and Boosting for Imbalanced Data Streams , 2013, IEEE Transactions on Knowledge and Data Engineering.
[16] Talel Abdessalem,et al. Adaptive random forests for evolving data stream classification , 2017, Machine Learning.
[17] Albert Bifet,et al. GNUsmail: Open Framework for On-line Email Classification , 2010, ECAI.
[18] Leo Breiman,et al. Random Forests , 2001, Machine Learning.
[19] Jayadeva,et al. Discovery of rare cells from voluminous single cell expression data , 2018, Nature Communications.
[20] R. Schapire. The Strength of Weak Learnability , 1990, Machine Learning.
[21] Talel Abdessalem,et al. Scikit-Multiflow: A Multi-output Streaming Framework , 2018, J. Mach. Learn. Res..
[22] Geoff Holmes,et al. Fast Perceptron Decision Tree Learning from Evolving Data Streams , 2010, PAKDD.
[23] Ralf Klinkenberg,et al. Learning drifting concepts: Example selection vs. example weighting , 2004, Intell. Data Anal..
[24] Yoav Freund,et al. Boosting the margin: A new explanation for the effectiveness of voting methods , 1997, ICML.
[25] Robi Polikar,et al. Incremental Learning of Concept Drift in Nonstationary Environments , 2011, IEEE Transactions on Neural Networks.
[26] Stuart J. Russell,et al. Experimental comparisons of online and batch versions of bagging and boosting , 2001, KDD '01.
[27] Vasant Honavar,et al. Learn++: an incremental learning algorithm for supervised neural networks , 2001, IEEE Trans. Syst. Man Cybern. Part C.
[28] Stuart J. Russell,et al. Online bagging and boosting , 2005, 2005 IEEE International Conference on Systems, Man and Cybernetics.
[29] Charu C. Aggarwal,et al. Subspace Outlier Detection in Linear Time with Randomized Hashing , 2016, 2016 IEEE 16th International Conference on Data Mining (ICDM).
[30] Nitesh V. Chawla,et al. SMOTE: Synthetic Minority Over-sampling Technique , 2002, J. Artif. Intell. Res..
[31] Charu C. Aggarwal,et al. Recommendations For Streaming Data , 2016, CIKM.
[32] Thomas G. Dietterich. An Experimental Comparison of Three Methods for Constructing Ensembles of Decision Trees: Bagging, Boosting, and Randomization , 2000, Machine Learning.