Deep Streaming Label Learning
暂无分享,去创建一个
[1] Shiming Xiang,et al. Multi-Label Image Classification via Knowledge Distillation from Weakly-Supervised Detection , 2018, ACM Multimedia.
[2] Tony X. Han,et al. Learning Efficient Object Detection Models with Knowledge Distillation , 2017, NIPS.
[3] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[4] Zhi-Hua Zhou,et al. ML-KNN: A lazy learning approach to multi-label learning , 2007, Pattern Recognit..
[5] Rich Caruana,et al. Do Deep Nets Really Need to be Deep? , 2013, NIPS.
[6] Johannes Fürnkranz,et al. Large-Scale Multi-label Text Classification - Revisiting Neural Networks , 2013, ECML/PKDD.
[7] Yu-Chiang Frank Wang,et al. Order-Free RNN with Visual Attention for Multi-Label Classification , 2017, AAAI.
[8] Ankit Singh Rawat,et al. Multilabel reductions: what is my loss optimising? , 2019, NeurIPS.
[9] Zhi-Hua Zhou,et al. A Unified View of Multi-Label Performance Measures , 2016, ICML.
[10] Quanquan Gu,et al. Generalization Error Bounds of Gradient Descent for Learning Over-Parameterized Deep ReLU Networks , 2019, AAAI.
[11] Liu He,et al. Bounding the efficiency gain of differentiable road pricing for EVs and GVs to manage congestion and emissions , 2020, PloS one.
[12] Manik Varma,et al. FastXML: a fast, accurate and stable tree-classifier for extreme multi-label learning , 2014, KDD.
[13] Dacheng Tao,et al. Streaming Label Learning for Modeling Labels on the Fly , 2016, ArXiv.
[14] Robert E. Schapire,et al. Hierarchical multi-label prediction of gene function , 2006, Bioinform..
[15] Zhi-Hua Zhou,et al. Multilabel Neural Networks with Applications to Functional Genomics and Text Categorization , 2006, IEEE Transactions on Knowledge and Data Engineering.
[16] Krista A. Ehinger,et al. SUN database: Large-scale scene recognition from abbey to zoo , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.
[17] Johannes Fürnkranz,et al. Learning Context-dependent Label Permutations for Multi-label Classification , 2019, ICML.
[18] Yuanzhi Li,et al. A Convergence Theory for Deep Learning via Over-Parameterization , 2018, ICML.
[19] Prateek Jain,et al. Sparse Local Embeddings for Extreme Multi-label Classification , 2015, NIPS.
[20] Weiwei Liu,et al. Sparse Extreme Multi-label Learning with Oracle Property , 2019, ICML.
[21] Piotr Szymanski,et al. A scikit-based Python environment for performing multi-label classification , 2017, ArXiv.
[22] Nicolò Cesa-Bianchi,et al. Synergy of multi-label hierarchical ensembles, data fusion, and cost-sensitive methods for gene functional inference , 2012, Machine Learning.
[23] Sangdoo Yun,et al. A Comprehensive Overhaul of Feature Distillation , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[24] Yue Gao,et al. Physiological Signals-based Emotion Recognition via High-order Correlation Learning , 2019 .
[25] Andreas Maurer,et al. A Vector-Contraction Inequality for Rademacher Complexities , 2016, ALT.
[26] Min-Ling Zhang,et al. A Review on Multi-Label Learning Algorithms , 2014, IEEE Transactions on Knowledge and Data Engineering.
[27] Liwei Wang,et al. Gradient Descent Finds Global Minima of Deep Neural Networks , 2018, ICML.
[28] Jürgen Schmidhuber,et al. Deep learning in neural networks: An overview , 2014, Neural Networks.
[29] Inderjit S. Dhillon,et al. Large-scale Multi-label Learning with Missing Labels , 2013, ICML.
[30] Geoff Holmes,et al. Classifier chains for multi-label classification , 2009, Machine Learning.
[31] Zhi-Hua Zhou,et al. Multi-Label Learning with Emerging New Labels , 2018, IEEE Transactions on Knowledge and Data Engineering.
[32] Manik Varma,et al. Extreme Multi-label Loss Functions for Recommendation, Tagging, Ranking & Other Missing Label Applications , 2016, KDD.
[33] Graham Neubig,et al. Understanding Knowledge Distillation in Non-autoregressive Machine Translation , 2019, ICLR.
[34] Irwin King,et al. Few Shot Network Compression via Cross Distillation , 2020, AAAI.
[35] Ambuj Tewari,et al. Regularization Techniques for Learning with Matrices , 2009, J. Mach. Learn. Res..
[36] Jeff A. Bilmes,et al. Deep Canonical Correlation Analysis , 2013, ICML.
[37] Yoshua Bengio,et al. FitNets: Hints for Thin Deep Nets , 2014, ICLR.
[38] Yu-Chiang Frank Wang,et al. Learning Deep Latent Spaces for Multi-Label Classification , 2017, ArXiv.
[39] Rui Zhang,et al. DBSVEC: Density-Based Clustering Using Support Vector Expansion , 2019, 2019 IEEE 35th International Conference on Data Engineering (ICDE).
[40] Yuan Cao,et al. Stochastic Gradient Descent Optimizes Over-parameterized Deep ReLU Networks , 2018, ArXiv.
[41] Johannes Fürnkranz,et al. Maximizing Subset Accuracy with Recurrent Neural Networks in Multi-label Classification , 2017, NIPS.
[42] Guigang Zhang,et al. Deep Learning , 2016, Int. J. Semantic Comput..