Online Knowledge Distillation via Collaborative Learning
暂无分享,去创建一个
Xiaolin Hu | Xinjiang Wang | Ding Liang | Ping Luo | Qiushan Guo | Zhipeng Yu | Yichao Wu | Yichao Wu | P. Luo | Xinjiang Wang | Xiaolin Hu | Ding Liang | Qiushan Guo | Zhipeng Yu
[1] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[2] Guocong Song,et al. Collaborative Learning for Deep Neural Networks , 2018, NeurIPS.
[3] Rich Caruana,et al. Do Deep Nets Really Need to be Deep? , 2013, NIPS.
[4] Ali Farhadi,et al. Label Refinery: Improving ImageNet Classification through Label Progression , 2018, ArXiv.
[5] Junmo Kim,et al. A Gift from Knowledge Distillation: Fast Optimization, Network Minimization and Transfer Learning , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[6] Graham W. Taylor,et al. Improved Regularization of Convolutional Neural Networks with Cutout , 2017, ArXiv.
[7] Nikos Komodakis,et al. Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer , 2016, ICLR.
[8] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[9] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[10] Mark Sandler,et al. MobileNetV2: Inverted Residuals and Linear Bottlenecks , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[11] Jin Young Choi,et al. Knowledge Distillation with Adversarial Samples Supporting Decision Boundary , 2018, AAAI.
[12] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[13] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[14] Yoshua Bengio,et al. FitNets: Hints for Thin Deep Nets , 2014, ICLR.
[15] Wei Tang,et al. Ensembling neural networks: Many could be better than all , 2002, Artif. Intell..
[16] Ludmila I. Kuncheva,et al. Measures of Diversity in Classifier Ensembles and Their Relationship with the Ensemble Accuracy , 2003, Machine Learning.
[17] Kaiming He,et al. Feature Pyramid Networks for Object Detection , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[18] Yann LeCun,et al. Regularization of Neural Networks using DropConnect , 2013, ICML.
[19] Ross B. Girshick,et al. Mask R-CNN , 2017, 1703.06870.
[20] François Fleuret,et al. Knowledge Transfer with Jacobian Matching , 2018, ICML.
[21] Mark J. van der Laan,et al. The relative performance of ensemble methods with deep convolutional neural networks for image classification , 2017, Journal of applied statistics.
[22] Yan Lu,et al. Relational Knowledge Distillation , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[23] Kilian Q. Weinberger,et al. Deep Networks with Stochastic Depth , 2016, ECCV.
[24] Dumitru Erhan,et al. Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[25] Geoffrey E. Hinton,et al. Large scale distributed neural network training through online distillation , 2018, ICLR.
[26] Kaiming He,et al. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[27] Alan L. Yuille,et al. Snapshot Distillation: Teacher-Student Optimization in One Generation , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[28] Huchuan Lu,et al. Deep Mutual Learning , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[29] David H. Wolpert,et al. Stacked generalization , 1992, Neural Networks.
[30] Xu Lan,et al. Knowledge Distillation by On-the-Fly Native Ensemble , 2018, NeurIPS.
[31] Pietro Perona,et al. Microsoft COCO: Common Objects in Context , 2014, ECCV.
[32] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.