Dual Student: Breaking the Limits of the Teacher in Semi-Supervised Learning
暂无分享,去创建一个
Rynson W. H. Lau | Qiong Yan | Rynson W.H. Lau | Jimmy S. J. Ren | Zhanghan Ke | Jimmy Ren | Daoye Wang | Qiong Yan | Zhanghan Ke | Daoye Wang
[1] Geoffrey French,et al. Self-ensembling for domain adaptation , 2017, ArXiv.
[2] Zhi-Hua Zhou. When semi-supervised learning meets ensemble learning , 2011 .
[3] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[4] Il-Chul Moon,et al. Adversarial Dropout for Supervised and Semi-supervised Learning , 2017, AAAI.
[5] Andrew Y. Ng,et al. Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .
[6] Tapani Raiko,et al. Semi-supervised Learning with Ladder Networks , 2015, NIPS.
[7] Augustus Odena,et al. Semi-Supervised Learning with Generative Adversarial Networks , 2016, ArXiv.
[8] Jost Tobias Springenberg,et al. Unsupervised and Semi-supervised Learning with Categorical Generative Adversarial Networks , 2015, ICLR.
[9] Andrew Gordon Wilson,et al. Averaging Weights Leads to Wider Optima and Better Generalization , 2018, UAI.
[10] Boris Polyak,et al. Acceleration of stochastic approximation by averaging , 1992 .
[11] Colin Raffel,et al. Realistic Evaluation of Semi-Supervised Learning Algorithms , 2018, ICLR.
[12] Philip Bachman,et al. Learning with Pseudo-Ensembles , 2014, NIPS.
[13] Ole Winther,et al. Auxiliary Deep Generative Models , 2016, ICML.
[14] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[15] Bhuvana Ramabhadran,et al. Efficient Knowledge Distillation from an Ensemble of Teachers , 2017, INTERSPEECH.
[16] Bo Zhang,et al. Smooth Neighbors on Teacher Graphs for Semi-Supervised Learning , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[17] Xiaojin Zhu,et al. --1 CONTENTS , 2006 .
[18] Zhuowen Tu,et al. Aggregated Residual Transformations for Deep Neural Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[19] Wei Li,et al. WebVision Database: Visual Learning and Understanding from Web Data , 2017, ArXiv.
[20] Andrew Gordon Wilson,et al. There Are Many Consistent Explanations of Unlabeled Data: Why You Should Average , 2018, ICLR.
[21] Alexander Zien,et al. Semi-Supervised Learning , 2006 .
[22] Bo Wang,et al. Deep Co-Training for Semi-Supervised Image Recognition , 2018, ECCV.
[23] Enhua Wu,et al. Squeeze-and-Excitation Networks , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[24] Jocelyn Sietsma,et al. Creating artificial neural networks that generalize , 1991, Neural Networks.
[25] Zhi-Hua Zhou,et al. Tri-net for Semi-Supervised Deep Learning , 2018, IJCAI.
[26] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[27] Sergey Ioffe,et al. Batch Renormalization: Towards Reducing Minibatch Dependence in Batch-Normalized Models , 2017, NIPS.
[28] Avrim Blum,et al. The Bottleneck , 2021, Monopsony Capitalism.
[29] Shin Ishii,et al. Virtual Adversarial Training: A Regularization Method for Supervised and Semi-Supervised Learning , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[30] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[31] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[32] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[33] Harri Valpola,et al. Weight-averaged consistency targets improve semi-supervised deep learning results , 2017, ArXiv.