暂无分享,去创建一个
Greg Mori | Ye Yu | Mei Chen | Yu Gong | Gaurav Mittal | Greg Mori | Gaurav Mittal | Ye Yu | Mei Chen | Yunpeng Gong
[1] Yoshua Bengio,et al. Mutual Information Neural Estimation , 2018, ICML.
[2] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[3] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[4] Il-Chul Moon,et al. Refine Myself by Teaching Myself: Feature Refinement via Self-Knowledge Distillation , 2021, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[5] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[6] Quoc V. Le,et al. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks , 2019, ICML.
[7] William Bialek,et al. Entropy and information in neural spike trains: progress on the sampling problem. , 2003, Physical review. E, Statistical, nonlinear, and soft matter physics.
[8] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[9] A. Kraskov,et al. Estimating mutual information. , 2003, Physical review. E, Statistical, nonlinear, and soft matter physics.
[10] Kilian Q. Weinberger,et al. Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[11] J. Kinney,et al. Equitability, mutual information, and the maximal information coefficient , 2013, Proceedings of the National Academy of Sciences.
[12] Vijay Vasudevan,et al. Learning Transferable Architectures for Scalable Image Recognition , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[13] Chen Change Loy,et al. Learning Lightweight Lane Detection CNNs by Self Attention Distillation , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[14] Pietro Perona,et al. Microsoft COCO: Common Objects in Context , 2014, ECCV.
[15] Nojun Kwak,et al. Feature-map-level Online Adversarial Knowledge Distillation , 2020, ICML.
[16] Xiangyu Zhang,et al. ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[17] Neil D. Lawrence,et al. Variational Information Distillation for Knowledge Transfer , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[18] Xiaolin Hu,et al. Online Knowledge Distillation via Collaborative Learning , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[19] Thomas Brox,et al. Striving for Simplicity: The All Convolutional Net , 2014, ICLR.
[20] Cheng-Lin Liu,et al. Data-Distortion Guided Self-Distillation for Deep Neural Networks , 2019, AAAI.
[21] Yonglong Tian,et al. Contrastive Representation Distillation , 2019, ICLR.
[22] Oriol Vinyals,et al. Representation Learning with Contrastive Predictive Coding , 2018, ArXiv.
[23] Alan L. Yuille,et al. Snapshot Distillation: Teacher-Student Optimization in One Generation , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[24] Michael Tschannen,et al. On Mutual Information Maximization for Representation Learning , 2019, ICLR.
[25] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[26] Naiyan Wang,et al. Like What You Like: Knowledge Distill via Neuron Selectivity Transfer , 2017, ArXiv.
[27] Nikos Komodakis,et al. Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer , 2016, ICLR.
[28] Huchuan Lu,et al. Deep Mutual Learning , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[29] Xu Lan,et al. Knowledge Distillation by On-the-Fly Native Ensemble , 2018, NeurIPS.
[30] M. V. Van Hulle,et al. Edgeworth Approximation of Multivariate Differential Entropy , 2005, Neural Computation.
[31] R Devon Hjelm,et al. Learning Representations by Maximizing Mutual Information Across Views , 2019, NeurIPS.
[32] Kaisheng Ma,et al. Be Your Own Teacher: Improve the Performance of Convolutional Neural Networks via Self Distillation , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[33] Jinwoo Shin,et al. Regularizing Class-Wise Predictions via Self-Knowledge Distillation , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[34] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[35] Yoshua Bengio,et al. FitNets: Hints for Thin Deep Nets , 2014, ICLR.
[36] Phillip Isola,et al. Contrastive Multiview Coding , 2019, ECCV.
[37] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[38] Zhiqiang Shen,et al. MEAL: Multi-Model Ensemble via Adversarial Learning , 2018, AAAI.
[39] Yoshua Bengio,et al. Learning deep representations by mutual information estimation and maximization , 2018, ICLR.