Harmonized Dense Knowledge Distillation Training for Multi-Exit Architectures
暂无分享,去创建一个
[1] Yingming Li,et al. Gradient Deconfliction-Based Training For Multi-Exit Architectures , 2020, 2020 IEEE International Conference on Image Processing (ICIP).
[2] Hugo Larochelle,et al. Optimization as a Model for Few-Shot Learning , 2016, ICLR.
[3] Jiwen Lu,et al. Runtime Neural Pruning , 2017, NIPS.
[4] Yafei Dai,et al. MSD: Multi-Self-Distillation Learning via Multi-classifiers within Deep Neural Networks , 2019, ArXiv.
[5] Anastasios Tefas,et al. Efficient adaptive inference for deep convolutional neural networks using hierarchical early exits , 2020, Pattern Recognit..
[6] Ge Wang,et al. Structurally-Sensitive Multi-Scale Deep Neural Network for Low-Dose CT Denoising , 2018, IEEE Access.
[7] Bo Chen,et al. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications , 2017, ArXiv.
[8] Yoshua Bengio,et al. FitNets: Hints for Thin Deep Nets , 2014, ICLR.
[9] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[10] Jia Deng,et al. Dynamic Deep Neural Networks: Optimizing Accuracy-Efficiency Trade-offs by Selective Execution , 2017, AAAI.
[11] Xin Wang,et al. SkipNet: Learning Dynamic Routing in Convolutional Networks , 2017, ECCV.
[12] Kaiming He,et al. Accurate, Large Minibatch SGD: Training ImageNet in 1 Hour , 2017, ArXiv.
[13] Christoph H. Lampert,et al. Distillation-Based Training for Multi-Exit Architectures , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[14] Ruigang Yang,et al. Improved Techniques for Training Adaptive Deep Networks , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[15] Song Han,et al. Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.
[16] Yan Wang,et al. Resource Aware Person Re-identification Across Multiple Resolutions , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[17] Joshua Achiam,et al. On First-Order Meta-Learning Algorithms , 2018, ArXiv.
[18] Kilian Q. Weinberger,et al. Multi-Scale Dense Networks for Resource Efficient Image Classification , 2017, ICLR.
[19] Serge J. Belongie,et al. Convolutional Networks with Adaptive Inference Graphs , 2017, International Journal of Computer Vision.
[20] S. Levine,et al. Gradient Surgery for Multi-Task Learning , 2020, NeurIPS.
[21] Larry S. Davis,et al. BlockDrop: Dynamic Inference Paths in Residual Networks , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[22] Honggang Qi,et al. Multi-Scale Structure-Aware Network for Human Pose Estimation , 2018, ECCV.
[23] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[24] David Duvenaud,et al. Stochastic Hyperparameter Optimization through Hypernetworks , 2018, ArXiv.
[25] Jonathon Shlens,et al. Recurrent Segmentation for Variable Computational Budgets , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[26] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[27] Kilian Q. Weinberger,et al. Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[28] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[29] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[30] Suyog Gupta,et al. To prune, or not to prune: exploring the efficacy of pruning for model compression , 2017, ICLR.
[31] Yuan Yu,et al. TensorFlow: A system for large-scale machine learning , 2016, OSDI.
[32] Yoshua Bengio,et al. BinaryNet: Training Deep Neural Networks with Weights and Activations Constrained to +1 or -1 , 2016, ArXiv.
[33] Ali Farhadi,et al. XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks , 2016, ECCV.
[34] Sergey Levine,et al. Meta-Learning with Implicit Gradients , 2019, NeurIPS.
[35] Venkatesh Saligrama,et al. Adaptive Neural Networks for Efficient Inference , 2017, ICML.
[36] Raquel Urtasun,et al. Graph HyperNetworks for Neural Architecture Search , 2018, ICLR.
[37] Yu Liu,et al. Gradient Harmonized Single-stage Detector , 2018, AAAI.
[38] Kaisheng Ma,et al. Be Your Own Teacher: Improve the Performance of Convolutional Neural Networks via Self Distillation , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[39] H. T. Kung,et al. BranchyNet: Fast inference via early exiting from deep neural networks , 2016, 2016 23rd International Conference on Pattern Recognition (ICPR).