Recurrent Knowledge Distillation
暂无分享,去创建一个
[1] Alexandre Boulch. ShaResNet: reducing residual network parameter number by sharing weights , 2017, ArXiv.
[2] Kaiming He,et al. Feature Pyramid Networks for Object Detection , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[3] Junmo Kim,et al. A Gift from Knowledge Distillation: Fast Optimization, Network Minimization and Transfer Learning , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[4] Dacheng Tao,et al. Beyond Filters: Compact Feature Map for Portable Deep Model , 2017, ICML.
[5] Jian Sun,et al. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[6] Yoshua Bengio,et al. Maxout Networks , 2013, ICML.
[7] Yi Li,et al. R-FCN: Object Detection via Region-based Fully Convolutional Networks , 2016, NIPS.
[8] Trevor Darrell,et al. Fully Convolutional Networks for Semantic Segmentation , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[9] William Chan,et al. Transferring knowledge from a RNN to a DNN , 2015, INTERSPEECH.
[10] Ross B. Girshick,et al. Mask R-CNN , 2017, 1703.06870.
[11] Jian Sun,et al. Accelerating Very Deep Convolutional Networks for Classification and Detection , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[12] Mohammed Bennamoun,et al. Resfeats: Residual network based features for image classification , 2016, 2017 IEEE International Conference on Image Processing (ICIP).
[13] Jürgen Schmidhuber,et al. Training Very Deep Networks , 2015, NIPS.
[14] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[15] Serge J. Belongie,et al. Residual Networks are Exponential Ensembles of Relatively Shallow Networks , 2016, ArXiv.
[16] Tianqi Chen,et al. Net2Net: Accelerating Learning via Knowledge Transfer , 2015, ICLR.
[17] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[18] Jürgen Schmidhuber,et al. Highway and Residual Networks learn Unrolled Iterative Estimation , 2016, ICLR.
[19] Dacheng Tao,et al. On Compressing Deep Models by Low Rank and Sparse Decomposition , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[20] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[21] Jitendra Malik,et al. Cross Modal Distillation for Supervision Transfer , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[22] Yoram Singer,et al. Sketching and Neural Networks , 2016, ArXiv.
[23] Martin A. Riedmiller,et al. Improving Deep Neural Networks with Probabilistic Maxout Units , 2013, ICLR.
[24] Victor S. Lempitsky,et al. Fast ConvNets Using Group-Wise Brain Damage , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[25] Yoshua Bengio,et al. FitNets: Hints for Thin Deep Nets , 2014, ICLR.
[26] Tomaso A. Poggio,et al. Bridging the Gaps Between Residual Learning, Recurrent Neural Networks and Visual Cortex , 2016, ArXiv.
[27] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[28] Qiang Chen,et al. Network In Network , 2013, ICLR.
[29] Shih-Fu Chang,et al. An Exploration of Parameter Redundancy in Deep Networks with Circulant Projections , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[30] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[31] Bolei Zhou,et al. Places: A 10 Million Image Database for Scene Recognition , 2018, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[32] Xiaolin Hu,et al. Recurrent convolutional neural network for object recognition , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[33] Nikos Komodakis,et al. Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer , 2016, ICLR.
[34] Antonio Torralba,et al. Recognizing indoor scenes , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.
[35] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[36] Rich Caruana,et al. Do Deep Nets Really Need to be Deep? , 2013, NIPS.
[37] Jürgen Schmidhuber,et al. Highway Networks , 2015, ArXiv.