CoDiNet: Path Distribution Modeling With Consistency and Diversity for Dynamic Routing
暂无分享,去创建一个
Huanyu Wang | Xi Li | Zequn Qin | Songyuan Li | Huanyu Wang | Songyuan Li | Xi Li | Zequn Qin
[1] Niraj K. Jha,et al. Fully Dynamic Inference With Deep Neural Networks , 2020, IEEE Transactions on Emerging Topics in Computing.
[2] Kate Saenko,et al. AR-Net: Adaptive Frame Resolution for Efficient Action Recognition , 2020, ECCV.
[3] Matti Pietikäinen,et al. Dynamic Group Convolution for Accelerating Convolutional Neural Networks , 2020, ECCV.
[4] William Robson Schwartz,et al. Discriminative Layer Pruning for Convolutional Neural Networks , 2020, IEEE Journal of Selected Topics in Signal Processing.
[5] Fei Sun,et al. Computation on Sparse Neural Networks: an Inspiration for Future Hardware , 2020, ArXiv.
[6] Mihir Jain,et al. TimeGate: Conditional Gating of Segments in Long-range Activities , 2020, ArXiv.
[7] Zheng Zhang,et al. Spatially Adaptive Inference with Stochastic Feature Sampling and Interpolation , 2020, ECCV.
[8] Le Yang,et al. Resolution Adaptive Networks for Efficient Inference , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[9] Lu Yuan,et al. Dynamic Convolution: Attention Over Convolution Kernels , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[10] Tinne Tuytelaars,et al. Dynamic Convolutions: Exploiting Spatial Sparsity for Faster Inference , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[11] Ross B. Girshick,et al. Momentum Contrast for Unsupervised Visual Representation Learning , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[12] Chuang Gan,et al. Once for All: Train One Network and Specialize it for Efficient Deployment , 2019, ICLR.
[13] Yue Wang,et al. Dual Dynamic Inference: Enabling More Efficient, Adaptive, and Controllable Deep Inference , 2019, IEEE Journal of Selected Topics in Signal Processing.
[14] Jiwen Lu,et al. Runtime Network Routing for Efficient Image Classification , 2019, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[15] Shifeng Zhang,et al. DARTS+: Improved Differentiable Architecture Search with Early Stopping , 2019, ArXiv.
[16] Ruigang Yang,et al. Improved Techniques for Training Adaptive Deep Networks , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[17] Pengyi Zhang,et al. SlimYOLOv3: Narrower, Faster and Better for Real-Time UAV Applications , 2019, 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW).
[18] Junjie Yan,et al. Dynamic Recursive Neural Network , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[19] Quoc V. Le,et al. AutoAugment: Learning Augmentation Strategies From Data , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[20] Yu Cao,et al. Efficient Network Construction Through Structural Plasticity , 2019, IEEE Journal on Emerging and Selected Topics in Circuits and Systems.
[21] Quoc V. Le,et al. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks , 2019, ICML.
[22] Qi Tian,et al. Progressive Differentiable Architecture Search: Bridging the Depth Gap Between Search and Evaluation , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[23] Thomas S. Huang,et al. Universally Slimmable Networks and Improved Training Techniques , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[24] Ning Xu,et al. Slimmable Neural Networks , 2018, ICLR.
[25] Yuandong Tian,et al. FBNet: Hardware-Aware Efficient ConvNet Design via Differentiable Neural Architecture Search , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[26] Alexander Wong,et al. Dynamic Representations Toward Efficient Inference on Deep Neural Networks by Decision Gates , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[27] Song Han,et al. ProxylessNAS: Direct Neural Architecture Search on Target Task and Hardware , 2018, ICLR.
[28] Cheng-Zhong Xu,et al. Dynamic Channel Pruning: Feature Boosting and Suppression , 2018, ICLR.
[29] Bo Chen,et al. MnasNet: Platform-Aware Neural Architecture Search for Mobile , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[30] Serge J. Belongie,et al. Convolutional Networks with Adaptive Inference Graphs , 2017, International Journal of Computer Vision.
[31] Noam Shazeer,et al. HydraNets: Specialized Dynamic Architectures for Efficient Inference , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[32] Dan Alistarh,et al. Model compression via distillation and quantization , 2018, ICLR.
[33] Pavlo Molchanov,et al. IamNN: Iterative and Adaptive Mobile Neural Network for Efficient Image Classification , 2018, ICLR.
[34] Zhiguang Cao,et al. Distilling the Knowledge From Handcrafted Features for Human Activity Recognition , 2018, IEEE Transactions on Industrial Informatics.
[35] Jonathon Shlens,et al. Recurrent Segmentation for Variable Computational Budgets , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[36] Xin Wang,et al. SkipNet: Learning Dynamic Routing in Convolutional Networks , 2017, ECCV.
[37] Kilian Q. Weinberger,et al. CondenseNet: An Efficient DenseNet Using Learned Group Convolutions , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[38] Larry S. Davis,et al. BlockDrop: Dynamic Inference Paths in Residual Networks , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[39] Larry S. Davis,et al. NISP: Pruning Networks Using Neuron Importance Score Propagation , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[40] Kilian Q. Weinberger,et al. Multi-Scale Dense Networks for Resource Efficient Image Classification , 2017, ICLR.
[41] Tony X. Han,et al. Learning Efficient Object Detection Models with Knowledge Distillation , 2017, NIPS.
[42] Jianxin Wu,et al. ThiNet: A Filter Level Pruning Method for Deep Neural Network Compression , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[43] Xiangyu Zhang,et al. Channel Pruning for Accelerating Very Deep Neural Networks , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[44] Yi Yang,et al. More is Less: A More Complicated Network with Less Inference Complexity , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[45] Li Zhang,et al. Spatially Adaptive Computation Time for Residual Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[46] Ben Poole,et al. Categorical Reparameterization with Gumbel-Softmax , 2016, ICLR.
[47] Hanan Samet,et al. Pruning Filters for Efficient ConvNets , 2016, ICLR.
[48] Kilian Q. Weinberger,et al. Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[49] H. T. Kung,et al. BranchyNet: Fast inference via early exiting from deep neural networks , 2016, 2016 23rd International Conference on Pattern Recognition (ICPR).
[50] Yiran Chen,et al. Learning Structured Sparsity in Deep Neural Networks , 2016, NIPS.
[51] Serge J. Belongie,et al. Residual Networks Behave Like Ensembles of Relatively Shallow Networks , 2016, NIPS.
[52] Alex Graves,et al. Adaptive Computation Time for Recurrent Neural Networks , 2016, ArXiv.
[53] Jian Cheng,et al. Quantized Convolutional Neural Networks for Mobile Devices , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[54] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[55] Hugo Larochelle,et al. Dynamic Capacity Networks , 2015, ICML.
[56] Roberto Cipolla,et al. Training CNNs with Low-Rank Filters for Efficient Image Classification , 2015, ICLR.
[57] Xiaogang Wang,et al. Convolutional neural networks with low-rank regularization , 2015, ICLR.
[58] Song Han,et al. Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.
[59] Pushmeet Kohli,et al. PerforatedCNNs: Acceleration through Elimination of Redundant Convolutions , 2015, NIPS.
[60] Martial Hebert,et al. Efficient Feature Group Sequencing for Anytime Linear Prediction , 2014, UAI.
[61] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[62] Andrew Zisserman,et al. Speeding up Convolutional Neural Networks with Low Rank Expansions , 2014, BMVC.
[63] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[64] Yann LeCun,et al. Regularization of Neural Networks using DropConnect , 2013, ICML.
[65] Lev Reyzin,et al. Boosting on a Budget: Sampling for Feature-Efficient Prediction , 2011, ICML.
[66] Andrew Y. Ng,et al. Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .
[67] Fei-Fei Li,et al. ImageNet: A large-scale hierarchical image database , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.
[68] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .