Vision Transformer Computation and Resilience for Dynamic Inference
暂无分享,去创建一个
[1] K. Asanović,et al. AuRORA: Virtualized Accelerator Orchestration for Multi-Tenant Workloads , 2023, 2023 56th IEEE/ACM International Symposium on Microarchitecture (MICRO).
[2] S. Keckler,et al. Augmenting Legacy Networks for Flexible Inference , 2023, 2023 IEEE Intelligent Vehicles Symposium (IV).
[3] Jack Choquette. NVIDIA Hopper H100 GPU: Scaling Performance , 2023, IEEE Micro.
[4] Ross B. Girshick,et al. Segment Anything , 2023, 2023 IEEE/CVF International Conference on Computer Vision (ICCV).
[5] Henrique Pondé de Oliveira Pinto,et al. GPT-4 Technical Report , 2023, 2303.08774.
[6] Yu Liu,et al. DETRs with Collaborative Hybrid Assignments Training , 2022, 2023 IEEE/CVF International Conference on Computer Vision (ICCV).
[7] Haoran You,et al. ViTCoD: Vision Transformer Acceleration via Dedicated Algorithm and Accelerator Co-Design , 2022, 2023 IEEE International Symposium on High-Performance Computer Architecture (HPCA).
[8] Xiangyu Zhang,et al. Anchor DETR: Query Design for Transformer-Based Detector , 2022, AAAI.
[9] J. Kautz,et al. Global Context Vision Transformers , 2022, ICML.
[10] C. T. Gray,et al. A 17–95.6 TOPS/W Deep Learning Inference Accelerator with Per-Vector Scaled 4-bit Quantization for Transformers in 5nm , 2022, 2022 IEEE Symposium on VLSI Technology and Circuits (VLSI Technology and Circuits).
[11] KenLi Li,et al. DiVIT: Algorithm and architecture co-design of differential attention in vision transformer , 2022, J. Syst. Archit..
[12] K. Zeng,et al. FPGA-based accelerator for object detection: a comprehensive survey , 2022, Journal of Supercomputing.
[13] Mao Yang,et al. Towards efficient vision transformer inference: a first study of transformers on mobile devices , 2022, HotMobile.
[14] H. Shum,et al. DINO: DETR with Improved DeNoising Anchor Boxes for End-to-End Object Detection , 2022, ICLR.
[15] Hang Su,et al. DAB-DETR: Dynamic Anchor Boxes are Better Queries for DETR , 2022, ICLR.
[16] Armand Joulin,et al. Detecting Twenty-thousand Classes using Image-level Supervision , 2022, ECCV.
[17] Saehoon Kim,et al. Sparse DETR: Efficient End-to-End Object Detection with Learnable Sparsity , 2021, ICLR.
[18] Lu Yuan,et al. Dynamic DETR: End-to-End Object Detection with Dynamic Attention , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[19] Gang Zeng,et al. Conditional DETR for Fast Training Convergence , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[20] David J. Crandall,et al. A Survey on Deep Learning Technique for Video Segmentation , 2021, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[21] Siwei Ma,et al. Post-Training Quantization for Vision Transformer , 2021, NeurIPS.
[22] Anima Anandkumar,et al. SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers , 2021, NeurIPS.
[23] Boxun Li,et al. Efficient DETR: Improving End-to-End Object Detector with Dense Prior , 2021, ArXiv.
[24] Quanfu Fan,et al. CrossViT: Cross-Attention Multi-Scale Vision Transformer for Image Classification , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[25] Xiang Li,et al. Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction without Convolutions , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[26] Gao Huang,et al. Dynamic Neural Networks: A Survey , 2021, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[27] Fahad Shahbaz Khan,et al. Transformers in Vision: A Survey , 2021, ACM Comput. Surv..
[28] Michael R. Lyu,et al. BinaryBERT: Pushing the Limit of BERT Quantization , 2020, ACL.
[29] Tao Xiang,et al. Rethinking Semantic Segmentation from a Sequence-to-Sequence Perspective with Transformers , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[30] Matthieu Cord,et al. Training data-efficient image transformers & distillation through attention , 2020, ICML.
[31] Alexander M. Rush,et al. EdgeBERT: Sentence-Level Energy Optimizations for Latency-Aware Multi-Task NLP Inference , 2020, MICRO.
[32] Xiaogang Wang,et al. End-to-End Object Detection with Adaptive Clustering Transformer , 2020, BMVC.
[33] S. Gelly,et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale , 2020, ICLR.
[34] Bin Li,et al. Deformable DETR: Deformable Transformers for End-to-End Object Detection , 2020, ICLR.
[35] Ke Xu,et al. BERT Loses Patience: Fast and Robust Inference with Early Exit , 2020, NeurIPS.
[36] Song Han,et al. HAT: Hardware-Aware Transformers for Efficient Natural Language Processing , 2020, ACL.
[37] Tom B. Brown,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[38] Nicolas Usunier,et al. End-to-End Object Detection with Transformers , 2020, ECCV.
[39] Jimmy J. Lin,et al. DeeBERT: Dynamic Early Exiting for Accelerating BERT Inference , 2020, ACL.
[40] Peng Zhou,et al. FastBERT: a Self-distilling BERT with Adaptive Inference Time , 2020, ACL.
[41] Tianlong Chen,et al. Triple Wins: Boosting Accuracy, Robustness and Efficiency Together by Enabling Input-Adaptive Inference , 2020, ICLR.
[42] William J. Dally,et al. MAGNet: A Modular Accelerator Generator for Neural Networks , 2019, 2019 IEEE/ACM International Conference on Computer-Aided Design (ICCAD).
[43] Peter J. Liu,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[44] Moshe Wasserblat,et al. Q8BERT: Quantized 8Bit BERT , 2019, 2019 Fifth Workshop on Energy Efficient Machine Learning and Cognitive Computing - NeurIPS Edition (EMC2-NIPS).
[45] Thomas Wolf,et al. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter , 2019, ArXiv.
[46] Chuang Gan,et al. Once for All: Train One Network and Specialize it for Efficient Deployment , 2019, ICLR.
[47] Yu Cheng,et al. Patient Knowledge Distillation for BERT Model Compression , 2019, EMNLP.
[48] Khan M. Iftekharuddin,et al. Survey on Deep Neural Networks in Speech and Vision Systems , 2019, Neurocomputing.
[49] Yue Wang,et al. Dual Dynamic Inference: Enabling More Efficient, Adaptive, and Controllable Deep Inference , 2019, IEEE Journal of Selected Topics in Signal Processing.
[50] Quoc V. Le,et al. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks , 2019, ICML.
[51] Fedor Moiseev,et al. Analyzing Multi-Head Self-Attention: Specialized Heads Do the Heavy Lifting, the Rest Can Be Pruned , 2019, ACL.
[52] Jieping Ye,et al. Object Detection in 20 Years: A Survey , 2019, Proceedings of the IEEE.
[53] Omer Levy,et al. Are Sixteen Heads Really Better than One? , 2019, NeurIPS.
[54] Klaus C. J. Dietmayer,et al. Deep Multi-Modal Object Detection and Semantic Segmentation for Autonomous Driving: Datasets, Methods, and Challenges , 2019, IEEE Transactions on Intelligent Transportation Systems.
[55] Tudor Dumitras,et al. Shallow-Deep Networks: Understanding and Mitigating Network Overthinking , 2018, ICML.
[56] Mingyu Gao,et al. Interstellar: Using Halide's Scheduling Language to Analyze DNN Accelerators , 2018, ASPLOS.
[57] Wei Sun,et al. Methods and datasets on semantic segmentation: A review , 2018, Neurocomputing.
[58] Yuning Jiang,et al. Unified Perceptual Parsing for Scene Understanding , 2018, ECCV.
[59] Lukasz Kaiser,et al. Universal Transformers , 2018, ICLR.
[60] Martin Jägersand,et al. A Comparative Study of Real-Time Semantic Segmentation for Autonomous Driving , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[61] Bo Chen,et al. NetAdapt: Platform-Aware Neural Network Adaptation for Mobile Applications , 2018, ECCV.
[62] Xin Wang,et al. SkipNet: Learning Dynamic Routing in Convolutional Networks , 2017, ECCV.
[63] Yoshua Bengio,et al. Residual Connections Encourage Iterative Inference , 2017, ICLR.
[64] Bolei Zhou,et al. Scene Parsing through ADE20K Dataset , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[65] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[66] Ross B. Girshick,et al. Mask R-CNN , 2017, 1703.06870.
[67] Li Zhang,et al. Spatially Adaptive Computation Time for Residual Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[68] H. T. Kung,et al. BranchyNet: Fast inference via early exiting from deep neural networks , 2016, 2016 23rd International Conference on Pattern Recognition (ICPR).
[69] Sebastian Ramos,et al. The Cityscapes Dataset for Semantic Urban Scene Understanding , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[70] Alex Graves,et al. Adaptive Computation Time for Recurrent Neural Networks , 2016, ArXiv.
[71] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[72] Wei Liu,et al. SSD: Single Shot MultiBox Detector , 2015, ECCV.
[73] Sergey Ioffe,et al. Rethinking the Inception Architecture for Computer Vision , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[74] Ali Farhadi,et al. You Only Look Once: Unified, Real-Time Object Detection , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[75] Kaiming He,et al. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[76] Ross B. Girshick,et al. Fast R-CNN , 2015, 1504.08083.
[77] Trevor Darrell,et al. Fully convolutional networks for semantic segmentation , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[78] John Tran,et al. cuDNN: Efficient Primitives for Deep Learning , 2014, ArXiv.
[79] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[80] Pietro Perona,et al. Microsoft COCO: Common Objects in Context , 2014, ECCV.
[81] Trevor Darrell,et al. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation , 2013, 2014 IEEE Conference on Computer Vision and Pattern Recognition.
[82] Fei-Fei Li,et al. ImageNet: A large-scale hierarchical image database , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.
[83] GPT-4V(ision) System Card , 2023 .
[84] P. Varshney,et al. Resource Scheduling for Multi-Target Tracking in Multi-Radar Systems With Imperfect Detection , 2022, IEEE Transactions on Signal Processing.
[85] Stephen Lin,et al. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[86] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[87] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[88] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .