ViT-Calibrator: Decision Stream Calibration for Vision Transformer
暂无分享,去创建一个
Zunlei Feng | Jie Lei | Min-Gyoo Song | Lechao Cheng | Tian Qiu | Lin Chen | Zhijie Jia
[1] Ibrahim M. Alabdulmohsin,et al. FlexiViT: One Model for All Patch Sizes , 2022, 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[2] Ledell Yu Wu,et al. EVA: Exploring the Limits of Masked Visual Representation Learning at Scale , 2022, 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[3] Zunlei Feng,et al. Model Doctor: A Simple Gradient Aggregation Strategy for Diagnosing and Treating CNN Classifiers , 2021, AAAI.
[4] Li Dong,et al. BEiT: BERT Pre-Training of Image Transformers , 2021, ICLR.
[5] Quoc V. Le,et al. CoAtNet: Marrying Convolution and Attention for All Data Sizes , 2021, NeurIPS.
[6] Zunlei Feng,et al. Edge-competing Pathological Liver Vessel Segmentation with Limited Labels , 2021, AAAI.
[7] Julien Mairal,et al. Emerging Properties in Self-Supervised Vision Transformers , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[8] Matthieu Cord,et al. Going deeper with Image Transformers , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[9] Quanfu Fan,et al. CrossViT: Cross-Attention Multi-Scale Vision Transformer for Image Classification , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[10] Ari S. Morcos,et al. ConViT: improving vision transformers with soft convolutional inductive biases , 2021, ICML.
[11] Xiang Li,et al. Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction without Convolutions , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[12] Francis E. H. Tay,et al. Tokens-to-Token ViT: Training Vision Transformers from Scratch on ImageNet , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[13] Matthieu Cord,et al. Training data-efficient image transformers & distillation through attention , 2020, ICML.
[14] Lior Wolf,et al. Transformer Interpretability Beyond Attention Visualization , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[15] S. Gelly,et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale , 2020, ICLR.
[16] Stephen Lin,et al. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[17] Willem Zuidema,et al. Quantifying Attention Flow in Transformers , 2020, ACL.
[18] Andrea Vedaldi,et al. Understanding Deep Networks via Extremal Perturbations and Smooth Masks , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[19] Le Song,et al. L-Shapley and C-Shapley: Efficient Model Interpretation for Structured Data , 2018, ICLR.
[20] Yang Wang,et al. Manifold: A Model-Agnostic Framework for Interpretation and Diagnosis of Machine Learning Models , 2018, IEEE Transactions on Visualization and Computer Graphics.
[21] Bolei Zhou,et al. Interpreting Deep Visual Representations via Network Dissection , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[22] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[23] Jiajun Bu,et al. Understanding the Prediction Process of Deep Networks by Forests , 2018, 2018 IEEE Fourth International Conference on Multimedia Big Data (BigMM).
[24] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[25] Osbert Bastani,et al. Interpretability via Model Extraction , 2017, ArXiv.
[26] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[27] Scott Lundberg,et al. A Unified Approach to Interpreting Model Predictions , 2017, NIPS.
[28] Yarin Gal,et al. Real Time Image Saliency for Black Box Classifiers , 2017, NIPS.
[29] Yindalon Aphinyanagphongs,et al. A Workflow for Visual Diagnostics of Binary Classifiers using Instance-Level Explanations , 2017, 2017 IEEE Conference on Visual Analytics Science and Technology (VAST).
[30] Andrea Vedaldi,et al. Interpretable Explanations of Black Boxes by Meaningful Perturbation , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[31] Bram van Ginneken,et al. A survey on deep learning in medical image analysis , 2017, Medical Image Anal..
[32] Zhe L. Lin,et al. Top-Down Neural Attention by Excitation Backprop , 2016, International Journal of Computer Vision.
[33] Kenney Ng,et al. Interacting with Predictions: Visual Inspection of Black-box Machine Learning Models , 2016, CHI.
[34] Xin Zhang,et al. End to End Learning for Self-Driving Cars , 2016, ArXiv.
[35] Bolei Zhou,et al. Learning Deep Features for Discriminative Localization , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[36] Ran Gilad-Bachrach,et al. Debugging Machine Learning Models , 2016 .
[37] Ashish Kapoor,et al. FeatureInsight: Visual support for error-driven feature ideation in text classification , 2015, 2015 IEEE Conference on Visual Analytics Science and Technology (VAST).
[38] Weng-Keen Wong,et al. Principles of Explanatory Debugging to Personalize Interactive Machine Learning , 2015, IUI.
[39] Rosane Minghim,et al. An Approach to Supporting Incremental Visual Data Classification , 2015, IEEE Transactions on Visualization and Computer Graphics.
[40] Rob Fergus,et al. Visualizing and Understanding Convolutional Networks , 2013, ECCV.
[41] Weng-Keen Wong,et al. Explanatory Debugging: Supporting End-User Debugging of Machine-Learned Programs , 2010, VL/HCC.
[42] Fei-Fei Li,et al. ImageNet: A large-scale hierarchical image database , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.
[43] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[44] J. Démos. Getting Started with Neurofeedback , 2005 .
[45] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.