暂无分享,去创建一个
[1] Remco C. Veltkamp,et al. Saliency Tubes: Visual Explanations for Spatio-Temporal Convolutions , 2019, 2019 IEEE International Conference on Image Processing (ICIP).
[2] Martin Wattenberg,et al. SmoothGrad: removing noise by adding noise , 2017, ArXiv.
[3] Rob Fergus,et al. Visualizing and Understanding Convolutional Networks , 2013, ECCV.
[4] Motoaki Kawanabe,et al. How to Explain Individual Classification Decisions , 2009, J. Mach. Learn. Res..
[5] Thomas Brox,et al. Striving for Simplicity: The All Convolutional Net , 2014, ICLR.
[6] Alexander Binder,et al. Evaluating the Visualization of What a Deep Neural Network Has Learned , 2015, IEEE Transactions on Neural Networks and Learning Systems.
[7] Mubarak Shah,et al. UCF101: A Dataset of 101 Human Actions Classes From Videos in The Wild , 2012, ArXiv.
[8] Bolei Zhou,et al. Learning Deep Features for Discriminative Localization , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[9] Donghyun Kim,et al. Excitation Backprop for RNNs , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[10] Alexandros Stergiou,et al. Class Feature Pyramids for Video Explanation , 2019, 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW).
[11] Andrea Vedaldi,et al. Interpretable Explanations of Black Boxes by Meaningful Perturbation , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[12] Wojciech Samek,et al. Methods for interpreting and understanding deep neural networks , 2017, Digit. Signal Process..
[13] Yarin Gal,et al. Real Time Image Saliency for Black Box Classifiers , 2017, NIPS.
[14] Been Kim,et al. Sanity Checks for Saliency Maps , 2018, NeurIPS.
[15] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.
[16] Andrew Zisserman,et al. Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[17] Fuxin Li,et al. Visualizing Deep Networks by Optimizing with Integrated Gradients , 2019, CVPR Workshops.
[18] Yang Wang,et al. Video Summarization Using Fully Convolutional Sequence Networks , 2018, ECCV.
[19] Yann LeCun,et al. A Closer Look at Spatiotemporal Convolutions for Action Recognition , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[20] John Folkesson,et al. Interpreting video features: a comparison of 3D convolutional networks and convolutional LSTM networks , 2020, ACCV.
[21] Sven Behnke,et al. Interpretable and Fine-Grained Visual Explanations for Convolutional Neural Networks , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[22] Kate Saenko,et al. RISE: Randomized Input Sampling for Explanation of Black-box Models , 2018, BMVC.
[23] Andrea Vedaldi,et al. Understanding Deep Networks via Extremal Perturbations and Smooth Masks , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[24] Alexander Binder,et al. On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation , 2015, PloS one.
[25] Andrew Zisserman,et al. Two-Stream Convolutional Networks for Action Recognition in Videos , 2014, NIPS.
[26] Zhe L. Lin,et al. Top-Down Neural Attention by Excitation Backprop , 2016, International Journal of Computer Vision.
[27] Ankur Taly,et al. Axiomatic Attribution for Deep Networks , 2017, ICML.
[28] Dima Damen,et al. Scaling Egocentric Vision: The EPIC-KITCHENS Dataset , 2018, ArXiv.
[29] Avanti Shrikumar,et al. Learning Important Features Through Propagating Activation Differences , 2017, ICML.
[30] Andrew Zisserman,et al. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps , 2013, ICLR.