Modeling Latent Attention Within Neural Networks
暂无分享,去创建一个
Lawson L. S. Wong | Christopher Grimm | Michael L. Littman | Dilip Arumugam | David Abel | Siddharth Karamcheti | Lawson L.S. Wong | M. Littman | Siddharth Karamcheti | David Abel | Dilip Arumugam | Christopher Grimm
[1] G. David Garson,et al. Interpreting neural-network connection weights , 1991 .
[2] Tamás D. Gedeon,et al. Data Mining of Inputs: Analysing Magnitude and Functional Measures , 1997, Int. J. Neural Syst..
[3] Derek Partridge,et al. Assessing the Impact of Input Features in a Feedforward Neural Network , 2000, Neural Computing & Applications.
[4] Alfonso Palmer,et al. Numeric sensitivity analysis applied to feedforward neural networks , 2003, Neural Computing & Applications.
[5] Laurenz Wiskott,et al. Slow feature analysis yields a rich repertoire of complex cell properties. , 2005, Journal of vision.
[6] M. Gevrey,et al. Two-way interaction of input variables in the sensitivity analysis of neural network models , 2006 .
[7] M.A. Mazurowski,et al. Limitations of sensitivity analysis for neural networks in cases with dependent inputs , 2006, 2006 IEEE International Conference on Computational Cybernetics.
[8] Eliezer Yudkowsky. Artificial Intelligence as a Positive and Negative Factor in Global Risk , 2006 .
[9] Marko Robnik-Sikonja,et al. Explaining Classifications For Individual Instances , 2008, IEEE Transactions on Knowledge and Data Engineering.
[10] Pascal Vincent,et al. Visualizing Higher-Layer Features of a Deep Network , 2009 .
[11] Motoaki Kawanabe,et al. How to Explain Individual Classification Decisions , 2009, J. Mach. Learn. Res..
[12] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[13] Graham W. Taylor,et al. Adaptive deconvolutional networks for mid and high level feature learning , 2011, 2011 International Conference on Computer Vision.
[14] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[15] Andrew L. Maas. Rectifier Nonlinearities Improve Neural Network Acoustic Models , 2013 .
[16] Rob Fergus,et al. Visualizing and Understanding Convolutional Networks , 2013, ECCV.
[17] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[18] Alex Graves,et al. Recurrent Models of Visual Attention , 2014, NIPS.
[19] Andrew Zisserman,et al. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps , 2013, ICLR.
[20] Hod Lipson,et al. Understanding Neural Networks Through Deep Visualization , 2015, ArXiv.
[21] Geoffrey E. Hinton,et al. Grammar as a Foreign Language , 2014, NIPS.
[22] Yoshua Bengio,et al. Describing Multimedia Content Using Attention-Based Encoder-Decoder Networks , 2015, IEEE Transactions on Multimedia.
[23] Alex Graves,et al. DRAW: A Recurrent Neural Network For Image Generation , 2015, ICML.
[24] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[25] Phil Blunsom,et al. Teaching Machines to Read and Comprehend , 2015, NIPS.
[26] Hal Daumé,et al. Deep Unordered Composition Rivals Syntactic Methods for Text Classification , 2015, ACL.
[27] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[28] Dumitru Erhan,et al. Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[29] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[30] Alexander J. Smola,et al. Stacked Attention Networks for Image Question Answering , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[31] Richard Socher,et al. Ask Me Anything: Dynamic Memory Networks for Natural Language Processing , 2015, ICML.
[32] Regina Barzilay,et al. Rationalizing Neural Predictions , 2016, EMNLP.
[33] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.
[34] Honglak Lee,et al. Control of Memory, Active Perception, and Action in Minecraft , 2016, ICML.
[35] Jason Yosinski,et al. Multifaceted Feature Visualization: Uncovering the Different Types of Features Learned By Each Neuron in Deep Neural Networks , 2016, ArXiv.
[36] Yash Goyal,et al. Towards Transparent AI Systems: Interpreting Visual Question Answering Models , 2016, 1608.08974.
[37] Max Welling,et al. Visualizing Deep Neural Network Decisions: Prediction Difference Analysis , 2017, ICLR.