暂无分享,去创建一个
Marc G. Bellemare | Rui Wang | Felipe Petroski Such | Jeff Clune | Joel Lehman | Rosanne Liu | Vashisht Madhavan | Yulun Li | Ludwig Schubert | Pablo Samuel Castro | Marc Bellemare | J. Clune | P. S. Castro | J. Lehman | Rosanne Liu | F. Such | Ludwig Schubert | Vashisht Madhavan | Rui Wang | Yulun Li
[1] Giulio Sandini,et al. 2nd European conference on computer vision , 1992, Image Vis. Comput..
[2] Deborah Silver,et al. Feature Visualization , 1994, Scientific Visualization.
[3] Geoffrey E. Hinton,et al. Visualizing Data using t-SNE , 2008 .
[4] Pascal Vincent,et al. Visualizing Higher-Layer Features of a Deep Network , 2009 .
[5] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[6] Yoshua Bengio,et al. How transferable are features in deep neural networks? , 2014, NIPS.
[7] Trevor Darrell,et al. Caffe: Convolutional Architecture for Fast Feature Embedding , 2014, ACM Multimedia.
[8] Andrew Zisserman,et al. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps , 2013, ICLR.
[9] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[10] Hod Lipson,et al. Understanding Neural Networks Through Deep Visualization , 2015, ArXiv.
[11] Mikhail Pavlov,et al. Deep Attention Recurrent Q-Network , 2015, ArXiv.
[12] Sergey Levine,et al. Trust Region Policy Optimization , 2015, ICML.
[13] Geoffrey E. Hinton,et al. Deep Learning , 2015, Nature.
[14] Hod Lipson,et al. Convergent Learning: Do different neural networks learn the same representations? , 2015, FE@NIPS.
[15] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[16] Honglak Lee,et al. Action-Conditional Video Prediction using Deep Networks in Atari Games , 2015, NIPS.
[17] Peter Stone,et al. Deep Recurrent Q-Learning for Partially Observable MDPs , 2015, AAAI Fall Symposia.
[18] Marc G. Bellemare,et al. The Arcade Learning Environment: An Evaluation Platform for General Agents (Extended Abstract) , 2012, IJCAI.
[19] Tom Schaul,et al. Dueling Network Architectures for Deep Reinforcement Learning , 2015, ICML.
[20] Katja Hofmann,et al. Memory Lens: How Much Memory Does an Agent Use? , 2016, ArXiv.
[21] Alex Graves,et al. Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.
[22] Shie Mannor,et al. Graying the black box: Understanding DQNs , 2016, ICML.
[23] Andrea Vedaldi,et al. Visualizing Deep Convolutional Neural Networks Using Natural Pre-images , 2015, International Journal of Computer Vision.
[24] Jason Yosinski,et al. Multifaceted Feature Visualization: Uncovering the Different Types of Features Learned By Each Neuron in Deep Neural Networks , 2016, ArXiv.
[25] Thomas Brox,et al. Synthesizing the preferred inputs for neurons in neural networks via deep generator networks , 2016, NIPS.
[26] Yoshua Bengio,et al. Plug & Play Generative Networks: Conditional Iterative Generation of Images in Latent Space , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[27] Alexei A. Efros,et al. Curiosity-Driven Exploration by Self-Supervised Prediction , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[28] Zeb Kurth-Nelson,et al. Learning to reinforcement learn , 2016, CogSci.
[29] Xi Chen,et al. Evolution Strategies as a Scalable Alternative to Reinforcement Learning , 2017, ArXiv.
[30] Jascha Sohl-Dickstein,et al. SVCCA: Singular Vector Canonical Correlation Analysis for Deep Learning Dynamics and Interpretability , 2017, NIPS.
[31] Kenneth O. Stanley,et al. Deep Neuroevolution: Genetic Algorithms Are a Competitive Alternative for Training Deep Neural Networks for Reinforcement Learning , 2017, ArXiv.
[32] Marc G. Bellemare,et al. A Distributional Perspective on Reinforcement Learning , 2017, ICML.
[33] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[34] Tom Schaul,et al. Reinforcement Learning with Unsupervised Auxiliary Tasks , 2016, ICLR.
[35] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[36] Tom Schaul,et al. Rainbow: Combining Improvements in Deep Reinforcement Learning , 2017, AAAI.
[37] David Budden,et al. Distributed Prioritized Experience Replay , 2018, ICLR.
[38] Kenneth O. Stanley,et al. Improving Exploration in Evolution Strategies for Deep Reinforcement Learning via a Population of Novelty-Seeking Agents , 2017, NeurIPS.
[39] Arvind Satyanarayan,et al. The Building Blocks of Interpretability , 2018 .
[40] Katia P. Sycara,et al. Transparency and Explanation in Deep Reinforcement Learning Neural Networks , 2018, AIES.
[41] Shane Legg,et al. IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures , 2018, ICML.
[42] Kenneth O. Stanley,et al. ES is more than just a traditional finite-difference approximator , 2017, GECCO.
[43] Nicola Pezzotti,et al. Differentiable Image Parameterizations , 2018, Distill.
[44] Marc G. Bellemare,et al. Dopamine: A Research Framework for Deep Reinforcement Learning , 2018, ArXiv.
[45] Jonathan Dodge,et al. Visualizing and Understanding Atari Agents , 2017, ICML.
[46] Lior Wolf,et al. Visual Analogies between Atari Games for Studying Transfer Learning in RL , 2018, ArXiv.
[47] Tom Schaul,et al. Deep Q-learning From Demonstrations , 2017, AAAI.
[48] Katia P. Sycara,et al. Towards Better Interpretability in Deep Q-Networks , 2018, AAAI.