暂无分享,去创建一个
Pushmeet Kohli | Edward Grefenstette | Hanjun Dai | Thomas Kipf | Peter W. Battaglia | Yujia Li | Vinícius Flores Zambaldi | Pushmeet Kohli | P. Battaglia | V. Zambaldi | Yujia Li | Thomas Kipf | Edward Grefenstette | H. Dai
[1] Tom Schaul,et al. FeUdal Networks for Hierarchical Reinforcement Learning , 2017, ICML.
[2] Hyeonwoo Noh,et al. Neural Program Synthesis from Diverse Demonstration Videos , 2018, ICML.
[3] Honglak Lee,et al. Zero-Shot Task Generalization with Multi-Task Deep Reinforcement Learning , 2017, ICML.
[4] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[5] Yuval Tassa,et al. DeepMind Control Suite , 2018, ArXiv.
[6] Doina Precup,et al. The Option-Critic Architecture , 2016, AAAI.
[7] Juan Carlos Niebles,et al. Dense-Captioning Events in Videos , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[8] Pushmeet Kohli,et al. CompILE: Compositional Imitation Learning and Execution , 2018, ICML.
[9] Gaurav S. Sukhatme,et al. Multi-Modal Imitation Learning from Unstructured Demonstrations using Generative Adversarial Nets , 2017, NIPS.
[10] Ryan P. Adams,et al. Composing graphical models with neural networks for structured representations and fast inference , 2016, NIPS.
[11] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[12] Yu Zhang,et al. Latent Sequence Decompositions , 2016, ICLR.
[13] Christopher Burgess,et al. beta-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework , 2016, ICLR 2016.
[14] Gerald Tesauro,et al. Learning Abstract Options , 2018, NeurIPS.
[15] Joshua B. Tenenbaum,et al. Hierarchical Deep Reinforcement Learning: Integrating Temporal Abstraction and Intrinsic Motivation , 2016, NIPS.
[16] Dawn Xiaodong Song,et al. Parametrized Hierarchical Procedures for Neural Programming , 2018, ICLR.
[17] Daan Wierstra,et al. Stochastic Backpropagation and Approximate Inference in Deep Generative Models , 2014, ICML.
[18] Doina Precup,et al. Between MDPs and Semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning , 1999, Artif. Intell..
[19] Le Song,et al. Recurrent Hidden Semi-Markov Model , 2017, ICLR.
[20] Samy Bengio,et al. Show and tell: A neural image caption generator , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[21] Alex Graves,et al. Supervised Sequence Labelling , 2012 .
[22] Chong Wang,et al. Sequence Modeling via Segmentations , 2017, ICML.
[23] L. Davachi,et al. What Constitutes an Episode in Episodic Memory? , 2011, Psychological science.
[24] T. Griffiths,et al. A Bayesian framework for word segmentation: Exploring the effects of context , 2009, Cognition.
[25] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[26] Bernard Ghanem,et al. DAPs: Deep Action Proposals for Action Understanding , 2016, ECCV.
[27] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[28] Dan Klein,et al. Modular Multitask Reinforcement Learning with Policy Sketches , 2016, ICML.
[29] Oliver Kroemer,et al. Towards learning hierarchical skills for multi-phase manipulation tasks , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).
[30] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[31] Pieter Abbeel,et al. Stochastic Neural Networks for Hierarchical Reinforcement Learning , 2016, ICLR.
[32] Ion Stoica,et al. Multi-Level Discovery of Deep Options , 2017, ArXiv.
[33] Ion Stoica,et al. DDCO: Discovery of Deep Continuous Options for Robot Learning from Demonstrations , 2017, CoRL.
[34] Jeffrey M. Zacks,et al. Event boundaries in memory and cognition , 2017, Current Opinion in Behavioral Sciences.
[35] Jeffrey M. Zacks,et al. Perceiving, remembering, and communicating structure in events. , 2001, Journal of experimental psychology. General.
[36] Scott Niekum,et al. Incremental Semantically Grounded Learning from Demonstration , 2013, Robotics: Science and Systems.
[37] Ben Poole,et al. Categorical Reparameterization with Gumbel-Softmax , 2016, ICLR.
[38] Jakob Uszkoreit,et al. A Decomposable Attention Model for Natural Language Inference , 2016, EMNLP.
[39] Yee Whye Teh,et al. The Concrete Distribution: A Continuous Relaxation of Discrete Random Variables , 2016, ICLR.
[40] Honglak Lee,et al. Learning Structured Output Representation using Deep Conditional Generative Models , 2015, NIPS.
[41] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[42] Stefan Bauer,et al. Adaptive Skip Intervals: Temporal Abstraction for Recurrent Dynamical Models , 2018, NeurIPS.
[43] Shimon Whiteson,et al. TACO: Learning Task Decomposition via Temporal Alignment for Control , 2018, ICML.
[44] Jeffrey M. Zacks,et al. Constructing Experience: Event Models from Perception to Action , 2017, Trends in Cognitive Sciences.
[45] Shane Legg,et al. IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures , 2018, ICML.
[46] Kenneth A. Norman,et al. Discovering Event Structure in Continuous Narrative Perception and Memory , 2016, Neuron.