暂无分享,去创建一个
Dongqi Han | Kenji Doya | Jun Tani | K. Doya | J. Tani | Dongqi Han
[1] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[2] M. Botvinick,et al. The hippocampus as a predictive map , 2016 .
[3] Ruben Villegas,et al. Learning Latent Dynamics for Planning from Pixels , 2018, ICML.
[4] Yoshua Bengio,et al. Z-Forcing: Training Stochastic Recurrent Networks , 2017, NIPS.
[5] Thomas H. B. FitzGerald,et al. Disruption of Dorsolateral Prefrontal Cortex Decreases Model-Based in Favor of Model-free Control in Humans , 2013, Neuron.
[6] Christian Osendorfer,et al. Learning Stochastic Recurrent Networks , 2014, NIPS 2014.
[7] Dongqi Han. Self-organization of action hierarchy and compositionality by reinforcement learning with recurrent networks. , 2019 .
[8] Martin A. Riedmiller,et al. Embed to Control: A Locally Linear Latent Dynamics Model for Control from Raw Images , 2015, NIPS.
[9] K. Doya,et al. Neural substrate of dynamic Bayesian inference in the cerebral cortex , 2016, Nature Neuroscience.
[10] Razvan Pascanu,et al. On the difficulty of training recurrent neural networks , 2012, ICML.
[11] Shimon Whiteson,et al. Deep Variational Reinforcement Learning for POMDPs , 2018, ICML.
[12] Jürgen Schmidhuber,et al. Recurrent World Models Facilitate Policy Evolution , 2018, NeurIPS.
[13] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[14] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[15] Jürgen Schmidhuber,et al. Reinforcement Learning in Markovian and Non-Markovian Environments , 1990, NIPS.
[16] Henry Zhu,et al. Soft Actor-Critic Algorithms and Applications , 2018, ArXiv.
[17] Jun Tani,et al. A Novel Predictive-Coding-Inspired Variational RNN Model for Online Prediction and Recognition , 2018, Neural Computation.
[18] P. Dayan,et al. States versus Rewards: Dissociable Neural Prediction Error Signals Underlying Model-Based and Model-Free Reinforcement Learning , 2010, Neuron.
[19] Yoshua Bengio,et al. A Recurrent Latent Variable Model for Sequential Data , 2015, NIPS.
[20] Sergey Levine,et al. Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor , 2018, ICML.
[21] Rémi Munos,et al. Recurrent Experience Replay in Distributed Reinforcement Learning , 2018, ICLR.
[22] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[23] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[24] Guy Lever,et al. Human-level performance in 3D multiplayer games with population-based reinforcement learning , 2018, Science.
[25] Sergey Levine,et al. Model-Based Reinforcement Learning for Atari , 2019, ICLR.
[26] Shinsuke Shimojo,et al. Neural Computations Underlying Arbitration between Model-Based and Model-free Learning , 2013, Neuron.
[27] Carl E. Rasmussen,et al. PILCO: A Model-Based and Data-Efficient Approach to Policy Search , 2011, ICML.
[28] Richard S. Sutton,et al. Neuronlike adaptive elements that can solve difficult learning control problems , 1983, IEEE Transactions on Systems, Man, and Cybernetics.
[29] Sergey Levine,et al. Stochastic Latent Actor-Critic: Deep Reinforcement Learning with a Latent Variable Model , 2019, NeurIPS.
[30] David Silver,et al. Memory-based control with recurrent neural networks , 2015, ArXiv.
[31] Yoshua Bengio,et al. Learning Dynamics Model in Reinforcement Learning by Incorporating the Long Term Future , 2019, ArXiv.
[32] Andrew McCallum,et al. Overcoming Incomplete Perception with Utile Distinction Memory , 1993, ICML.
[33] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[34] Leslie Pack Kaelbling,et al. Planning and Acting in Partially Observable Stochastic Domains , 1998, Artif. Intell..