暂无分享,去创建一个
Wenhao Ding | Baiming Chen | Mengdi Xu | Zuxin Liu | Ding Zhao | Jiacheng Zhu | Wenhao Ding | Ding Zhao | Baiming Chen | Zuxin Liu | Mengdi Xu | Jiacheng Zhu
[1] Junsoo Ha,et al. A Neural Dirichlet Process Mixture Model for Task-Free Continual Learning , 2020, ICLR.
[2] Yee Whye Teh,et al. Attentive Neural Processes , 2019, ICLR.
[3] Katja Hofmann,et al. Meta Reinforcement Learning with Latent Variable Gaussian Processes , 2018, UAI.
[4] Svetha Venkatesh,et al. Streaming Variational Inference for Dirichlet Process Mixtures , 2015, ACML.
[5] David B. Dahl,et al. Sequentially-Allocated Merge-Split Sampler for Conjugate and Nonconjugate Dirichlet Process Mixture Models , 2005 .
[6] Sergey Levine,et al. Learning to Adapt in Dynamic, Real-World Environments through Meta-Reinforcement Learning , 2018, ICLR.
[7] Tinne Tuytelaars,et al. Task-Free Continual Learning , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[8] Peter Stone,et al. Autonomous agents modelling other agents: A comprehensive survey and open problems , 2017, Artif. Intell..
[9] Emily B. Fox,et al. Streaming Variational Inference for Bayesian Nonparametric Mixture Models , 2014, AISTATS.
[10] Andre Wibisono,et al. Streaming Variational Bayes , 2013, NIPS.
[11] Bing Liu,et al. Lifelong machine learning: a paradigm for continuous learning , 2017, Frontiers of Computer Science.
[12] Sebastian Ruder,et al. An Overview of Multi-Task Learning in Deep Neural Networks , 2017, ArXiv.
[13] Richard E. Turner,et al. Variational Continual Learning , 2017, ICLR.
[14] Zeb Kurth-Nelson,et al. Learning to reinforcement learn , 2016, CogSci.
[15] Dahua Lin,et al. Online Learning of Nonparametric Mixture Models via Sequential Variational Approximation , 2013, NIPS.
[16] Alan Fern,et al. Multi-task reinforcement learning: a hierarchical Bayesian approach , 2007, ICML '07.
[17] Stefan Schaal,et al. 2008 Special Issue: Reinforcement learning of motor skills with policy gradients , 2008 .
[18] Carl E. Rasmussen,et al. Gaussian processes for machine learning , 2005, Adaptive computation and machine learning.
[19] Andrew Gordon Wilson,et al. GPyTorch: Blackbox Matrix-Matrix Gaussian Process Inference with GPU Acceleration , 2018, NeurIPS.
[20] Yishay Mansour,et al. An Information-Theoretic Analysis of Hard and Soft Assignment Methods for Clustering , 1997, UAI.
[21] Yuval Tassa,et al. MuJoCo: A physics engine for model-based control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[22] Pieter Abbeel,et al. Evolved Policy Gradients , 2018, NeurIPS.
[23] Philip C. Candy,et al. Self-Direction for Lifelong Learning: A Comprehensive Guide to Theory and Practice , 1991 .
[24] Michael I. Jordan,et al. A Sticky HDP-HMM With Application to Speaker Diarization , 2009, 0905.2592.
[25] Radford M. Neal,et al. A Split-Merge Markov chain Monte Carlo Procedure for the Dirichlet Process Mixture Model , 2004 .
[26] Sergey Levine,et al. Deep Online Learning via Meta-Learning: Continual Adaptation for Model-Based RL , 2018, ICLR.
[27] Yee Whye Teh,et al. Conditional Neural Processes , 2018, ICML.
[28] Radford M. Neal. Markov Chain Sampling Methods for Dirichlet Process Mixture Models , 2000 .
[29] Jean-Philippe Vert,et al. Clustered Multi-Task Learning: A Convex Formulation , 2008, NIPS.
[30] Sung Ju Hwang,et al. Lifelong Learning with Dynamically Expandable Networks , 2017, ICLR.
[31] Sergey Levine,et al. Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynamics Models , 2018, NeurIPS.
[32] Wojciech Zaremba,et al. OpenAI Gym , 2016, ArXiv.
[33] Joseph J. Lim,et al. Multimodal Model-Agnostic Meta-Learning via Task-Aware Modulation , 2019, NeurIPS.
[34] Qiang Yang,et al. An Overview of Multi-task Learning , 2018 .
[35] Nhat Ho,et al. On posterior contraction of parameters and interpretability in Bayesian mixture modeling , 2019, Bernoulli.
[36] Dustin Tran,et al. Variational Gaussian Process , 2015, ICLR.
[37] Massimiliano Pontil,et al. Regularized multi--task learning , 2004, KDD.
[38] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[39] Yoshua Bengio,et al. The effects of negative adaptation in Model-Agnostic Meta-Learning , 2018, ArXiv.
[40] Peter L. Bartlett,et al. RL$^2$: Fast Reinforcement Learning via Slow Reinforcement Learning , 2016, ArXiv.
[41] Thomas L. Griffiths,et al. Reconciling meta-learning and continual learning with online mixtures of tasks , 2018, NeurIPS.
[42] P. Candy,et al. Self-Direction for Lifelong Learning , 1993 .
[43] Joshua B. Tenenbaum,et al. Human-level concept learning through probabilistic program induction , 2015, Science.
[44] Michael I. Jordan,et al. Variational inference for Dirichlet process mixtures , 2006 .
[45] Yee Whye Teh,et al. Meta-Learning surrogate models for sequential decision making , 2019, ArXiv.
[46] Carl E. Rasmussen,et al. Infinite Mixtures of Gaussian Process Experts , 2001, NIPS.
[47] Tamim Asfour,et al. ProMP: Proximal Meta-Policy Search , 2018, ICLR.
[48] Michalis K. Titsias,et al. Variational Learning of Inducing Variables in Sparse Gaussian Processes , 2009, AISTATS.
[49] Wenshuo Wang,et al. Recurrent Attentive Neural Process for Sequential Data , 2019, ArXiv.
[50] Yee Whye Teh,et al. Functional Regularisation for Continual Learning using Gaussian Processes , 2019, ICLR.