暂无分享,去创建一个
Barnabás Póczos | Sanket Vaibhav Mehta | Jaime Carbonell | Zirui Wang | B. Póczos | J. Carbonell | Zirui Wang
[1] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.
[2] Jeff Johnson,et al. Billion-Scale Similarity Search with GPUs , 2017, IEEE Transactions on Big Data.
[3] Fan-Keng Sun,et al. LAMOL: LAnguage MOdeling for Lifelong Language Learning , 2020, ICLR.
[4] Yoshua Bengio,et al. Toward Training Recurrent Neural Networks for Lifelong Learning , 2018, Neural Computation.
[5] Mona Attariyan,et al. Parameter-Efficient Transfer Learning for NLP , 2019, ICML.
[6] Eunsol Choi,et al. TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension , 2017, ACL.
[7] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[8] R'emi Louf,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[9] Jiwon Kim,et al. Continual Learning with Deep Generative Replay , 2017, NIPS.
[10] Xiang Zhang,et al. Character-level Convolutional Networks for Text Classification , 2015, NIPS.
[11] Colin Raffel,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[12] Jaime G. Carbonell,et al. Towards more Reliable Transfer Learning , 2018, ECML/PKDD.
[13] Richard Socher,et al. The Natural Language Decathlon: Multitask Learning as Question Answering , 2018, ArXiv.
[14] Razvan Pascanu,et al. Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.
[15] Michael McCloskey,et al. Catastrophic Interference in Connectionist Networks: The Sequential Learning Problem , 1989 .
[16] James L. McClelland,et al. Why there are complementary learning systems in the hippocampus and neocortex: insights from the successes and failures of connectionist models of learning and memory. , 1995, Psychological review.
[17] Yee Whye Teh,et al. Progress & Compress: A scalable framework for continual learning , 2018, ICML.
[18] Sebastian Ruder,et al. Episodic Memory in Lifelong Language Learning , 2019, NeurIPS.
[19] Qiang Yang,et al. A Survey on Transfer Learning , 2010, IEEE Transactions on Knowledge and Data Engineering.
[20] Victor S. Lempitsky,et al. Unsupervised Domain Adaptation by Backpropagation , 2014, ICML.
[21] Yoshua Bengio,et al. An Empirical Study of Example Forgetting during Deep Neural Network Learning , 2018, ICLR.
[22] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[23] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[24] Paul N. Bennett,et al. Dual Strategy Active Learning , 2007, ECML.
[25] Sung Ju Hwang,et al. Lifelong Learning with Dynamically Expandable Networks , 2017, ICLR.
[26] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[27] Razvan Pascanu,et al. Progressive Neural Networks , 2016, ArXiv.
[28] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[29] David Rolnick,et al. Experience Replay for Continual Learning , 2018, NeurIPS.
[30] Surya Ganguli,et al. Continual Learning Through Synaptic Intelligence , 2017, ICML.
[31] Marc'Aurelio Ranzato,et al. Gradient Episodic Memory for Continual Learning , 2017, NIPS.
[32] Luca Antiga,et al. Automatic differentiation in PyTorch , 2017 .
[33] Lei Yu,et al. Learning and Evaluating General Linguistic Intelligence , 2019, ArXiv.
[34] Marta Garnelo,et al. Adaptive Posterior Learning: few-shot learning with a surprise-based memory module , 2019, ICLR.
[35] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[36] Jaime G. Carbonell,et al. Characterizing and Avoiding Negative Transfer , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[37] Ankur Bapna,et al. Simple, Scalable Adaptation for Neural Machine Translation , 2019, EMNLP.
[38] Omer Levy,et al. Generalization through Memorization: Nearest Neighbor Language Models , 2020, ICLR.
[39] Eunsol Choi,et al. QuAC: Question Answering in Context , 2018, EMNLP.
[40] Jing Gao,et al. On handling negative transfer and imbalanced distributions in multiple source transfer learning , 2014, SDM.
[41] Stefan Wermter,et al. Continual Lifelong Learning with Neural Networks: A Review , 2019, Neural Networks.
[42] Marc'Aurelio Ranzato,et al. Efficient Lifelong Learning with A-GEM , 2018, ICLR.