暂无分享,去创建一个
Marc'Aurelio Ranzato | Honglak Lee | Myle Ott | Arthur Szlam | Ann Lee | Lajanugen Logeswaran | Myle Ott | Marc'Aurelio Ranzato | Honglak Lee | Ann Lee | Arthur Szlam | Lajanugen Logeswaran
[1] Lav R. Varshney,et al. CTRL: A Conditional Transformer Language Model for Controllable Generation , 2019, ArXiv.
[2] Massimiliano Pontil,et al. Sparse coding for multitask and transfer learning , 2012, ICML.
[3] Marta R. Costa-jussà,et al. Findings of the 2019 Conference on Machine Translation (WMT19) , 2019, WMT.
[4] Geoffrey E. Hinton,et al. Using Fast Weights to Attend to the Recent Past , 2016, NIPS.
[5] Marc'Aurelio Ranzato,et al. Task-Driven Modular Networks for Zero-Shot Compositional Learning , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[6] Mona Attariyan,et al. Parameter-Efficient Transfer Learning for NLP , 2019, ICML.
[7] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[8] Katja Hofmann,et al. Fast Context Adaptation via Meta-Learning , 2018, ICML.
[9] Richard S. Zemel,et al. Prototypical Networks for Few-shot Learning , 2017, NIPS.
[10] Razvan Pascanu,et al. Meta-Learning with Latent Embedding Optimization , 2018, ICLR.
[11] Iain Murray,et al. BERT and PALs: Projected Attention Layers for Efficient Adaptation in Multi-Task Learning , 2019, ICML.
[12] Honglak Lee,et al. Exploring the structure of a real-time, arbitrary neural artistic stylization network , 2017, BMVC.
[13] Pieter Abbeel,et al. A Simple Neural Attentive Meta-Learner , 2017, ICLR.
[14] Michael Fink,et al. Object Classification from a Single Example Utilizing Class Relevance Metrics , 2004, NIPS.
[15] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[16] Marcin Andrychowicz,et al. Learning to learn by gradient descent by gradient descent , 2016, NIPS.
[17] Hal Daumé,et al. Learning Task Grouping and Overlap in Multi-task Learning , 2012, ICML.
[18] Razvan Pascanu,et al. Stabilizing Transformers for Reinforcement Learning , 2019, ICML.
[19] Aaron C. Courville,et al. FiLM: Visual Reasoning with a General Conditioning Layer , 2017, AAAI.
[20] J. Schulman,et al. Reptile: a Scalable Metalearning Algorithm , 2018 .
[21] Marc'Aurelio Ranzato,et al. Mixture Models for Diverse Machine Translation: Tricks of the Trade , 2019, ICML.
[22] Ali Farhadi,et al. Defending Against Neural Fake News , 2019, NeurIPS.
[23] Sepp Hochreiter,et al. Learning to Learn Using Gradient Descent , 2001, ICANN.
[24] Jürgen Schmidhuber,et al. Learning to Control Fast-Weight Memories: An Alternative to Dynamic Recurrent Networks , 1992, Neural Computation.
[25] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[26] Pietro Perona,et al. One-shot learning of object categories , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[27] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[28] Regina Barzilay,et al. Few-shot Text Classification with Distributional Signatures , 2019, ICLR.
[29] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[30] 知秀 柴田. 5分で分かる!? 有名論文ナナメ読み:Jacob Devlin et al. : BERT : Pre-training of Deep Bidirectional Transformers for Language Understanding , 2020 .
[31] Brenden M. Lake,et al. Compositional generalization through meta sequence-to-sequence learning , 2019, NeurIPS.
[32] Geoffrey E. Hinton. Using fast weights to deblur old memories , 1987 .
[33] Oriol Vinyals,et al. Matching Networks for One Shot Learning , 2016, NIPS.