暂无分享,去创建一个
Jakob N. Foerster | Tim Rocktaschel | Jakob Foerster | Danielle Rothermel | Margaret Li | Tim Rocktaschel | Dan Rothermel | Margaret Li
[1] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[2] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[3] Sanja Fidler,et al. Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[4] Danqi Chen,et al. of the Association for Computational Linguistics: , 2001 .
[5] Christopher Potts,et al. Learning Word Vectors for Sentiment Analysis , 2011, ACL.
[6] Jie Hou,et al. DeepSF: deep convolutional neural network for mapping protein sequences to folds , 2017, Bioinform..
[7] Nicola De Cao,et al. KILT: a Benchmark for Knowledge Intensive Language Tasks , 2020, NAACL.
[8] 知秀 柴田. 5分で分かる!? 有名論文ナナメ読み:Jacob Devlin et al. : BERT : Pre-training of Deep Bidirectional Transformers for Language Understanding , 2020 .
[9] Omer Levy,et al. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding , 2018, BlackboxNLP@EMNLP.
[10] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[11] Steven E. Brenner,et al. SCOPe: Structural Classification of Proteins—extended, integrating SCOP and ASTRAL data and classification of new structures , 2013, Nucleic Acids Res..
[12] Samuel R. Bowman,et al. ListOps: A Diagnostic Dataset for Latent Tree Learning , 2018, NAACL.
[13] Rémi Louf,et al. Transformers : State-ofthe-art Natural Language Processing , 2019 .
[14] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[15] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[16] Kenneth O. Stanley,et al. Differentiable plasticity: training plastic neural networks with backpropagation , 2018, ICML.
[17] Pieter Abbeel,et al. Pretrained Transformers as Universal Computation Engines , 2021, ArXiv.
[18] Omer Levy,et al. SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems , 2019, NeurIPS.
[19] Xiang Zhang,et al. Character-level Convolutional Networks for Text Classification , 2015, NIPS.
[20] Thorsten Brants,et al. One billion word benchmark for measuring progress in statistical language modeling , 2013, INTERSPEECH.
[21] Chris Dyer,et al. On the State of the Art of Evaluation in Neural Language Models , 2017, ICLR.
[22] John Canny,et al. Evaluating Protein Transfer Learning with TAPE , 2019, bioRxiv.
[23] Jaehoon Lee,et al. On Empirical Comparisons of Optimizers for Deep Learning , 2019, ArXiv.
[24] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[25] Sebastian Riedel,et al. Language Models as Knowledge Bases? , 2019, EMNLP.
[26] Liu Yang,et al. Long Range Arena: A Benchmark for Efficient Transformers , 2020, ICLR.