Language modelling for biological sequences – curated datasets and baselines
暂无分享,去创建一个
Henrik Nielsen | Alexander Rosenberg Johansen | Ole Winther | Jose Juan Almagro Armenteros | O. Winther | H. Nielsen | J. J. Almagro Armenteros | alexander rosenberg johansen
[1] The UniProt Consortium,et al. UniProt: a worldwide hub of protein knowledge , 2018, Nucleic Acids Res..
[2] James H. Martin,et al. Speech and Language Processing, 2nd Edition , 2008 .
[3] Yoshua Bengio,et al. Extracting and composing robust features with denoising autoencoders , 2008, ICML '08.
[4] Konstantinos D. Tsirigos,et al. SignalP 5.0 improves signal peptide predictions using deep neural networks , 2019, Nature Biotechnology.
[5] John Canny,et al. Evaluating Protein Transfer Learning with TAPE , 2019, bioRxiv.
[6] Johannes Söding,et al. MMseqs2: sensitive protein sequence searching for the analysis of massive data sets , 2017, bioRxiv.
[7] Yiming Yang,et al. Transformer-XL: Attentive Language Models beyond a Fixed-Length Context , 2019, ACL.
[8] C. Sander,et al. Database of homology‐derived protein structures and the structural meaning of sequence alignment , 1991, Proteins.
[9] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[10] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[11] Burkhard Rost,et al. Modeling the language of life – Deep Learning Protein Sequences , 2019, bioRxiv.
[12] Omer Levy,et al. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding , 2018, BlackboxNLP@EMNLP.
[13] George M. Church,et al. Unified rational protein engineering with sequence-based deep representation learning , 2019, Nature Methods.
[14] Luca Antiga,et al. Automatic differentiation in PyTorch , 2017 .
[15] Richard Socher,et al. Regularizing and Optimizing LSTM Language Models , 2017, ICLR.
[16] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[17] Yoshua Bengio,et al. A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..
[18] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[19] Steve Renals,et al. INTERSPEECH 2010 11th Annual Conference of the International Speech Communication Association , 2010 .
[20] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[21] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[22] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[23] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[24] Myle Ott,et al. Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences , 2019, Proceedings of the National Academy of Sciences.