暂无分享,去创建一个
[1] Premkumar T. Devanbu,et al. On the naturalness of software , 2016, Commun. ACM.
[2] Truyen Tran,et al. A deep language model for software code , 2016, FSE 2016.
[3] Christopher C. Cummins,et al. Synthesizing benchmarks for predictive modeling , 2017, 2017 IEEE/ACM International Symposium on Code Generation and Optimization (CGO).
[4] Xiaodong Gu,et al. Deep API learning , 2016, SIGSOFT FSE.
[5] Kevin Gimpel,et al. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations , 2019, ICLR.
[6] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[7] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[8] José Nelson Amaral,et al. Syntax and sensibility: Using language models to detect and correct syntax errors , 2018, 2018 IEEE 25th International Conference on Software Analysis, Evolution and Reengineering (SANER).
[9] Ali Mesbah,et al. DeepDelta: learning to repair compilation errors , 2019, ESEC/SIGSOFT FSE.
[10] Premkumar T. Devanbu,et al. A Survey of Machine Learning for Big Code and Naturalness , 2017, ACM Comput. Surv..
[11] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[12] Noah Constant,et al. Character-Level Language Modeling with Deeper Self-Attention , 2018, AAAI.
[13] Miltiadis Allamanis,et al. The adverse effects of code duplication in machine learning models of code , 2018, Onward!.
[14] Earl T. Barr,et al. Learning Python Code Suggestion with a Sparse Pointer Network , 2016, ArXiv.
[15] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[16] Sumit Gulwani,et al. Compilation Error Repair: For the Student Programs, From the Student Programs , 2018, 2018 IEEE/ACM 40th International Conference on Software Engineering: Software Engineering Education and Training (ICSE-SEET).
[17] Rafael-Michael Karampatsis,et al. Maybe Deep Neural Networks are the Best Choice for Modeling Source Code , 2019, ArXiv.
[18] Yiming Yang,et al. Transformer-XL: Attentive Language Models beyond a Fixed-Length Context , 2019, ACL.
[19] Emily Hill,et al. Towards automatically generating summary comments for Java methods , 2010, ASE.
[20] J. Wrachtrup,et al. Proposal for a room-temperature diamond maser , 2015, Nature communications.
[21] Eran Yahav,et al. Code completion with statistical language models , 2014, PLDI.
[22] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[23] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[24] Rishabh Singh,et al. Automated Correction for Syntax Errors in Programming Assignments using Recurrent Neural Networks , 2016, ArXiv.
[25] Hailong Sun,et al. A Novel Neural Source Code Representation Based on Abstract Syntax Tree , 2019, 2019 IEEE/ACM 41st International Conference on Software Engineering (ICSE).
[26] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[27] Martin White,et al. Toward Deep Learning Software Repositories , 2015, 2015 IEEE/ACM 12th Working Conference on Mining Software Repositories.
[28] Charles A. Sutton,et al. Learning natural coding conventions , 2014, SIGSOFT FSE.
[29] Rahul Gupta,et al. DeepFix: Fixing Common C Language Errors by Deep Learning , 2017, AAAI.
[30] Charles A. Sutton,et al. Mining source code repositories at massive scale using language modeling , 2013, 2013 10th Working Conference on Mining Software Repositories (MSR).
[31] Kevin Gimpel,et al. Gaussian Error Linear Units (GELUs) , 2016 .
[32] Mohammad Shoeybi,et al. Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism , 2019, ArXiv.
[33] Premkumar T. Devanbu,et al. On the localness of software , 2014, SIGSOFT FSE.