IntelliCode compose: code generation using transformer
暂无分享,去创建一个
Neel Sundaresan | Alexey Svyatkovskiy | Shao Kun Deng | Shengyu Fu | Neel Sundaresan | Alexey Svyatkovskiy | Shengyu Fu
[1] Andrew M. Dai,et al. Gmail Smart Compose: Real-Time Assisted Writing , 2019, KDD.
[2] Hakan Inan,et al. Tying Word Vectors and Word Classifiers: A Loss Framework for Language Modeling , 2016, ICLR.
[3] Nathanael Chambers,et al. A Corpus and Cloze Evaluation for Deeper Understanding of Commonsense Stories , 2016, NAACL.
[4] Yejin Choi,et al. SWAG: A Large-Scale Adversarial Dataset for Grounded Commonsense Inference , 2018, EMNLP.
[5] Song Wang,et al. Automatically Learning Semantic Features for Defect Prediction , 2016, 2016 IEEE/ACM 38th International Conference on Software Engineering (ICSE).
[6] Mira Mezini,et al. Learning from examples to improve code completion systems , 2009, ESEC/SIGSOFT FSE.
[7] Mira Mezini,et al. Intelligent Code Completion with Bayesian Networks , 2015, ACM Trans. Softw. Eng. Methodol..
[8] Alexander Sergeev,et al. Horovod: fast and easy distributed deep learning in TensorFlow , 2018, ArXiv.
[9] Anh Tuan Nguyen,et al. A statistical semantic language model for source code , 2013, ESEC/FSE 2013.
[10] Yi Zhang,et al. Automatic parameter recommendation for practical API usage , 2012, 2012 34th International Conference on Software Engineering (ICSE).
[11] Lysandre Debut,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[12] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[13] Thomas Wolf,et al. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter , 2019, ArXiv.
[14] Colin Raffel,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[15] Chin-Yew Lin,et al. Automatic Evaluation of Machine Translation Quality Using Longest Common Subsequence and Skip-Bigram Statistics , 2004, ACL.
[16] Alessandro Orso,et al. Automated API-usage update for Android apps , 2019, ISSTA.
[17] Yue Luo,et al. Nomen est Omen: Exploring and Exploiting Similarities between Argument and Parameter Names , 2016, 2016 IEEE/ACM 38th International Conference on Software Engineering (ICSE).
[18] R'emi Louf,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[19] Eran Yahav,et al. Code completion with statistical language models , 2014, PLDI.
[20] Neel Sundaresan,et al. Pythia: AI-assisted Code Completion System , 2019, KDD.
[21] Oleksandr Polozov,et al. Program Synthesis and Semantic Parsing with Learned Code Idioms , 2019, NeurIPS.
[22] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[23] Lav R. Varshney,et al. CTRL: A Conditional Transformer Language Model for Controllable Generation , 2019, ArXiv.
[24] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[25] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[26] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[27] Chanchal Kumar Roy,et al. CSCC: Simple, Efficient, Context Sensitive Code Completion , 2014, 2014 IEEE International Conference on Software Maintenance and Evolution.
[28] Premkumar T. Devanbu,et al. On the naturalness of software , 2016, Commun. ACM.
[29] Omer Levy,et al. code2seq: Generating Sequences from Structured Representations of Code , 2018, ICLR.
[30] Xiaodong Gu,et al. Deep Code Search , 2018, 2018 IEEE/ACM 40th International Conference on Software Engineering (ICSE).
[31] Premkumar T. Devanbu,et al. A Survey of Machine Learning for Big Code and Naturalness , 2017, ACM Comput. Surv..
[32] Jimmy J. Lin,et al. Distilling Task-Specific Knowledge from BERT into Simple Neural Networks , 2019, ArXiv.
[33] Chin-Yew Lin,et al. ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.
[34] Mohammad Shoeybi,et al. Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism , 2019, ArXiv.
[35] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[36] Maik Riechert,et al. Fast and Memory-Efficient Neural Code Completion , 2020, ArXiv.
[37] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[38] Quoc V. Le,et al. The Evolved Transformer , 2019, ICML.