暂无分享,去创建一个
Yelong Shen | Dinghan Shen | Weizhu Chen | Yanru Qu | Mingzhi Zheng | Dinghan Shen | Weizhu Chen | Yelong Shen | Yanru Qu | Ming Zheng
[1] Avrim Blum,et al. The Bottleneck , 2021, Monopsony Capitalism.
[2] Graham Neubig,et al. SwitchOut: an Efficient Data Augmentation Algorithm for Neural Machine Translation , 2018, EMNLP.
[3] Larry S. Davis,et al. Adversarial Training for Free! , 2019, NeurIPS.
[4] Diyi Yang,et al. MixText: Linguistically-Informed Interpolation of Hidden Space for Semi-Supervised Text Classification , 2020, ACL.
[5] Jianfeng Gao,et al. Adversarial Training for Large Neural Language Models , 2020, ArXiv.
[6] Jianfeng Gao,et al. DeBERTa: Decoding-enhanced BERT with Disentangled Attention , 2020, ICLR.
[7] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[8] O. Chapelle,et al. Semi-Supervised Learning (Chapelle, O. et al., Eds.; 2006) [Book reviews] , 2009, IEEE Transactions on Neural Networks.
[9] Tie-Yan Liu,et al. Multi-branch Attentive Transformer , 2020, ArXiv.
[10] Jianfeng Gao,et al. UniLMv2: Pseudo-Masked Language Models for Unified Language Model Pre-Training , 2020, ICML.
[11] Sanjoy Dasgupta,et al. PAC Generalization Bounds for Co-training , 2001, NIPS.
[12] Joelle Pineau,et al. An Actor-Critic Algorithm for Sequence Prediction , 2016, ICLR.
[13] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[14] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[15] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[16] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[17] Mikhail Belkin,et al. A Co-Regularization Approach to Semi-supervised Learning with Multiple Views , 2005 .
[18] Balaji Lakshminarayanan,et al. AugMix: A Simple Data Processing Method to Improve Robustness and Uncertainty , 2020, ICLR.
[19] Quoc V. Le,et al. Semi-Supervised Sequence Modeling with Cross-View Training , 2018, EMNLP.
[20] Nipun Kwatra,et al. Wide-minima Density Hypothesis and the Explore-Exploit Learning Rate Schedule , 2020, ArXiv.
[21] Yu Cheng,et al. FreeLB: Enhanced Adversarial Training for Natural Language Understanding , 2020, ICLR.
[22] Quoc V. Le,et al. Unsupervised Data Augmentation for Consistency Training , 2019, NeurIPS.
[23] Bin Dong,et al. You Only Propagate Once: Accelerating Adversarial Training via Maximal Principle , 2019, NeurIPS.
[24] Philip Bachman,et al. Learning with Pseudo-Ensembles , 2014, NIPS.
[25] Bin Dong,et al. You Only Propagate Once: Painless Adversarial Training Using Maximal Principle , 2019 .
[26] Armen Aghajanyan,et al. Better Fine-Tuning by Reducing Representational Collapse , 2020, ICLR.
[27] Omer Levy,et al. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension , 2019, ACL.
[28] Quoc V. Le,et al. ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators , 2020, ICLR.
[29] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[30] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[31] Shafiq Joty,et al. Data Diversification: A Simple Strategy For Neural Machine Translation , 2020, NeurIPS.
[32] Quoc V. Le,et al. The Evolved Transformer , 2019, ICML.
[33] Jiawei Han,et al. Understanding the Difficulty of Training Transformers , 2020, EMNLP.
[34] Richard Socher,et al. Weighted Transformer Network for Machine Translation , 2017, ArXiv.
[35] Ashish Agarwal,et al. Hallucinations in Neural Machine Translation , 2018 .
[36] Xiaodong Liu,et al. SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization , 2020, ACL.
[37] Omer Levy,et al. SpanBERT: Improving Pre-training by Representing and Predicting Spans , 2019, TACL.
[38] Yang Song,et al. Improving the Robustness of Deep Neural Networks via Stability Training , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[39] Dilin Wang,et al. Improving Neural Language Modeling via Adversarial Training , 2019, ICML.
[40] Tie-Yan Liu,et al. Sequence Generation with Mixed Representations , 2020, ICML.
[41] Harini Kannan,et al. Adversarial Logit Pairing , 2018, NIPS 2018.
[42] Shin Ishii,et al. Virtual Adversarial Training: A Regularization Method for Supervised and Semi-Supervised Learning , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[43] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[44] Andrew M. Dai,et al. Adversarial Training Methods for Semi-Supervised Text Classification , 2016, ICLR.
[45] Yu Sun,et al. ERNIE: Enhanced Representation through Knowledge Integration , 2019, ArXiv.
[46] Dacheng Tao,et al. A Survey on Multi-view Learning , 2013, ArXiv.