LEWIS: Levenshtein Editing for Unsupervised Text Style Transfer
暂无分享,去创建一个
[1] Roy Ka-Wei Lee,et al. Text Style Transfer: A Review and Experiment Evaluation , 2020, ArXiv.
[2] Mohit Iyyer,et al. Reformulating Unsupervised Style Transfer as Paraphrase Generation , 2020, EMNLP.
[3] Yiming Yang,et al. Politeness Transfer: A Tag and Generate Approach , 2020, ACL.
[4] Graham Neubig,et al. A Probabilistic Formulation of Unsupervised Text Style Transfer , 2020, ICLR.
[5] Omer Levy,et al. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension , 2019, ACL.
[6] Sebastian Gehrmann,et al. exBERT: A Visual Analysis Tool to Explore Learned Representations in Transformers Models , 2019, ArXiv.
[7] Lysandre Debut,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[8] Marc'Aurelio Ranzato,et al. Revisiting Self-Training for Neural Sequence Generation , 2019, ICLR.
[9] Kilian Q. Weinberger,et al. BERTScore: Evaluating Text Generation with BERT , 2019, ICLR.
[10] Eric Malmi. Unsupervised Text Style Transfer with Padded Masked Language Models , 2020 .
[11] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[12] Aliaksei Severyn,et al. Encode, Tag, Realize: High-Precision Text Editing , 2019, EMNLP.
[13] Tao Zhang,et al. Mask and Infill: Applying Masked Language Model for Sentiment Transfer , 2019, IJCAI.
[14] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[15] Myle Ott,et al. fairseq: A Fast, Extensible Toolkit for Sequence Modeling , 2019, NAACL.
[16] Guillaume Lample,et al. Multiple-Attribute Text Rewriting , 2018, ICLR.
[17] Lili Mou,et al. Disentangled Representation Learning for Non-Parallel Text Style Transfer , 2018, ACL.
[18] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[19] Myle Ott,et al. Understanding Back-Translation at Scale , 2018, EMNLP.
[20] Omer Levy,et al. Ultra-Fine Entity Typing , 2018, ACL.
[21] Mirella Lapata,et al. Coarse-to-Fine Decoding for Neural Semantic Parsing , 2018, ACL.
[22] Yulia Tsvetkov,et al. Style Transfer Through Back-Translation , 2018, ACL.
[23] Matt Post,et al. A Call for Clarity in Reporting BLEU Scores , 2018, WMT.
[24] Percy Liang,et al. Delete, Retrieve, Generate: a Simple Approach to Sentiment and Style Transfer , 2018, NAACL.
[25] Dongyan Zhao,et al. Style Transfer in Text: Exploration and Evaluation , 2017, AAAI.
[26] Hao Wu,et al. Mixed Precision Training , 2017, ICLR.
[27] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[28] Regina Barzilay,et al. Style Transfer from Non-Parallel Text by Cross-Alignment , 2017, NIPS.
[29] Julian J. McAuley,et al. Ups and Downs: Modeling the Visual Evolution of Fashion Trends with One-Class Collaborative Filtering , 2016, WWW.
[30] Rico Sennrich,et al. Improving Neural Machine Translation Models with Monolingual Data , 2015, ACL.
[31] Eugene Charniak,et al. Coarse-to-Fine n-Best Parsing and MaxEnt Discriminative Reranking , 2005, ACL.
[32] Philipp Koehn,et al. Statistical Significance Tests for Machine Translation Evaluation , 2004, EMNLP.
[33] Salim Roukos,et al. Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.
[34] Vladimir I. Levenshtein,et al. Binary codes capable of correcting deletions, insertions, and reversals , 1965 .