Text Editing by Command

A prevailing paradigm in neural text generation is one-shot generation, where text is produced in a single step. The one-shot setting is inadequate, however, when the constraints the user wishes to impose on the generated text are dynamic, especially when authoring longer documents. We address this limitation with an interactive text generation setting in which the user interacts with the system by issuing commands to edit existing text. To this end, we propose a novel text editing task, and introduce WikiDocEdits, a dataset of single-sentence edits crawled from Wikipedia. We show that our Interactive Editor, a transformer-based model trained on this dataset, outperforms baselines and obtains positive results in both automatic and human evaluations. We present empirical and qualitative analyses of this model's performance.

[1]  Michael Lebowitz,et al.  Story-telling as planning and learning , 1985 .

[2]  Xu Sun,et al.  A Skeleton-Based Model for Promoting Coherence Among Sentences in Narrative Story Generation , 2018, EMNLP.

[3]  Alexander M. Rush,et al.  Challenges in Data-to-Document Generation , 2017, EMNLP.

[4]  Diyi Yang,et al.  Automatically Neutralizing Subjective Bias in Text , 2019, AAAI.

[5]  Yutaka Matsuo,et al.  Variational Inference for Learning Representations of Natural Language Edits , 2021, AAAI.

[6]  Anirban Laha,et al.  Story Generation from Sequence of Independent Short Descriptions , 2017, ArXiv.

[7]  Jianfeng Gao,et al.  PlotMachines: Outline-Conditioned Generation with Dynamic Plot State Tracking , 2020, EMNLP.

[8]  Hang Li,et al.  Fact-based Text Editing , 2020, ACL.

[9]  Mark O. Riedl,et al.  Toward Automated Story Generation with Markov Chain Monte Carlo Methods and Deep Neural Networks , 2021, AIIDE Workshops.

[10]  Chris Callison-Burch,et al.  Unsupervised Hierarchical Story Infilling , 2019, Proceedings of the First Workshop on Narrative Understanding.

[11]  Ankush Gupta,et al.  A Deep Generative Framework for Paraphrase Generation , 2017, AAAI.

[12]  Ilya Sutskever,et al.  Language Models are Unsupervised Multitask Learners , 2019 .

[13]  Colin Raffel,et al.  Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..

[14]  Lysandre Debut,et al.  HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.

[15]  Mark Chen,et al.  Language Models are Few-Shot Learners , 2020, NeurIPS.

[16]  Graham Neubig,et al.  Learning to Represent Edits , 2018, ICLR.

[17]  Mirella Lapata,et al.  Text Generation from Knowledge Graphs with Graph Transformers , 2019, NAACL.

[18]  Danqi Chen,et al.  Knowledge Guided Text Retrieval and Reading for Open Domain Question Answering , 2019, ArXiv.

[19]  Mark O. Riedl,et al.  Controllable Neural Story Plot Generation via Reward Shaping , 2019, IJCAI.

[20]  Chris Quirk,et al.  Towards Content Transfer through Grounded Text Generation , 2019, NAACL.

[21]  Lav R. Varshney,et al.  CTRL: A Conditional Transformer Language Model for Controllable Generation , 2019, ArXiv.

[22]  Jason Yosinski,et al.  Plug and Play Language Models: A Simple Approach to Controlled Text Generation , 2020, ICLR.

[23]  Aaron Halfaker,et al.  Identifying Semantic Edit Intentions from Revisions in Wikipedia , 2017, EMNLP.

[24]  Zhifang Sui,et al.  Table-to-text Generation by Structure-aware Seq2seq Learning , 2017, AAAI.

[25]  Mark O. Riedl,et al.  Event Representations for Automated Story Generation with Deep Neural Nets , 2017, AAAI.

[26]  Zhe Gan,et al.  Pointer: Constrained Text Generation via Insertion-based Generative Pre-training , 2020, EMNLP.

[27]  Ming-Wei Chang,et al.  A Knowledge-Grounded Neural Conversation Model , 2017, AAAI.

[28]  Jason Weston,et al.  Reading Wikipedia to Answer Open-Domain Questions , 2017, ACL.

[29]  Dongyan Zhao,et al.  Plan-And-Write: Towards Better Automatic Storytelling , 2018, AAAI.

[30]  Jason Weston,et al.  Wizard of Wikipedia: Knowledge-Powered Conversational agents , 2018, ICLR.

[31]  Wiebke Wagner,et al.  Steven Bird, Ewan Klein and Edward Loper: Natural Language Processing with Python, Analyzing Text with the Natural Language Toolkit , 2010, Lang. Resour. Evaluation.

[32]  Kilian Q. Weinberger,et al.  BERTScore: Evaluating Text Generation with BERT , 2019, ICLR.

[33]  Tibor Kiss,et al.  Unsupervised Multilingual Sentence Boundary Detection , 2006, CL.

[34]  Salim Roukos,et al.  Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.

[35]  Eric P. Xing,et al.  Toward Controlled Generation of Text , 2017, ICML.

[36]  George R. Doddington,et al.  Automatic Evaluation of Machine Translation Quality Using N-gram Co-Occurrence Statistics , 2002 .

[37]  Nanyun Peng,et al.  Towards Controllable Story Generation , 2018 .

[38]  Percy Liang,et al.  Generating Sentences by Editing Prototypes , 2017, TACL.

[39]  Zhoujun Li,et al.  Response Generation by Context-aware Prototype Editing , 2018, AAAI.

[40]  Chang-Tien Lu,et al.  Modeling the Relationship between User Comments and Edits in Document Revision , 2019, EMNLP.

[41]  Ming-Wei Chang,et al.  REALM: Retrieval-Augmented Language Model Pre-Training , 2020, ICML.

[42]  Lukasz Kaiser,et al.  Generating Wikipedia by Summarizing Long Sequences , 2018, ICLR.

[43]  Qun Liu,et al.  Lexically Constrained Decoding for Sequence Generation Using Grid Beam Search , 2017, ACL.

[44]  P. Pirolli,et al.  The Sensemaking Process and Leverage Points for Analyst Technology as Identified Through Cognitive Task Analysis , 2015 .

[45]  Yann Dauphin,et al.  Hierarchical Neural Story Generation , 2018, ACL.

[46]  Manaal Faruqui,et al.  WikiAtomicEdits: A Multilingual Corpus of Wikipedia Edits for Modeling Language and Discourse , 2018, EMNLP.

[47]  Lihong Li,et al.  Neural Approaches to Conversational AI , 2019, Found. Trends Inf. Retr..

[48]  Ewan Klein,et al.  Natural Language Processing with Python , 2009 .