Injecting Entity Types into Entity-Guided Text Generation

Recent successes in deep generative modeling have led to significant advances in natural language generation (NLG). Incorporating entities into neural generation models has demonstrated great improvements by assisting to infer the summary topic and to generate coherent content. In order to enhance the role of entity in NLG, in this paper, we aim to model the entity type in the decoding phase to generate contextual words accurately. We develop a novel NLG model to produce a target sequence (i.e., a news article) based on a given list of entities. The generation quality depends significantly on whether the input entities are logically connected and expressed in the output. Our model has a multi-step decoder that injects the entity types into the process of entity mention generation. It first predicts the token of being a contextual word or an entity, then if an entity, predicts the entity mention. It effectively embeds the entity's meaning into hidden states, making the generated words precise. Experiments on two public datasets demonstrate type injection performs better than type embedding concatenation baselines.

[1]  Arun Kumar Sangaiah,et al.  Automatic Generation of News Comments Based on Gated Attention Neural Networks , 2018, IEEE Access.

[2]  Xiaocheng Feng,et al.  Topic-to-Essay Generation with Neural Networks , 2018, IJCAI.

[3]  Xiaodong Liu,et al.  Unified Language Model Pre-training for Natural Language Understanding and Generation , 2019, NeurIPS.

[4]  Zhe Gan,et al.  Pointer: Constrained Text Generation via Insertion-based Generative Pre-training , 2020, EMNLP.

[5]  Yann Dauphin,et al.  Hierarchical Neural Story Generation , 2018, ACL.

[6]  Zhiting Hu,et al.  A Survey of Knowledge-Enhanced Text Generation , 2020, ArXiv.

[7]  Xing Xie,et al.  Personalized Reason Generation for Explainable Song Recommendation , 2019, ACM Trans. Intell. Syst. Technol..

[8]  Ke Wang,et al.  SentiGAN: Generating Sentimental Texts via Mixture Adversarial Networks , 2018, IJCAI.

[9]  Yaser Al-Onaizan,et al.  Training Neural Machine Translation to Apply Terminology Constraints , 2019, ACL.

[10]  Christopher D. Manning,et al.  Incorporating Non-local Information into Information Extraction Systems by Gibbs Sampling , 2005, ACL.

[11]  Eric P. Xing,et al.  Toward Controlled Generation of Text , 2017, ICML.

[12]  Seung-won Hwang,et al.  Entity Commonsense Representation for Neural Abstractive Summarization , 2018, NAACL.

[13]  Ani Nenkova,et al.  Entity-driven Rewrite for Multi-document Summarization , 2008, IJCNLP.

[14]  Zhe Hu,et al.  An Entity-Driven Framework for Abstractive Summarization , 2019, EMNLP.

[15]  Lei Li,et al.  CGMH: Constrained Sentence Generation by Metropolis-Hastings Sampling , 2018, AAAI.

[16]  Ilya Sutskever,et al.  Language Models are Unsupervised Multitask Learners , 2019 .

[17]  Scott Weinstein,et al.  Centering: A Framework for Modeling the Local Coherence of Discourse , 1995, CL.

[18]  悠太 菊池,et al.  大規模要約資源としてのNew York Times Annotated Corpus , 2015 .

[19]  Dongyan Zhao,et al.  Stick to the Facts: Learning towards a Fidelity-oriented E-Commerce Product Description Generation , 2019, EMNLP.

[20]  Yoshua Bengio,et al.  Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.

[21]  Heng Ji,et al.  Entity-aware Image Caption Generation , 2018, EMNLP.

[22]  Lei Li,et al.  Enhancing Topic-to-Essay Generation with External Commonsense Knowledge , 2019, ACL.

[23]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[24]  Minlie Huang,et al.  A Knowledge-Enhanced Pretraining Model for Commonsense Story Generation , 2020, TACL.

[25]  Tao Yu,et al.  TypeSQL: Knowledge-Based Type-Aware Neural Text-to-SQL Generation , 2018, NAACL.

[26]  Mirella Lapata,et al.  Data-to-text Generation with Entity Modeling , 2019, ACL.

[27]  Touseef Iqbal,et al.  The survey: Text generation models in deep learning , 2020, J. King Saud Univ. Comput. Inf. Sci..

[28]  Percy Liang,et al.  Delete, Retrieve, Generate: a Simple Approach to Sentiment and Style Transfer , 2018, NAACL.

[29]  Yejin Choi,et al.  Dynamic Entity Representations in Neural Language Models , 2017, EMNLP.

[30]  Regina Barzilay,et al.  Blank Language Models , 2020, EMNLP.

[31]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[32]  Tapio Salakoski,et al.  Template-free Data-to-Text Generation of Finnish Sports News , 2019, NODALIDA.

[33]  Qiaozhu Mei,et al.  Neural Language Generation: Formulation, Methods, and Evaluation , 2020, ArXiv.

[34]  Hang Li,et al.  “ Tony ” DNN Embedding for “ Tony ” Selective Read for “ Tony ” ( a ) Attention-based Encoder-Decoder ( RNNSearch ) ( c ) State Update s 4 SourceVocabulary Softmax Prob , 2016 .

[35]  Quoc V. Le,et al.  Sequence to Sequence Learning with Neural Networks , 2014, NIPS.

[36]  Yoshua Bengio,et al.  Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.

[37]  Dongyan Zhao,et al.  Plan-And-Write: Towards Better Automatic Storytelling , 2018, AAAI.