Paraphrase Generation with Collaboration between the Forward and the Backward Decoder

Automatic paraphrase generation plays a key role in many natural language applications. The dominant paraphrase generation models are the encoder-decoder neural networks with attention, where the decoder uses the information of the source text while predicting target text. However, the outputs of these paraphrase models often suffer the semantic error problem. This problem is caused by the inadequate information of the decoder. In this work, we introduce a novel neural model to solve this problem, called Collaboration between the Forward and the Backward Decoder. Specifically, the hidden states of the backward decoder are used as supplementary information of the forward decoder. Therefore, the forward decoder can generate more reasonable paraphrase text using the target-side future contextual. Conversely, the backward decoder employs the hidden states of the forward decoder to prevent the semantic error problem. As two experimental examples show, the proposed model can generate the high-quality paraphrase through this collaboration mechanism. The empirical study on two benchmark datasets demonstrates that our model outperforms some baselines and achieves the state-of-the-art performance.

[1]  Oladimeji Farri,et al.  Neural Paraphrase Generation with Stacked Residual LSTM Networks , 2016, COLING.

[2]  Salim Roukos,et al.  Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.

[3]  Emiel Krahmer,et al.  Sentence Simplification by Monolingual Machine Translation , 2012, ACL.

[4]  Chris Quirk,et al.  Monolingual Machine Translation for Paraphrase Generation , 2004, EMNLP.

[5]  Yoshua Bengio,et al.  Professor Forcing: A New Algorithm for Training Recurrent Networks , 2016, NIPS.

[6]  Yoshua Bengio,et al.  Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.

[7]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[8]  Ming Zhou,et al.  Combining Multiple Resources to Improve SMT-based Paraphrasing Model , 2008, ACL.

[9]  Wenjie Li,et al.  Joint Copying and Restricted Generation for Paraphrase , 2016, AAAI.

[10]  Regina Barzilay,et al.  Learning to Paraphrase: An Unsupervised Approach Using Multiple-Sequence Alignment , 2003, NAACL.

[11]  Marc'Aurelio Ranzato,et al.  Sequence Level Training with Recurrent Neural Networks , 2015, ICLR.

[12]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[13]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[14]  Kathleen McKeown,et al.  Paraphrasing Questions Using Given and new information , 1983, CL.

[15]  Xu Sun,et al.  A Hierarchical End-to-End Model for Jointly Improving Text Summarization and Sentiment Classification , 2018, IJCAI.

[16]  Eduard H. Hovy,et al.  Automatic Evaluation of Summaries Using N-gram Co-occurrence Statistics , 2003, NAACL.

[17]  Rongrong Ji,et al.  Asynchronous Bidirectional Decoding for Neural Machine Translation , 2018, AAAI.

[18]  Iryna Gurevych,et al.  A Monolingual Tree-based Translation Model for Sentence Simplification , 2010, COLING.

[19]  Gholamreza Haffari,et al.  Towards Decoding as Continuous Optimisation in Neural Machine Translation , 2017, EMNLP.

[20]  Chengqing Zong,et al.  Approach to Spoken Chinese Paraphrasing Based on Feature Extraction , 2001, NLPRS.

[21]  Shashi Narayan,et al.  Hybrid Simplification using Deep Semantics and Machine Translation , 2014, ACL.

[22]  Ankush Gupta,et al.  A Deep Generative Framework for Paraphrase Generation , 2017, AAAI.

[23]  Sergiu Nisioi,et al.  Exploring Neural Text Simplification Models , 2017, ACL.

[24]  Kang Yang,et al.  Bidirectional Attentional Encoder-Decoder Model and Bidirectional Beam Search for Abstractive Summarization , 2018, ArXiv.

[25]  Taro Watanabe,et al.  Paraphrasing as Machine Translation (自然言語処理特集号「言い換え」) , 2004 .

[26]  Joelle Pineau,et al.  How NOT To Evaluate Your Dialogue System: An Empirical Study of Unsupervised Evaluation Metrics for Dialogue Response Generation , 2016, EMNLP.

[27]  Xu Sun,et al.  Query and Output: Generating Words by Querying Distributed Word Representations for Paraphrase Generation , 2018, NAACL.

[28]  Lemao Liu,et al.  Agreement on Target-bidirectional Neural Machine Translation , 2016, NAACL.

[29]  Mirella Lapata,et al.  Sentence Simplification with Deep Reinforcement Learning , 2017, EMNLP.

[30]  Hang Li,et al.  Paraphrase Generation with Deep Reinforcement Learning , 2017, EMNLP.

[31]  Alexander M. Rush,et al.  Sequence-to-Sequence Learning as Beam-Search Optimization , 2016, EMNLP.

[32]  Satoshi Sato,et al.  Verb Paraphrase based on Case Frame Alignment , 2002, ACL.