Syn-QG: Syntactic and Shallow Semantic Rules for Question Generation

Question Generation (QG) is fundamentally a simple syntactic transformation; however, many aspects of semantics influence what questions are good to form. We implement this observation by developing Syn-QG, a set of transparent syntactic rules leveraging universal dependencies, shallow semantic parsing, lexical resources, and custom rules which transform declarative sentences into question-answer pairs. We utilize PropBank argument descriptions and VerbNet state predicates to incorporate shallow semantic content, which helps generate questions of a descriptive nature and produce inferential and semantically richer questions than existing systems. In order to improve syntactic fluency and eliminate grammatically incorrect questions, we employ back-translation over the output of these syntactic rules. A set of crowd-sourced evaluations shows that our system can generate a larger number of highly grammatical and relevant questions than previous QG systems and that back-translation drastically improves grammaticality at a slight cost of generating irrelevant questions.

[1]  Paul Piwek,et al.  The First Question Generation Shared Task Evaluation Challenge , 2010, Dialogue Discourse.

[2]  Martha Palmer,et al.  Verbnet: a broad-coverage, comprehensive verb lexicon , 2005 .

[3]  Mohammed J. Zaki,et al.  Natural Question Generation with Reinforcement Learning Based Graph-to-Sequence Model , 2019, ArXiv.

[4]  Furu Wei,et al.  MiniLM: Deep Self-Attention Distillation for Task-Agnostic Compression of Pre-Trained Transformers , 2020, NeurIPS.

[5]  Noah A. Smith,et al.  Good Question! Statistical Ranking for Question Generation , 2010, NAACL.

[6]  Jian Zhang,et al.  SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.

[7]  Roger Levy,et al.  Tregex and Tsurgeon: tools for querying and manipulating tree data structures , 2006, LREC.

[8]  Jiusheng Chen,et al.  ProphetNet: Predicting Future N-gram for Sequence-to-Sequence Pre-training , 2020, EMNLP.

[9]  Lidong Bing,et al.  Improving Question Generation With to the Point Context , 2019, EMNLP.

[10]  James Pustejovsky,et al.  VerbNet Representations: Subevent Semantics for Transfer Verbs , 2019, Proceedings of the First International Workshop on Designing Meaning Representations.

[11]  Sebastian Riedel,et al.  Evaluating Rewards for Question Generation Models , 2019, NAACL.

[12]  Xiaodong Liu,et al.  Unified Language Model Pre-training for Natural Language Understanding and Generation , 2019, NeurIPS.

[13]  Bijan Parsia,et al.  A Systematic Review of Automatic Question Generation for Educational Purposes , 2019, International Journal of Artificial Intelligence in Education.

[14]  Christiane Fellbaum,et al.  Book Reviews: WordNet: An Electronic Lexical Database , 1999, CL.

[15]  Tao Qin,et al.  Question Answering and Question Generation as Dual Tasks , 2017, ArXiv.

[16]  Daniel Jurafsky,et al.  Noising and Denoising Natural Language: Diverse Backtranslation for Grammar Correction , 2018, NAACL.

[17]  Igor Labutov,et al.  Deep Questions without Deep Understanding , 2015, ACL.

[18]  Ming Zhou,et al.  Neural Question Generation from Text: A Preliminary Study , 2017, NLPCC.

[19]  Mohit Bansal,et al.  Addressing Semantic Drift in Question Generation for Semi-Supervised Question Answering , 2019, EMNLP.

[20]  Balaraman Ravindran,et al.  Let’s Ask Again: Refine Network for Automatic Question Generation , 2019, EMNLP.

[21]  Salim Roukos,et al.  Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.

[22]  Xinya Du,et al.  Learning to Ask: Neural Question Generation for Reading Comprehension , 2017, ACL.

[23]  Slav Petrov,et al.  Syntactic Annotations for the Google Books NGram Corpus , 2012, ACL.

[24]  Myle Ott,et al.  fairseq: A Fast, Extensible Toolkit for Sequence Modeling , 2019, NAACL.

[25]  Jack Mostow,et al.  Generating Instruction Automatically for the Reading Strategy of Self-Questioning , 2009, AIED.

[26]  Joakim Nivre,et al.  Universal Stanford dependencies: A cross-linguistic typology , 2014, LREC.

[27]  Philip Bachman,et al.  Machine Comprehension by Text-to-Text Neural Question Generation , 2017, Rep4NLP@ACL.

[28]  Kyomin Jung,et al.  Improving Neural Question Generation using Answer Separation , 2018, AAAI.

[29]  Tong Wang,et al.  A Joint Model for Question Answering and Question Generation , 2017, ArXiv.

[30]  Hao Tian,et al.  ERNIE-GEN: An Enhanced Multi-Flow Pre-training and Fine-tuning Framework for Natural Language Generation , 2020, ArXiv.

[31]  Michael Flor,et al.  A Semantic Role-based Approach to Open-Domain Automatic Question Generation , 2018, BEA@NAACL-HLT.

[32]  Zhiguo Wang,et al.  A Unified Query-based Generative Model for Question Generation and Question Answering , 2017, ArXiv.

[33]  Luke S. Zettlemoyer,et al.  AllenNLP: A Deep Semantic Natural Language Processing Platform , 2018, ArXiv.

[34]  Nan Yang,et al.  Sequential Copying Networks , 2018, AAAI.

[35]  Xuchen Yao,et al.  Question Generation with Minimal Recursion Semantics , 2010 .

[36]  Yao Zhao,et al.  Paragraph-level Neural Question Generation with Maxout Pointer and Gated Self-attention Networks , 2018, EMNLP.

[37]  Lauri Karttunen,et al.  Simple and Phrasal Implicatives , 2012, *SEMEVAL.

[38]  Paul Tarau,et al.  Infusing NLU into Automatic Question Generation , 2016, INLG.

[39]  Jianfeng Gao,et al.  A Human Generated MAchine Reading COmprehension Dataset , 2018 .

[40]  Jianfeng Gao,et al.  UniLMv2: Pseudo-Masked Language Models for Unified Language Model Pre-Training , 2020, ICML.

[41]  Noah A. Smith,et al.  Question Generation via Overgenerating Transformations and Ranking , 2009 .

[42]  Di Niu,et al.  Asking Questions the Human Way: Scalable Question-Answer Generation from Text Corpus , 2020, WWW.

[43]  John C. Nesbit,et al.  Generating Natural Language Questions to Support Learning On-Line , 2013, ENLG.

[44]  Rodney D. Nielsen,et al.  Leveraging Multiple Views of Text for Automatic Question Generation , 2015, AIED.

[45]  Rodney D. Nielsen,et al.  Linguistic Considerations in Automatic Question Generation , 2014, ACL.

[46]  Regina Barzilay,et al.  Capturing Greater Context for Question Generation , 2019, AAAI.

[47]  Yanjun Ma,et al.  Answer-focused and Position-aware Neural Question Generation , 2018, EMNLP.

[48]  Yu Xu,et al.  Learning to Generate Questions by LearningWhat not to Generate , 2019, WWW.