Multi-Document Answer Generation for Non-Factoid Questions
暂无分享,去创建一个
The current research will be devoted to the challenging and under-investigated task of multi-source answer generation for complex non-factoid questions. We will start with experimenting with generative models on one particular type of non-factoid questions - instrumental/procedural questions which often start with "how-to". For this, a new dataset, comprised of more than 100,000 QA-pairs which were crawled from a dedicated web-resource where each answer has a set of references to the articles it was written upon, will be used. We will also compare different ways of model evaluation to choose a metric which better correlates with human assessment. To be able to do this, the way people evaluate answers to non-factoid questions and set some formal criteria of what makes a good quality answer is needed to be understood. Eye-tracking and crowdsourcing methods will be employed to study how users interact with answers and evaluate them, and how the answer features correlate with task complexity. We hope that our research will help to redefine the way users interact and work with search engines so as to transform IR finally into the answer retrieval systems that users have always desired.
[1] Chin-Yew Lin,et al. ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.
[2] Sameer Singh,et al. Evaluating Question Answering Evaluation , 2019, EMNLP.
[3] Jason Weston,et al. ELI5: Long Form Question Answering , 2019, ACL.
[4] Andrew A. Tawfik,et al. Role of questions in inquiry-based instruction: towards a design taxonomy for question-asking and implications for design , 2020 .