QUANT - Question Answering Benchmark Curator

Question answering engines have become one of the most popular type of applications driven by Semantic Web technologies. Consequently, the provision of means to quantify the performance of current question answering approaches on current datasets has become ever more important. However, a large percentage of the queries found in popular question answering benchmarks cannot be executed on current versions of their reference dataset. There is a consequently a clear need to curate question answering benchmarks periodically. However, the manual alteration of question answering benchmarks is often error-prone. We alleviate this problem by presenting QUANT, a novel framework for the creation and curation of question answering benchmarks. QUANT supports the curation of benchmarks by generating smart edit suggestions for question-query pair and for the corresponding metadata. In addition, our framework supports the creation of new benchmark entries by providing predefined quality checks for queries. We evaluate QUANT on 653 questions obtained from QALD-1 to QALD-8 with 10 users. Our results show that our framework generates reliable suggestions and can reduce the curation effort for QA benchmarks by up to 91%.

[1]  Octavian Udrea,et al.  Apples and oranges: a comparison of RDF benchmarks and real RDF datasets , 2011, SIGMOD '11.

[2]  Jens Lehmann,et al.  Survey on challenges of Question Answering in the Semantic Web , 2017, Semantic Web.

[3]  Felix Conrads,et al.  Benchmarking question answering systems , 2019, Semantic Web.

[4]  Ellen M. Voorhees,et al.  The TREC-8 Question Answering Track Report , 1999, TREC.

[5]  Markus Krötzsch,et al.  Getting the Most Out of Wikidata: Semantic Technology Usage in Wikipedia's Knowledge Graph , 2018, SEMWEB.

[6]  Shweta,et al.  Evaluation of Inter-Rater Agreement and Inter-Rater Reliability for Observational Data: An Overview of Concepts and Methods , 2015 .

[7]  Jens Lehmann,et al.  LC-QuAD: A Corpus for Complex Question Answering over Knowledge Graphs , 2017, SEMWEB.

[8]  Axel-Cyrille Ngonga Ngomo,et al.  Ensemble Learning of Named Entity Recognition Algorithms using Multilayer Perceptron for the Multilingual Web of Data , 2017, K-CAP.

[9]  Axel-Cyrille Ngonga Ngomo,et al.  LANCE: A Generic Benchmark Generator for Linked Data , 2015, International Semantic Web Conference.

[10]  Giorgio Maria Di Nunzio,et al.  10 Years of CLEF Data in DIRECT: Where We Are and Where We Can Go , 2010, EVIA@NTCIR.

[11]  Elena Cabrio,et al.  6th Open Challenge on Question Answering over Linked Data (QALD-6) , 2016, SemWebEval@ESWC.

[12]  Axel-Cyrille Ngonga Ngomo,et al.  7th Open Challenge on Question Answering over Linked Data (QALD-7) , 2017, SemWebEval@ESWC.

[13]  Jens Lehmann,et al.  Overcoming Challenges of Semantic Question Answering in the Semantic Web , 2015 .

[14]  Vanessa López,et al.  Core techniques of question answering systems over knowledge bases: a survey , 2017, Knowledge and Information Systems.

[15]  Ming-Wei Chang,et al.  The Value of Semantic Parse Labeling for Knowledge Base Question Answering , 2016, ACL.

[16]  Axel-Cyrille Ngonga Ngomo,et al.  All that Glitters Is Not Gold - Rule-Based Curation of Reference Datasets for Named Entity Recognition and Entity Linking , 2017, ESWC.

[17]  Alexander Yates,et al.  Large-scale Semantic Parsing via Schema Matching and Lexicon Extension , 2013, ACL.

[18]  Andrew Chou,et al.  Semantic Parsing on Freebase from Question-Answer Pairs , 2013, EMNLP.