CaRB: A Crowdsourced Benchmark for Open IE

Open Information Extraction (Open IE) systems have been traditionally evaluated via manual annotation. Recently, an automated evaluator with a benchmark dataset (OIE2016) was released – it scores Open IE systems automatically by matching system predictions with predictions in the benchmark dataset. Unfortunately, our analysis reveals that its data is rather noisy, and the tuple matching in the evaluator has issues, making the results of automated comparisons less trustworthy. We contribute CaRB, an improved dataset and framework for testing Open IE systems. To the best of our knowledge, CaRB is the first crowdsourced Open IE dataset and it also makes substantive changes in the matching code and metrics. NLP experts annotate CaRB’s dataset to be more accurate than OIE2016. Moreover, we find that on one pair of Open IE systems, CaRB framework provides contradictory results to OIE2016. Human assessment verifies that CaRB’s ranking of the two systems is the accurate ranking. We release the CaRB framework along with its crowdsourced dataset.

[1]  Ido Dagan,et al.  Creating a Large Benchmark for Open Information Extraction , 2016, EMNLP.

[2]  Harinder Pal,et al.  Demonyms and Compound Relational Nouns in Nominal Open IE , 2016, AKBC@NAACL-HLT.

[3]  Oren Etzioni,et al.  Open Information Extraction: The Second Generation , 2011, IJCAI.

[4]  Mausam,et al.  Open Information Extraction Systems and Downstream Applications , 2016, IJCAI.

[5]  Harinder Pal,et al.  Bootstrapping for Numerical Open IE , 2017, ACL.

[6]  Ido Dagan,et al.  Getting More Out Of Syntax with PropS , 2016, ArXiv.

[7]  Oren Etzioni,et al.  An analysis of open information extraction based on semantic role labeling , 2011, K-CAP '11.

[8]  Luke S. Zettlemoyer,et al.  Question-Answer Driven Semantic Role Labeling: Using Natural Language to Annotate Natural Language , 2015, EMNLP.

[9]  Philippe Langlais,et al.  WiRe57 : A Fine-Grained Benchmark for Open Information Extraction , 2018, LAW@ACL.

[10]  Ming Zhou,et al.  Neural Open Information Extraction , 2018, ACL.

[11]  Alexander Löser,et al.  Analysing Errors of Open Information Extraction Systems , 2017, ArXiv.

[12]  Oren Etzioni,et al.  Open Language Learning for Information Extraction , 2012, EMNLP.

[13]  Mausam,et al.  Open Information Extraction from Conjunctive Sentences , 2018, COLING.

[14]  Ido Dagan,et al.  Open IE as an Intermediate Structure for Semantic Tasks , 2015, ACL.

[15]  Luciano Del Corro,et al.  ClausIE: clause-based open information extraction , 2013, WWW.

[16]  Oren Etzioni,et al.  Identifying Relations for Open Information Extraction , 2011, EMNLP.

[17]  Oren Etzioni,et al.  Open Information Extraction from the Web , 2007, CACM.

[18]  Xiaojun Chen,et al.  Supervised Neural Models Revitalize the Open Relation Extraction , 2019, ArXiv.

[19]  Ido Dagan,et al.  Supervised Open Information Extraction , 2018, NAACL.