Mastering the Task of Open Information Extraction with Large Language Models and Consistent Reasoning Environment

Open Information Extraction (OIE) aims to extract objective structured knowledge from natural texts, which has attracted growing attention to build dedicated models with human experience. As the large language models (LLMs) have exhibited remarkable in-context learning capabilities, a question arises as to whether the task of OIE can be effectively tackled with this paradigm? In this paper, we explore solving the OIE problem by constructing an appropriate reasoning environment for LLMs. Specifically, we first propose a method to effectively estimate the discrepancy of syntactic distribution between a LLM and test samples, which can serve as correlation evidence for preparing positive demonstrations. Upon the evidence, we introduce a simple yet effective mechanism to establish the reasoning environment for LLMs on specific tasks. Without bells and whistles, experimental results on the standard CaRB benchmark demonstrate that our $6$-shot approach outperforms state-of-the-art supervised method, achieving an $55.3$ $F_1$ score. Further experiments on TACRED and ACE05 show that our method can naturally generalize to other information extraction tasks, resulting in improvements of $5.7$ and $6.8$ $F_1$ scores, respectively.

[1]  Hao Peng,et al.  The Devil is in the Details: On the Pitfalls of Event Extraction Evaluation , 2023, ACL.

[2]  Xiaozhi Wang,et al.  Preserving Knowledge Invariance: Rethinking Robustness Evaluation of Open Information Extraction , 2023, EMNLP.

[3]  Lu Liu,et al.  Is Information Extraction Solved by ChatGPT? An Analysis of Performance, Evaluation Criteria, Robustness and Errors , 2023, ArXiv.

[4]  S. Kurohashi,et al.  GPT-RE: In-context Learning for Relation Extraction using Large Language Models , 2023, EMNLP.

[5]  Ningyu Zhang,et al.  How to Unleash the Power of Large Language Models for Few-shot Relation Extraction? , 2023, SUSTAINLP.

[6]  Quansen Wang,et al.  Evaluating ChatGPT's Information Extraction Capabilities: An Assessment of Performance, Explainability, Calibration, and Faithfulness , 2023, ArXiv.

[7]  S. Ananiadou,et al.  Zero-shot Temporal Relation Extraction with ChatGPT , 2023, BIONLP.

[8]  Huan Zhao,et al.  Exploring the Feasibility of ChatGPT for Event Extraction , 2023, ArXiv.

[9]  Naman Goyal,et al.  LLaMA: Open and Efficient Foundation Language Models , 2023, ArXiv.

[10]  Yufeng Chen,et al.  Zero-Shot Information Extraction via Chatting with ChatGPT , 2023, ArXiv.

[11]  Michihiro Yasunaga,et al.  Is ChatGPT a General-Purpose Natural Language Processing Task Solver? , 2023, EMNLP.

[12]  Davide Andreoletti,et al.  ChatGPT or Human? Detect and Explain. Explaining Decisions of Machine Learning Model for Detecting Short ChatGPT-generated Text , 2023, ArXiv.

[13]  Ziyuan Wang,et al.  How Close is ChatGPT to Human Experts? Comparison Corpus, Evaluation, and Detection , 2023, ArXiv.

[14]  Bin Xu,et al.  Syntactically Robust Training on Partially-Observed Data for Open Information Extraction , 2023, EMNLP.

[15]  Bowen Zhang,et al.  How would Stance Detection Techniques Evolve after the Launch of ChatGPT? , 2022, ArXiv.

[16]  Wei Ye,et al.  Reviewing Labels: Label Graph Network with Top-k Prediction Set for Relation Extraction , 2022, AAAI.

[17]  K. Chang,et al.  When to Use What: An In-Depth Comparative Empirical Analysis of OpenIE Systems for Downstream Applications , 2022, ACL.

[18]  Yiming Zhang,et al.  Active Example Selection for In-Context Learning , 2022, EMNLP.

[19]  Andrew M. Dai,et al.  Scaling Instruction-Finetuned Language Models , 2022, ArXiv.

[20]  Andrew M. Dai,et al.  PaLM: Scaling Language Modeling with Pathways , 2022, J. Mach. Learn. Res..

[21]  Ryan J. Lowe,et al.  Training language models to follow instructions with human feedback , 2022, NeurIPS.

[22]  Jonathan Berant,et al.  Learning To Retrieve Prompts for In-Context Learning , 2021, NAACL.

[23]  Rajarshi Das,et al.  Case-based Reasoning for Natural Language Queries over Knowledge Bases , 2021, EMNLP.

[24]  Dan Klein,et al.  Constrained Language Models Yield Few-Shot Semantic Parsers , 2021, EMNLP.

[25]  Weizhu Chen,et al.  What Makes Good In-Context Examples for GPT-3? , 2021, DEELIO.

[26]  Rainer Gemulla,et al.  On Aligning OpenIE Extractions with Knowledge Bases: A Case Study , 2020, EVAL4NLP.

[27]  Yukyung Lee,et al.  Multiˆ2OIE: Multilingual Open Information Extraction based on Multi-Head Attention with BERT , 2020, FINDINGS.

[28]  Mark Chen,et al.  Language Models are Few-Shot Learners , 2020, NeurIPS.

[29]  Mausam,et al.  IMoJIE: Iterative Memory-Based Joint Open Information Extraction , 2020, ACL.

[30]  Mausam,et al.  CaRB: A Crowdsourced Benchmark for Open IE , 2019, EMNLP.

[31]  Shimei Pan,et al.  Supervising Unsupervised Open Information Extraction Models , 2019, EMNLP.

[32]  Hannaneh Hajishirzi,et al.  Entity, Relation, and Event Extraction with Contextualized Span Representations , 2019, EMNLP.

[33]  Hai Zhao,et al.  Span Model for Open Information Extraction on Accurate Corpus , 2019, AAAI.

[34]  Ido Dagan,et al.  Supervised Open Information Extraction , 2018, NAACL.

[35]  Ming Zhou,et al.  Neural Open Information Extraction , 2018, ACL.

[36]  Miao Fan,et al.  Logician: A Unified End-to-End Neural Approach for Open-Domain Information Extraction , 2018, WSDM.

[37]  Danqi Chen,et al.  Position-aware Attention and Supervised Data Improve Slot Filling , 2017, EMNLP.

[38]  Harinder Pal,et al.  Bootstrapping for Numerical Open IE , 2017, ACL.

[39]  Shane Legg,et al.  Deep Reinforcement Learning from Human Preferences , 2017, NIPS.

[40]  Mausam,et al.  Open Information Extraction Systems and Downstream Applications , 2016, IJCAI.

[41]  Harinder Pal,et al.  Demonyms and Compound Relational Nouns in Nominal Open IE , 2016, AKBC@NAACL-HLT.

[42]  Christopher D. Manning,et al.  Leveraging Linguistic Structure For Open Domain Information Extraction , 2015, ACL.

[43]  Luciano Del Corro,et al.  ClausIE: clause-based open information extraction , 2013, WWW.

[44]  Oren Etzioni,et al.  Open Language Learning for Information Extraction , 2012, EMNLP.

[45]  Oren Etzioni,et al.  An analysis of open information extraction based on semantic role labeling , 2011, K-CAP '11.

[46]  Mascha Kaléko,et al.  What You Need… , 2010 .

[47]  Oren Etzioni,et al.  Open Information Extraction from the Web , 2007, CACM.

[48]  T. Qian,et al.  Graph-based Model Generation for Few-Shot Relation Extraction , 2022, EMNLP.

[49]  Yufeng Chen,et al.  Saliency as Evidence: Event Detection with Trigger Saliency Attribution , 2022, ACL.

[50]  Hongsong Zhu,et al.  Maximal Clique Based Non-Autoregressive Open Information Extraction , 2021, EMNLP.