Mastering the Task of Open Information Extraction with Large Language Models and Consistent Reasoning Environment
暂无分享,去创建一个
Xiaozhi Wang | Lei Hou | Jifan Yu | Juanzi Li | Kaisheng Zeng | Ji Qi | Kaixuan Ji | Bin Xu
[1] Hao Peng,et al. The Devil is in the Details: On the Pitfalls of Event Extraction Evaluation , 2023, ACL.
[2] Xiaozhi Wang,et al. Preserving Knowledge Invariance: Rethinking Robustness Evaluation of Open Information Extraction , 2023, EMNLP.
[3] Lu Liu,et al. Is Information Extraction Solved by ChatGPT? An Analysis of Performance, Evaluation Criteria, Robustness and Errors , 2023, ArXiv.
[4] S. Kurohashi,et al. GPT-RE: In-context Learning for Relation Extraction using Large Language Models , 2023, EMNLP.
[5] Ningyu Zhang,et al. How to Unleash the Power of Large Language Models for Few-shot Relation Extraction? , 2023, SUSTAINLP.
[6] Quansen Wang,et al. Evaluating ChatGPT's Information Extraction Capabilities: An Assessment of Performance, Explainability, Calibration, and Faithfulness , 2023, ArXiv.
[7] S. Ananiadou,et al. Zero-shot Temporal Relation Extraction with ChatGPT , 2023, BIONLP.
[8] Huan Zhao,et al. Exploring the Feasibility of ChatGPT for Event Extraction , 2023, ArXiv.
[9] Naman Goyal,et al. LLaMA: Open and Efficient Foundation Language Models , 2023, ArXiv.
[10] Yufeng Chen,et al. Zero-Shot Information Extraction via Chatting with ChatGPT , 2023, ArXiv.
[11] Michihiro Yasunaga,et al. Is ChatGPT a General-Purpose Natural Language Processing Task Solver? , 2023, EMNLP.
[12] Davide Andreoletti,et al. ChatGPT or Human? Detect and Explain. Explaining Decisions of Machine Learning Model for Detecting Short ChatGPT-generated Text , 2023, ArXiv.
[13] Ziyuan Wang,et al. How Close is ChatGPT to Human Experts? Comparison Corpus, Evaluation, and Detection , 2023, ArXiv.
[14] Bin Xu,et al. Syntactically Robust Training on Partially-Observed Data for Open Information Extraction , 2023, EMNLP.
[15] Bowen Zhang,et al. How would Stance Detection Techniques Evolve after the Launch of ChatGPT? , 2022, ArXiv.
[16] Wei Ye,et al. Reviewing Labels: Label Graph Network with Top-k Prediction Set for Relation Extraction , 2022, AAAI.
[17] K. Chang,et al. When to Use What: An In-Depth Comparative Empirical Analysis of OpenIE Systems for Downstream Applications , 2022, ACL.
[18] Yiming Zhang,et al. Active Example Selection for In-Context Learning , 2022, EMNLP.
[19] Andrew M. Dai,et al. Scaling Instruction-Finetuned Language Models , 2022, ArXiv.
[20] Andrew M. Dai,et al. PaLM: Scaling Language Modeling with Pathways , 2022, J. Mach. Learn. Res..
[21] Ryan J. Lowe,et al. Training language models to follow instructions with human feedback , 2022, NeurIPS.
[22] Jonathan Berant,et al. Learning To Retrieve Prompts for In-Context Learning , 2021, NAACL.
[23] Rajarshi Das,et al. Case-based Reasoning for Natural Language Queries over Knowledge Bases , 2021, EMNLP.
[24] Dan Klein,et al. Constrained Language Models Yield Few-Shot Semantic Parsers , 2021, EMNLP.
[25] Weizhu Chen,et al. What Makes Good In-Context Examples for GPT-3? , 2021, DEELIO.
[26] Rainer Gemulla,et al. On Aligning OpenIE Extractions with Knowledge Bases: A Case Study , 2020, EVAL4NLP.
[27] Yukyung Lee,et al. Multiˆ2OIE: Multilingual Open Information Extraction based on Multi-Head Attention with BERT , 2020, FINDINGS.
[28] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[29] Mausam,et al. IMoJIE: Iterative Memory-Based Joint Open Information Extraction , 2020, ACL.
[30] Mausam,et al. CaRB: A Crowdsourced Benchmark for Open IE , 2019, EMNLP.
[31] Shimei Pan,et al. Supervising Unsupervised Open Information Extraction Models , 2019, EMNLP.
[32] Hannaneh Hajishirzi,et al. Entity, Relation, and Event Extraction with Contextualized Span Representations , 2019, EMNLP.
[33] Hai Zhao,et al. Span Model for Open Information Extraction on Accurate Corpus , 2019, AAAI.
[34] Ido Dagan,et al. Supervised Open Information Extraction , 2018, NAACL.
[35] Ming Zhou,et al. Neural Open Information Extraction , 2018, ACL.
[36] Miao Fan,et al. Logician: A Unified End-to-End Neural Approach for Open-Domain Information Extraction , 2018, WSDM.
[37] Danqi Chen,et al. Position-aware Attention and Supervised Data Improve Slot Filling , 2017, EMNLP.
[38] Harinder Pal,et al. Bootstrapping for Numerical Open IE , 2017, ACL.
[39] Shane Legg,et al. Deep Reinforcement Learning from Human Preferences , 2017, NIPS.
[40] Mausam,et al. Open Information Extraction Systems and Downstream Applications , 2016, IJCAI.
[41] Harinder Pal,et al. Demonyms and Compound Relational Nouns in Nominal Open IE , 2016, AKBC@NAACL-HLT.
[42] Christopher D. Manning,et al. Leveraging Linguistic Structure For Open Domain Information Extraction , 2015, ACL.
[43] Luciano Del Corro,et al. ClausIE: clause-based open information extraction , 2013, WWW.
[44] Oren Etzioni,et al. Open Language Learning for Information Extraction , 2012, EMNLP.
[45] Oren Etzioni,et al. An analysis of open information extraction based on semantic role labeling , 2011, K-CAP '11.
[46] Mascha Kaléko,et al. What You Need… , 2010 .
[47] Oren Etzioni,et al. Open Information Extraction from the Web , 2007, CACM.
[48] T. Qian,et al. Graph-based Model Generation for Few-Shot Relation Extraction , 2022, EMNLP.
[49] Yufeng Chen,et al. Saliency as Evidence: Event Detection with Trigger Saliency Attribution , 2022, ACL.
[50] Hongsong Zhu,et al. Maximal Clique Based Non-Autoregressive Open Information Extraction , 2021, EMNLP.