Answering Yes/No Questions in Legal Bar Exams

The development of Question Answering (QA) systems has become important because it reveals research issues that require insight from a variety of disciplines, including Artificial Intelligence, Information Extraction, Natural Language Processing, and Psychology. Our goal here is to develop a QA approach to answer yes/no questions relevant to civil laws in legal bar exams. A bar examination is intended to determine whether a candidate is qualified to practice law in a given jurisdiction. We have found that the development of a QA system for this task provides insight into the challenges of formalizing reasoning about legal text, and about how to exploit advances in computational linguistics. We separate our QA approach into two steps. The first step is to identify legal documents relevant to the exam questions; the second step is to answer the questions by analyzing the relevant documents. In our initial approach described here, the first step has been already solved for us: the appropriate articles for each question have been identified by legal experts. So here, we focus on the second task, which can be considered as a form of Recognizing Textual Entailment (RTE), where input to the system is a question sentence and its corresponding civil law article(s), and the output is a binary answer: whether the question sentence is entailed from the article(s). We propose a hybrid method, which combines simple rules and an unsupervised learning model using deep linguistic features. We first construct a knowledge base for negation and antonym words for the legal domain. We then identify potential premise and conclusion components of input questions and documents, based on text patterns and separating commas. We further classify the questions into easy and difficult ones, and develop a two-phase method for answering yes/no questions. We answer easy questions by negation/antonym detection. For more difficult questions, we adapt an unsupervised machine learning method based on morphological, syntactic, and lexical semantic analysis on identified premises and conclusions. This provides the basis to compare the semantic correlation between a question and a legal article. Our experimental results show reasonable performance, which improves the baseline system, and outperforms an SVM-based supervised machine learning model.

[1]  Ricardo Gonçalves,et al.  Normative Systems Represented as Hybrid Knowledge Bases , 2011, CLIMA.

[2]  Christopher D. Manning,et al.  Natural Logic for Textual Inference , 2007, ACL-PASCAL@ACL.

[3]  Ken Satoh,et al.  PROLEG: An Implementation of the Presupposed Ultimate Fact Theory of Japanese Civil Code by PROLOG Technology , 2010, JSAI-isAI Workshops.

[4]  Dieter Merkl,et al.  En route to data mining in legal text corpora: clustering, neural computation, and international treaties , 1997, Database and Expert Systems Applications. 8th International Conference, DEXA '97. Proceedings.

[5]  Trevor J. M. Bench-Capon What Makes a System a Legal Expert? , 2012, JURIX.

[6]  Lenhart K. Schubert,et al.  Entailment Inference in a Natural Logic-like General Reasoner , 2010, AAAI Fall Symposium: Commonsense Knowledge.

[7]  Arti Arya,et al.  A Knowledge Based Approach for Recognizing Textual Entailment for Natural Language Inference using Data Mining , 2010 .

[8]  Ian H. Witten,et al.  The WEKA data mining software: an update , 2009, SKDD.

[9]  Arul Menezes,et al.  Effectively Using Syntax for Recognizing False Entailment , 2006, NAACL.

[10]  Yue Zhang,et al.  Recognizing Text Entailment via Syntactic Tree Matching , 2011, NTCIR.

[11]  Arul Menezes,et al.  Syntactic Contributions in the Entailment Task: an implementation , 2005 .

[12]  Andreas Hamfelt,et al.  Towards a dynamic metalogic implementation of legal argumentation , 2011, ICAIL.

[13]  Marcin Walas How to Answer Yes/No Spatial Questions Using Qualitative Reasoning? , 2012, CICLing.

[14]  Emiel Krahmer,et al.  Dependency-based paraphrasing for recognizing textual entailment , 2007, ACL-PASCAL@ACL.

[15]  Stefan Harmeling An Extensible Probabilistic Transformation-based Approach to the Third Recognizing Textual Entailment Challenge , 2007, ACL-PASCAL@ACL.

[16]  Christopher D. Manning,et al.  Learning to recognize features of valid textual entailments , 2006, NAACL.

[17]  E. Maier New Frontiers in Artificial Intelligence , 2009 .

[18]  Valentin Jijkoun,et al.  Recognizing Textual Entailment Using Lexical Similarity , 2005 .

[19]  Anselmo Peñas,et al.  Overview of ResPubliQA 2009: Question Answering Evaluation over European Legislation , 2009, CLEF.

[20]  Natheer K. Gharaibeh,et al.  Development of Yes/No Arabic Question Answering System , 2013, ArXiv.

[21]  Alexander Gelbukh,et al.  Computational Linguistics and Intelligent Text Processing , 2015, Lecture Notes in Computer Science.

[22]  Alessandro Moschitti,et al.  Learning textual entailment from examples , 2006 .

[23]  Mi-Young Kim,et al.  Resolving Ambiguity in Inter-chunk Dependency Parsing , 2001, NLPRS.

[24]  Trevor J. M. Bench-Capon,et al.  On the Instantiation of Knowledge Bases in Abstract Argumentation Frameworks , 2013, CLIMA.

[25]  Alexander F. Gelbukh,et al.  NLP for Shallow Question Answering of Legal Documents Using Graphs , 2009, CICLing.

[26]  Christopher D. Manning,et al.  A Phrase-Based Alignment Model for Natural Language Inference , 2008, EMNLP.