Toward constructing evidence-based legal arguments using legal decision documents and machine learning

This paper explores how to extract argumentation-relevant information automatically from a corpus of legal decision documents, and how to build new arguments using that information. For decision texts, we use the Vaccine/Injury Project (V/IP) Corpus, which contains default-logic annotations of argument structure. We supplement this with presuppositional annotations about entities, events, and relations that play important roles in argumentation, and about the level of confidence that arguments would be successful. We then propose how to integrate these semantic-pragmatic annotations with syntactic and domain-general semantic annotations, such as those generated in the DeepQA architecture, and outline how to apply machine learning and scoring techniques similar to those used in the IBM Watson system for playing the Jeopardy! question-answer game. We replace this game-playing goal, however, with the goal of learning to construct legal arguments.

[1]  Robert Tibshirani,et al.  An Introduction to the Bootstrap , 1994 .

[2]  Jennifer Chu-Carroll,et al.  Building Watson: An Overview of the DeepQA Project , 2010, AI Mag..

[3]  Michael C. McCord,et al.  Deep parsing in Watson , 2012, IBM J. Res. Dev..

[4]  Chang Wang,et al.  Relation extraction and scoring in DeepQA , 2012, IBM J. Res. Dev..

[5]  Marie-Francine Moens,et al.  Argumentation mining , 2011, Artificial Intelligence and Law.

[6]  David A. Ferrucci,et al.  Introduction to "This is Watson" , 2012, IBM J. Res. Dev..

[7]  Vern R. Walker A Default-Logic Paradigm for Legal Fact-Finding , 2007 .

[8]  Aditya Kalyanpur,et al.  Automatic knowledge extraction from documents , 2012, IBM J. Res. Dev..