We computed linguistic information at the lexical, syntactic, and semantic levels for the RITE (Recognizing Inference in TExt) tasks for both traditional and simplified Chinese in NTCIR-10. Techniques for syntactic parsing, named-entity recognition, and near synonym recognition were employed, and features like counts of common words, sentence lengths, negation words, and antonyms were considered to judge the logical relationships of two sentences, while we explored both heuristics-based functions and machine-learning approaches. We focused on the BC (binary classification) task at the preparatory stage, but participated in both BC and MC (multiple classes) evaluations. Three settings were submitted for the formal runs for each task. The best performing settings achieved the second best performance in BC tasks, and were listed in the top five performers in MC tasks for both traditional and simplified Chinese.
[1]
Shuming Shi,et al.
Overview of NTCIR-9 RITE: Recognizing Inference in TExt
,
2011,
NTCIR.
[2]
Ido Dagan,et al.
The Third PASCAL Recognizing Textual Entailment Challenge
,
2007,
ACL-PASCAL@ACL.
[3]
Rodney D. Nielsen,et al.
Recognizing entailment in intelligent tutoring systems*
,
2009,
Natural Language Engineering.
[4]
Chih-Jen Lin,et al.
LIBSVM: A library for support vector machines
,
2011,
TIST.
[5]
Changning Huang,et al.
Chinese Word Segmentation and Named Entity Recognition: A Pragmatic Approach
,
2005,
CL.