MICRON: Multigranular Interaction for Contextualizing RepresentatiON in Non-factoid Question Answering

This paper studies the problem of non-factoid question answering, where the answer may span over multiple sentences. Existing solutions can be categorized into representation- and interaction-focused approaches. We combine their complementary strength, by a hybrid approach allowing multi-granular interactions, but represented at word level, enabling an easy integration with strong word-level signals. Specifically, we propose MICRON: Multigranular Interaction for Contextualizing RepresentatiON, a novel approach which derives contextualized uni-gram representation from n-grams. Our contributions are as follows: First, we enable multi-granular matches between question and answer n-grams. Second, by contextualizing word representation with surrounding n-grams, MICRON can naturally utilize word-based signals for query term weighting, known to be effective in information retrieval. We validate MICRON in two public non-factoid question answering datasets: WikiPassageQA and InsuranceQA, showing our model achieves the state of the art among baselines with reported performances on both datasets.

[1]  Iryna Gurevych,et al.  Representation Learning for Answer Selection with LSTM-Based Importance Weighting , 2017, IWCS.

[2]  Kathleen R. McKeown,et al.  A Hybrid Approach for Answering Definitional Questions , 2003 .

[3]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[4]  Bowen Zhou,et al.  Attentive Pooling Networks , 2016, ArXiv.

[5]  Bowen Zhou,et al.  Applying deep learning to answer selection: A study and an open task , 2015, 2015 IEEE Workshop on Automatic Speech Recognition and Understanding (ASRU).

[6]  Pengfei Zhang,et al.  Collaborative Learning for Answer Selection in Question Answering , 2019, IEEE Access.

[7]  Iryna Gurevych,et al.  COALA: A Neural Coverage-Based Approach for Long Answer Selection with Small Data , 2019, AAAI.

[8]  W. Bruce Croft,et al.  A Deep Look into Neural Ranking Models for Information Retrieval , 2019, Inf. Process. Manag..

[9]  Bowen Zhou,et al.  Improved Representation Learning for Question Answer Matching , 2016, ACL.

[10]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[11]  W. Bruce Croft,et al.  End to End Long Short Term Memory Networks for Non-Factoid Question Answering , 2016, ICTIR.

[12]  Dong Liu,et al.  MIX: Multi-Channel Information Crossing for Text Matching , 2018, KDD.

[13]  Seung-won Hwang,et al.  Optimizing access cost for top-k queries over Web sources: a unified cost-based approach , 2005, 21st International Conference on Data Engineering (ICDE'05).

[14]  W. Bruce Croft,et al.  WikiPassageQA: A Benchmark Collection for Research on Non-factoid Answer Passage Retrieval , 2018, SIGIR.

[15]  Ramesh Nallapati,et al.  Passage Ranking with Weak Supervsion , 2019, ArXiv.

[16]  Christopher D. Manning,et al.  Enhanced English Universal Dependencies: An Improved Representation for Natural Language Understanding Tasks , 2016, LREC.

[17]  W. Bruce Croft,et al.  A Deep Relevance Matching Model for Ad-hoc Retrieval , 2016, CIKM.

[18]  Shuohang Wang,et al.  A Compare-Aggregate Model for Matching Text Sequences , 2016, ICLR.

[19]  Jeffrey Pennington,et al.  GloVe: Global Vectors for Word Representation , 2014, EMNLP.

[20]  Luke S. Zettlemoyer,et al.  Deep Contextualized Word Representations , 2018, NAACL.