SPLADE: Sparse Lexical and Expansion Model for First Stage Ranking
暂无分享,去创建一个
Benjamin Piwowarski | Stéphane Clinchant | Thibault Formal | S. Clinchant | Benjamin Piwowarski | Thibault Formal
[1] Barnabás Póczos,et al. Minimizing FLOPs to Learn Efficient Sparse Representations , 2020, ICLR.
[2] Jimmy J. Lin,et al. Distilling Dense Representations for Ranking using Tightly-Coupled Teachers , 2020, ArXiv.
[3] Tao Tao,et al. A formal study of information retrieval heuristics , 2004, SIGIR '04.
[4] Allan Hanbury,et al. Improving Efficient Neural Ranking Models with Cross-Architecture Knowledge Distillation , 2020, ArXiv.
[5] Jimmy J. Lin,et al. Document Expansion by Query Prediction , 2019, ArXiv.
[6] M. Zaharia,et al. ColBERT: Efficient and Effective Passage Search via Contextualized Late Interaction over BERT , 2020, SIGIR.
[7] Iryna Gurevych,et al. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks , 2019, EMNLP.
[8] D. Cheriton. From doc2query to docTTTTTquery , 2019 .
[9] Jun Xu,et al. SparTerm: Learning Term-based Sparse Representation for Fast Text Retrieval , 2020, ArXiv.
[10] Tiancheng Zhao,et al. SPARTA: Efficient Open-Domain Question Answering via Sparse Transformer Matching Retrieval , 2020, NAACL.
[11] Jimmy J. Lin,et al. Approximate Nearest Neighbor Search and Lightweight Dense Vector Reranking in Multi-Stage Retrieval Architectures , 2020, ICTIR.
[12] Jimmy J. Lin,et al. Critically Examining the "Neural Hype": Weak Baselines and the Additivity of Effectiveness Gains from Neural Ranking Models , 2019, SIGIR.
[13] Ming-Wei Chang,et al. REALM: Retrieval-Augmented Language Model Pre-Training , 2020, ICML.
[14] Danqi Chen,et al. Dense Passage Retrieval for Open-Domain Question Answering , 2020, EMNLP.
[15] Hua Wu,et al. RocketQA: An Optimized Training Approach to Dense Passage Retrieval for Open-Domain Question Answering , 2020, NAACL.
[16] Thomas Wolf,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[17] Siu Kwan Lam,et al. Numba: a LLVM-based Python JIT compiler , 2015, LLVM '15.
[18] Zhuyun Dai,et al. Context-Aware Sentence/Passage Term Importance Estimation For First Stage Retrieval , 2019, ArXiv.
[19] Raffaele Perego,et al. Expansion via Prediction of Importance with Contextualization , 2020, SIGIR.
[20] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[21] M. Zaharia,et al. ColBERT , 2020, Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval.
[22] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[23] Jamie Callan,et al. Context-Aware Term Weighting For First Stage Passage Retrieval , 2020, SIGIR.
[24] Ye Li,et al. Approximate Nearest Neighbor Negative Contrastive Learning for Dense Text Retrieval , 2020, ArXiv.
[25] Bhaskar Mitra,et al. Overview of the TREC 2019 deep learning track , 2020, ArXiv.
[26] Kyunghyun Cho,et al. Passage Re-ranking with BERT , 2019, ArXiv.
[27] James P. Callan,et al. Context-Aware Document Term Weighting for Ad-Hoc Search , 2020, WWW.
[28] W. Bruce Croft,et al. From Neural Re-Ranking to Neural Ranking: Learning a Sparse Representation for Inverted Indexing , 2018, CIKM.