Optimizing Dense Retrieval Model Training with Hard Negatives
暂无分享,去创建一个
Jiafeng Guo | Yiqun Liu | Shaoping Ma | Jiaxin Mao | Jingtao Zhan | Min Zhang | M. Zhang | Yiqun Liu | Shaoping Ma | Jingtao Zhan | Jiaxin Mao | Jiafeng Guo
[1] Christopher J. C. Burges,et al. From RankNet to LambdaRank to LambdaMART: An Overview , 2010 .
[2] Xuanhui Wang,et al. Learning-to-Rank with BERT in TF-Ranking , 2020, ArXiv.
[3] James Demmel,et al. Large Batch Optimization for Deep Learning: Training BERT in 76 minutes , 2019, ICLR.
[4] Pascal Vincent,et al. Representation Learning: A Review and New Perspectives , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[5] Chenliang Li,et al. IDST at TREC 2019 Deep Learning Track: Deep Cascade Ranking with Generation-based Document Expansion and Pre-trained Language Modeling , 2019, TREC.
[6] Jacob Eisenstein,et al. Sparse, Dense, and Attentional Representations for Text Retrieval , 2021, Transactions of the Association for Computational Linguistics.
[7] Danqi Chen,et al. Dense Passage Retrieval for Open-Domain Question Answering , 2020, EMNLP.
[8] Zhuyun Dai,et al. Context-Aware Sentence/Passage Term Importance Estimation For First Stage Retrieval , 2019, ArXiv.
[9] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[10] Yiqun Liu,et al. RepBERT: Contextualized Text Embeddings for First-Stage Retrieval , 2020, ArXiv.
[11] Hang Li,et al. Semantic Matching in Search , 2014, SMIR@SIGIR.
[12] Ye Li,et al. Approximate Nearest Neighbor Negative Contrastive Learning for Dense Text Retrieval , 2020, ArXiv.
[13] Stephen E. Robertson,et al. Some simple effective approximations to the 2-Poisson model for probabilistic weighted retrieval , 1994, SIGIR '94.
[14] Cordelia Schmid,et al. Product Quantization for Nearest Neighbor Search , 2011, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[15] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[16] Aapo Hyvärinen,et al. Noise-contrastive estimation: A new estimation principle for unnormalized statistical models , 2010, AISTATS.
[17] Jimmy J. Lin,et al. Document Expansion by Query Prediction , 2019, ArXiv.
[18] M. Zaharia,et al. ColBERT: Efficient and Effective Passage Search via Contextualized Late Interaction over BERT , 2020, SIGIR.
[19] Jimmy J. Lin,et al. Distilling Dense Representations for Ranking using Tightly-Coupled Teachers , 2020, ArXiv.
[20] Hua Wu,et al. RocketQA: An Optimized Training Approach to Dense Passage Retrieval for Open-Domain Question Answering , 2020, NAACL.
[21] Ming-Wei Chang,et al. REALM: Retrieval-Augmented Language Model Pre-Training , 2020, ICML.
[22] Jianfeng Gao,et al. A Human Generated MAchine Reading COmprehension Dataset , 2018 .
[23] Tie-Yan Liu. Learning to Rank for Information Retrieval , 2009, Found. Trends Inf. Retr..
[24] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[25] Frank Hutter,et al. Decoupled Weight Decay Regularization , 2017, ICLR.
[26] Tie-Yan Liu,et al. Learning to rank for information retrieval , 2009, SIGIR.
[27] Wei Liu,et al. Learning Binary Codes for Maximum Inner Product Search , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[28] Luyu Gao,et al. Complementing Lexical Retrieval with Semantic Residual Embedding , 2020, ArXiv.
[29] Bhaskar Mitra,et al. Overview of the TREC 2019 deep learning track , 2020, ArXiv.
[30] Kyunghyun Cho,et al. Passage Re-ranking with BERT , 2019, ArXiv.
[31] Jimmy J. Lin,et al. Anserini , 2018, Journal of Data and Information Quality.
[32] Jeff Johnson,et al. Billion-Scale Similarity Search with GPUs , 2017, IEEE Transactions on Big Data.
[33] Linjun Yang,et al. Embedding-based Retrieval in Facebook Search , 2020, KDD.
[34] Tao Qin,et al. A general approximation framework for direct optimization of information retrieval measures , 2010, Information Retrieval.
[35] King-Sun Fu,et al. IEEE Transactions on Pattern Analysis and Machine Intelligence Publication Information , 2004, IEEE Transactions on Pattern Analysis and Machine Intelligence.