Improved Utilization Methodology of BERT Specialized in Text Classification
暂无分享,去创建一个
Jongtae Rhee | Hyeonji So | J. Rhee | H. So
[1] Da Luo,et al. Improving BERT-Based Text Classification With Auxiliary Sentence and Domain Knowledge , 2019, IEEE Access.
[2] James Philbin,et al. FaceNet: A unified embedding for face recognition and clustering , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[3] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[4] Xuanjing Huang,et al. How to Fine-Tune BERT for Text Classification? , 2019, CCL.
[5] Tomas Mikolov,et al. Enriching Word Vectors with Subword Information , 2016, TACL.
[6] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[7] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[8] Thorsten Joachims,et al. Learning a Distance Metric from Relative Comparisons , 2003, NIPS.
[9] Donald E. Brown,et al. Text Classification Algorithms: A Survey , 2019, Inf..
[10] Philip S. Yu,et al. BERT Post-Training for Review Reading Comprehension and Aspect-based Sentiment Analysis , 2019, NAACL.
[11] Jeffrey Dean,et al. Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.
[12] Ellen M. Voorhees,et al. The TREC-8 Question Answering Track Evaluation , 2000, TREC.
[13] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[14] Dogu Araci,et al. FinBERT: Financial Sentiment Analysis with Pre-trained Language Models , 2019, ArXiv.
[15] Yoshua Bengio,et al. A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..
[16] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[17] Geraldo Xexéo,et al. Word Embeddings: A Survey , 2019, ArXiv.
[18] Kilian Q. Weinberger,et al. Distance Metric Learning for Large Margin Nearest Neighbor Classification , 2005, NIPS.
[19] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.