Pre-Trained BERT-GRU Model for Relation Extraction
暂无分享,去创建一个
[1] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[2] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[3] Daniel Jurafsky,et al. Distant supervision for relation extraction without labeled data , 2009, ACL.
[4] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[5] Danqi Chen,et al. Position-aware Attention and Supervised Data Improve Slot Filling , 2017, EMNLP.
[6] Razvan C. Bunescu,et al. A Shortest Path Dependency Kernel for Relation Extraction , 2005, HLT.
[7] Ming Yang,et al. Bidirectional Long Short-Term Memory Networks for Relation Classification , 2015, PACLIC.
[8] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[9] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[10] Sanda M. Harabagiu,et al. UTD: Classifying Semantic Relations by Combining Lexical and Semantic Resources , 2010, *SEMEVAL.
[11] Christopher D. Manning,et al. Graph Convolution over Pruned Dependency Trees Improves Relation Extraction , 2018, EMNLP.
[12] Jun Zhao,et al. Relation Classification via Convolutional Deep Neural Network , 2014, COLING.
[13] Preslav Nakov,et al. SemEval-2010 Task 8: Multi-Way Classification of Semantic Relations Between Pairs of Nominals , 2009, SEW@NAACL-HLT.
[14] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[15] Wei Shi,et al. Attention-Based Bidirectional Long Short-Term Memory Networks for Relation Classification , 2016, ACL.
[16] George Kurian,et al. Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.
[17] Dong Wang,et al. Relation Classification via Recurrent Neural Network , 2015, ArXiv.