暂无分享,去创建一个
[1] Philipp Koehn,et al. Abstract Meaning Representation for Sembanking , 2013, LAW@ACL.
[2] Samuel R. Bowman,et al. A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference , 2017, NAACL.
[3] Stephan Oepen,et al. Broad-Coverage Semantic Dependency Parsing , 2014 .
[4] Hwee Tou Ng,et al. Towards Robust Linguistic Analysis using OntoNotes , 2013, CoNLL.
[5] Ido Dagan,et al. Supervised Open Information Extraction , 2018, NAACL.
[6] Roland Vollgraf,et al. Contextual String Embeddings for Sequence Labeling , 2018, COLING.
[7] Andrew McCallum,et al. Linguistically-Informed Self-Attention for Semantic Role Labeling , 2018, EMNLP.
[8] Omer Levy,et al. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding , 2018, BlackboxNLP@EMNLP.
[9] Omer Levy,et al. Zero-Shot Relation Extraction via Reading Comprehension , 2017, CoNLL.
[10] Martha Palmer,et al. From TreeBank to PropBank , 2002, LREC.
[11] Luke S. Zettlemoyer,et al. Deep Semantic Role Labeling: What Works and What’s Next , 2017, ACL.
[12] Alex Wang,et al. What do you learn from context? Probing for sentence structure in contextualized word representations , 2019, ICLR.
[13] Erik F. Tjong Kim Sang,et al. Introduction to the CoNLL-2003 Shared Task: Language-Independent Named Entity Recognition , 2003, CoNLL.
[14] Ido Dagan,et al. Crowdsourcing Question-Answer Meaning Representations , 2017, NAACL.
[15] Ming-Wei Chang,et al. Guiding Semi-Supervision with Constraint-Driven Learning , 2007, ACL.
[16] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.
[17] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[18] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[19] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[20] Bernhard Schölkopf,et al. Fidelity-Weighted Learning , 2017, ICLR.
[21] Luke S. Zettlemoyer,et al. Higher-Order Coreference Resolution with Coarse-to-Fine Inference , 2018, NAACL.
[22] Timothy Dozat,et al. Simpler but More Accurate Semantic Dependency Parsing , 2018, ACL.
[23] Preslav Nakov,et al. SemEval-2010 Task 8: Multi-Way Classification of Semantic Relations Between Pairs of Nominals , 2009, SEW@NAACL-HLT.
[24] Dan Roth,et al. Question Answering as Global Reasoning Over Semantic Abstractions , 2018, AAAI.
[25] Luke S. Zettlemoyer,et al. End-to-end Neural Coreference Resolution , 2017, EMNLP.
[26] Luke S. Zettlemoyer,et al. Question-Answer Driven Semantic Role Labeling: Using Natural Language to Annotate Natural Language , 2015, EMNLP.
[27] Dipanjan Das,et al. BERT Rediscovers the Classical NLP Pipeline , 2019, ACL.
[28] Christopher D. Manning,et al. Natural Logic for Textual Inference , 2007, ACL-PASCAL@ACL.
[29] Christopher Potts,et al. Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank , 2013, EMNLP.
[30] Diego Marcheggiani,et al. Encoding Sentences with Graph Convolutional Networks for Semantic Role Labeling , 2017, EMNLP.
[31] Zhiyuan Liu,et al. Relation Classification via Multi-Level Attention CNNs , 2016, ACL.
[32] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[33] Wei Shi,et al. Attention-Based Bidirectional Long Short-Term Memory Networks for Relation Classification , 2016, ACL.
[34] Yuchen Zhang,et al. CoNLL-2012 Shared Task: Modeling Multilingual Unrestricted Coreference in OntoNotes , 2012, EMNLP-CoNLL Shared Task.
[35] Luke S. Zettlemoyer,et al. Large-Scale QA-SRL Parsing , 2018, ACL.
[36] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[37] Ming-Wei Chang,et al. Driving Semantic Parsing from the World’s Response , 2010, CoNLL.
[38] Dan Roth,et al. Incidental Supervision: Moving beyond Supervised Learning , 2017, AAAI.
[39] Dan Roth,et al. Partial Or Complete, That’s The Question , 2019, NAACL.
[40] Richard Socher,et al. Learned in Translation: Contextualized Word Vectors , 2017, NIPS.
[41] Ali Farhadi,et al. Bidirectional Attention Flow for Machine Comprehension , 2016, ICLR.