Understanding tables with intermediate pre-training
暂无分享,去创建一个
[1] Marcelo Finger,et al. A logical-based corpus for cross-lingual evaluation , 2019, EMNLP.
[2] Carsten Binnig,et al. DBPal: A Fully Pluggable NL2SQL Training Pipeline , 2020, SIGMOD Conference.
[3] Percy Liang,et al. Compositional Semantic Parsing on Semi-Structured Tables , 2015, ACL.
[4] Wei-Cheng Chang,et al. Pre-training Tasks for Embedding-based Large-scale Retrieval , 2020, ICLR.
[5] Matthew Crosby,et al. Association for the Advancement of Artificial Intelligence , 2014 .
[6] D. Sculley,et al. Google Vizier: A Service for Black-Box Optimization , 2017, KDD.
[7] Tibor Kiss,et al. Unsupervised Multilingual Sentence Boundary Detection , 2006, CL.
[8] Jonathan Berant,et al. Injecting Numerical Reasoning Skills into Language Models , 2020, ACL.
[9] Steven Bird,et al. NLTK: The Natural Language Toolkit , 2002, ACL 2006.
[10] Yidong Chen,et al. Bilingually-constrained Synthetic Data for Implicit Discourse Relation Recognition , 2016, EMNLP.
[11] Reut Tsarfaty,et al. Evaluating NLP Models via Contrast Sets , 2020, ArXiv.
[12] Oren Etzioni,et al. Question Answering via Integer Programming over Semi-Structured Knowledge , 2016, IJCAI.
[13] Andreas Vlachos,et al. Identification and Verification of Simple Claims about Statistical Properties , 2015, EMNLP.
[14] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[15] Ming-Wei Chang,et al. Search-based Neural Structured Learning for Sequential Question Answering , 2017, ACL.
[16] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[17] Quoc V. Le,et al. ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators , 2020, ICLR.
[18] Vivek Srikumar,et al. INFOTABS: Inference on Tables as Semi-structured Data , 2020, ACL.
[19] Cong Yu,et al. AggChecker: A Fact-Checking System for Text Summaries of Relational Data Sets , 2019, Proc. VLDB Endow..
[20] Ming-Wei Chang,et al. BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions , 2019, NAACL.
[21] Philip Bachman,et al. Deep Reinforcement Learning that Matters , 2017, AAAI.
[22] Wenhu Chen,et al. TabFact: A Large-scale Dataset for Table-based Fact Verification , 2019, ICLR.
[23] Thibault Sellam,et al. BLEURT: Learning Robust Metrics for Text Generation , 2020, ACL.
[24] Andrew Y. Ng,et al. Parsing with Compositional Vector Grammars , 2013, ACL.
[25] Elia Bruni,et al. The Fast and the Flexible: Training Neural Networks to Learn to Follow Instructions from Small Data , 2018, IWCS.
[26] Nan Duan,et al. LogicalFactChecker: Leveraging Logical Operations for Fact Checking with Graph Module Network , 2020, ACL.
[27] Samuel R. Bowman,et al. Intermediate-Task Transfer Learning with Pretrained Language Models: When and Why Does It Work? , 2020, ACL.
[28] Wenhan Xiong,et al. Pretrained Encyclopedia: Weakly Supervised Knowledge-Pretrained Language Model , 2019, ICLR.
[29] Ludovic Denoyer,et al. Unsupervised Question Answering by Cloze Translation , 2019, ACL.
[30] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[31] Yoav Artzi,et al. A Corpus of Natural Language for Visual Reasoning , 2017, ACL.
[32] Zhiyuan Liu,et al. NumNet: Machine Reading Comprehension with Numerical Reasoning , 2019, EMNLP.
[33] Ming-Wei Chang,et al. Latent Retrieval for Weakly Supervised Open Domain Question Answering , 2019, ACL.
[34] Michael Collins,et al. Synthetic QA Corpora Generation with Roundtrip Consistency , 2019, ACL.
[35] Jonathan Berant,et al. Building a Semantic Parser Overnight , 2015, ACL.
[36] Peter Clark,et al. Project Aristo: Towards Machines that Capture and Reason with Science Knowledge , 2019, K-CAP.
[37] Ido Dagan,et al. Recognizing textual entailment: Rational, evaluation and approaches , 2009 .
[38] Steven Bird,et al. NLTK: The Natural Language Toolkit , 2002, ACL.
[39] Christopher Potts,et al. Posing Fair Generalization Tasks for Natural Language Inference , 2019, EMNLP.
[40] Sameer Singh,et al. Do NLP Models Know Numbers? Probing Numeracy in Embeddings , 2019, EMNLP.
[41] Christopher Potts,et al. A large annotated corpus for learning natural language inference , 2015, EMNLP.
[42] Thomas Müller,et al. Answering Conversational Questions on Structured Data without Logical Forms , 2019, EMNLP.
[43] Holger Schwenk,et al. Supervised Learning of Universal Sentence Representations from Natural Language Inference Data , 2017, EMNLP.
[44] Christopher Joseph Pal,et al. Learning General Purpose Distributed Sentence Representations via Large Scale Multi-task Learning , 2018, ICLR.
[45] Roy Schwartz,et al. Show Your Work: Improved Reporting of Experimental Results , 2019, EMNLP.
[46] Ido Dagan,et al. The Third PASCAL Recognizing Textual Entailment Challenge , 2007, ACL-PASCAL@ACL.
[47] Sebastian Ruder,et al. Universal Language Model Fine-tuning for Text Classification , 2018, ACL.
[48] Yoav Artzi,et al. A Corpus for Reasoning about Natural Language Grounded in Photographs , 2018, ACL.
[49] Doug Downey,et al. Don’t Stop Pretraining: Adapt Language Models to Domains and Tasks , 2020, ACL.
[50] Eduard Hovy,et al. Learning the Difference that Makes a Difference with Counterfactually-Augmented Data , 2020, ICLR.
[51] Thomas Muller,et al. TaPas: Weakly Supervised Table Parsing via Pre-training , 2020, ACL.
[52] Dan Roth,et al. “Ask Not What Textual Entailment Can Do for You...” , 2010, ACL.
[53] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.