Rule Augmented Unsupervised Constituency Parsing
暂无分享,去创建一个
Rishabh K. Iyer | Rishabh Iyer | Atul Sahay | Ayush Maheshwari | Ganesh Ramakrishnan | Anshul Nasery | Ganesh Ramakrishnan | Ayush Maheshwari | Anshul Nasery | Atul Sahay
[1] Shiying Luo,et al. Weakly Supervised Sequence Tagging from Noisy Rules , 2020, AAAI.
[2] Lillian Lee,et al. Mostly-Unsupervised Statistical Segmentation of Japanese: Applications to Kanji , 2000, ANLP.
[3] Dan Klein,et al. Corpus-Based Induction of Syntactic Structure: Models of Dependency and Constituency , 2004, ACL.
[4] C. Lee Giles,et al. The Neural Network Pushdown Automaton: Model, Stack and Learning Simulations , 2017, ArXiv.
[5] Eric Brill,et al. Deducing Linguistic Structure from the Statistics of Large Corpora , 1990, HLT.
[6] Yoshimasa Tsuruoka,et al. Learning to Parse and Translate Improves Neural Machine Translation , 2017, ACL.
[7] Data Programming using Semi-Supervision and Subset Selection , 2020, ArXiv.
[8] Ganesh Ramakrishnan,et al. Robust Data Programming with Precision-guided Labeling Functions , 2020, AAAI.
[9] Gholamreza Haffari,et al. Incorporating Syntactic Uncertainty in Neural Machine Translation with a Forest-to-Sequence Model , 2017, COLING.
[10] Mihai Surdeanu,et al. The Stanford CoreNLP Natural Language Processing Toolkit , 2014, ACL.
[11] Jihun Choi,et al. Learning to Compose Task-Specific Tree Structures , 2017, AAAI.
[12] Mohit Yadav,et al. Unsupervised Latent Tree Induction with Deep Inside-Outside Recursive Auto-Encoders , 2019, NAACL.
[13] Pablo Gamallo,et al. Dependency-Based Open Information Extraction , 2012 .
[14] Wang Ling,et al. Learning to Compose Words into Sentences with Reinforcement Learning , 2016, ICLR.
[15] Christopher D. Manning,et al. Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks , 2015, ACL.
[16] Jihun Choi,et al. Are Pre-trained Language Models Aware of Phrases? Simple but Strong Baselines for Grammar Induction , 2020, ICLR.
[17] Samuel R. Bowman,et al. Do latent tree learning models identify meaningful structure in sentences? , 2017, TACL.
[18] Manjesh Kumar Hanawal,et al. Unsupervised Learning of Explainable Parse Trees for Improved Generalisation , 2021, 2021 International Joint Conference on Neural Networks (IJCNN).
[19] Sabine Buchholz,et al. Introduction to the CoNLL-2000 Shared Task Chunking , 2000, CoNLL/LLL.
[20] Aaron C. Courville,et al. Ordered Neurons: Integrating Tree Structures into Recurrent Neural Networks , 2018, ICLR.
[21] Yuchen Zhang,et al. CoNLL-2012 Shared Task: Modeling Multilingual Unrestricted Coreference in OntoNotes , 2012, EMNLP-CoNLL Shared Task.
[22] Aaron C. Courville,et al. Neural Language Modeling by Jointly Learning Syntax and Lexicon , 2017, ICLR.
[23] Jeffrey Pennington,et al. Semi-Supervised Recursive Autoencoders for Predicting Sentiment Distributions , 2011, EMNLP.
[24] Stephen Clark,et al. Jointly learning sentence embeddings and syntax with unsupervised Tree-LSTMs , 2017, Natural Language Engineering.
[25] Alexander M. Rush,et al. Compound Probabilistic Context-Free Grammars for Grammar Induction , 2019, ACL.
[26] Beatrice Santorini,et al. Building a Large Annotated Corpus of English: The Penn Treebank , 1993, CL.