Skip-Prop: Representing Sentences with One Vector Per Proposition
暂无分享,去创建一个
[1] Marco Marelli,et al. A SICK cure for the evaluation of compositional distributional semantic models , 2014, LREC.
[2] Sanja Fidler,et al. Skip-Thought Vectors , 2015, NIPS.
[3] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[4] Geoffrey E. Hinton,et al. Grammar as a Foreign Language , 2014, NIPS.
[5] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[6] Wei Xu,et al. End-to-end learning of semantic role labeling using recurrent neural networks , 2015, ACL.
[7] Benjamin Van Durme,et al. Annotated Gigaword , 2012, AKBC-WEKEX@NAACL-HLT.
[8] Eugene Charniak,et al. Effective Self-Training for Parsing , 2006, NAACL.
[9] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[10] John Hale,et al. The Information Conveyed by Words in Sentences , 2003, Journal of psycholinguistic research.
[11] Francis Ferraro,et al. Concretely Annotated Corpora , 2014 .
[12] Alexander M. Rush,et al. Character-Aware Neural Language Models , 2015, AAAI.
[13] Noah A. Smith,et al. Recurrent Neural Network Grammars , 2016, NAACL.
[14] Eugene Charniak,et al. Entropy Rate Constancy in Text , 2002, ACL.
[15] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[16] Richard Socher,et al. Ask Me Anything: Dynamic Memory Networks for Natural Language Processing , 2015, ICML.
[17] Thomas Hodgson,et al. Propositions , 2001, Peirce's Pragmatism.