暂无分享,去创建一个
[1] Joelle Pineau,et al. Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models , 2015, AAAI.
[2] Justin Domke,et al. Learning Graphical Model Parameters with Approximate Marginal Inference , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[3] Alexander M. Rush,et al. Latent Normalizing Flows for Discrete Sequences , 2019, ICML.
[4] Jason Lee,et al. Deterministic Non-Autoregressive Neural Sequence Modeling by Iterative Refinement , 2018, EMNLP.
[5] Hedvig Kjellström,et al. Advances in Variational Inference , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[6] Andreas Stolcke,et al. SRILM - an extensible language modeling toolkit , 2002, INTERSPEECH.
[7] Slav Petrov,et al. Globally Normalized Transition-Based Neural Networks , 2016, ACL.
[8] Wei Xu,et al. Bidirectional LSTM-CRF Models for Sequence Tagging , 2015, ArXiv.
[9] Vladlen Koltun,et al. Efficient Inference in Fully Connected CRFs with Gaussian Edge Potentials , 2011, NIPS.
[10] Stanislau Semeniuta,et al. On Accurate Evaluation of GANs for Language Generation , 2018, ArXiv.
[11] Ole Winther,et al. Sequential Neural Models with Stochastic Layers , 2016, NIPS.
[12] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[13] Max Welling,et al. Improved Variational Inference with Inverse Autoregressive Flow , 2016, NIPS 2016.
[14] Joelle Pineau,et al. Language GANs Falling Short , 2018, ICLR.
[15] Alexander M. Rush,et al. Structured Attention Networks , 2017, ICLR.
[16] Samy Bengio,et al. Scheduled Sampling for Sequence Prediction with Recurrent Neural Networks , 2015, NIPS.
[17] David M. Blei,et al. Variational Inference: A Review for Statisticians , 2016, ArXiv.
[18] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[19] Fernando Pereira,et al. Shallow Parsing with Conditional Random Fields , 2003, NAACL.
[20] Daan Wierstra,et al. Stochastic Back-propagation and Variational Inference in Deep Latent Gaussian Models , 2014, ArXiv.
[21] Jason Weston,et al. A Neural Attention Model for Abstractive Sentence Summarization , 2015, EMNLP.
[22] Yoshua Bengio,et al. Z-Forcing: Training Stochastic Recurrent Networks , 2017, NIPS.
[23] Anton Osokin,et al. SEARNN: Training RNNs with Global-Local Losses , 2017, ICLR.
[24] Andrew M. Dai,et al. MaskGAN: Better Text Generation via Filling in the ______ , 2018, ICLR.
[25] Geoffrey E. Hinton,et al. Deep Learning , 2015, Nature.
[26] Thomas Hofmann,et al. Deep State Space Models for Unconditional Word Generation , 2018, NeurIPS.
[27] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[28] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[29] Andrew McCallum,et al. End-to-End Learning for Structured Prediction Energy Networks , 2017, ICML.
[30] Lantao Yu,et al. SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient , 2016, AAAI.
[31] Doug Downey,et al. Controlling Global Statistics in Recurrent Neural Network Text Generation , 2018, AAAI.
[32] Yoshua Bengio,et al. Professor Forcing: A New Algorithm for Training Recurrent Networks , 2016, NIPS.
[33] Marc'Aurelio Ranzato,et al. Sequence Level Training with Recurrent Neural Networks , 2015, ICLR.
[34] Eduard H. Hovy,et al. End-to-end Sequence Labeling via Bi-directional LSTM-CNNs-CRF , 2016, ACL.
[35] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[36] Pratik Rane,et al. Self-Critical Sequence Training for Image Captioning , 2018 .
[37] Yang Liu,et al. Neural Machine Translation with Reconstruction , 2016, AAAI.
[38] Alexander M. Rush,et al. Sequence-to-Sequence Learning as Beam-Search Optimization , 2016, EMNLP.
[39] Christopher Potts,et al. A large annotated corpus for learning natural language inference , 2015, EMNLP.
[40] Jacob Goldberger,et al. CRF with deep class embedding for large scale classification , 2020, Comput. Vis. Image Underst..
[41] Jacob Goldberger,et al. Structured Image Classification from Conditional Random Field with Deep Class Embedding , 2017, ArXiv.
[42] Alex Graves,et al. Generating Sequences With Recurrent Neural Networks , 2013, ArXiv.
[43] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[44] Samy Bengio,et al. Generating Sentences from a Continuous Space , 2015, CoNLL.