暂无分享,去创建一个
Rishabh Singh | Bo Dai | Charles Sutton | Dale Schuurmans | Hanjun Dai | D. Schuurmans | Bo Dai | Rishabh Singh | H. Dai | Charles Sutton
[1] Geoffrey E. Hinton,et al. The Helmholtz Machine , 1995, Neural Computation.
[2] Emiel Hoogeboom,et al. Integer Discrete Flows and Lossless Compression , 2019, NeurIPS.
[3] Razvan Pascanu,et al. Learning Deep Generative Models of Graphs , 2018, ICLR 2018.
[4] Ben Poole,et al. Categorical Reparameterization with Gumbel-Softmax , 2016, ICLR.
[5] Yee Whye Teh,et al. The Concrete Distribution: A Continuous Relaxation of Discrete Random Variables , 2016, ICLR.
[6] Omer Levy,et al. Constant-Time Machine Translation with Conditional Masked Language Models , 2019, IJCNLP 2019.
[7] Percy Liang,et al. Learning Fast-Mixing Models for Structured Prediction , 2015, ICML.
[8] Kumar Krishna Agrawal,et al. Discrete Flows: Invertible Generative Models of Discrete Data , 2019, DGS@ICLR.
[9] Peter W. Glynn,et al. Likelihood ratio gradient estimation for stochastic systems , 1990, CACM.
[10] Yang Lu,et al. Cooperative Learning of Energy-Based Model and Latent Variable Model via MCMC Teaching , 2018, AAAI.
[11] Christopher D. Manning,et al. Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks , 2015, ACL.
[12] Barton P. Miller,et al. An empirical study of the reliability of UNIX utilities , 1990, Commun. ACM.
[13] Igor Mordatch,et al. Implicit Generation and Generalization with Energy Based Models , 2018 .
[14] Omer Levy,et al. Mask-Predict: Parallel Decoding of Conditional Masked Language Models , 2019, EMNLP.
[15] Le Song,et al. Kernel Exponential Family Estimation via Doubly Dual Embedding , 2018, AISTATS.
[16] Tijmen Tieleman,et al. Training restricted Boltzmann machines using approximations to the likelihood gradient , 2008, ICML '08.
[17] Fu Jie Huang,et al. A Tutorial on Energy-Based Learning , 2006 .
[18] Andrew McCallum,et al. Structured Prediction Energy Networks , 2015, ICML.
[19] Geoffrey E. Hinton,et al. The "wake-sleep" algorithm for unsupervised neural networks. , 1995, Science.
[20] Ronald J. Williams,et al. Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning , 2004, Machine Learning.
[21] Max Welling,et al. Semi-Supervised Classification with Graph Convolutional Networks , 2016, ICLR.
[22] Aapo Hyvärinen,et al. Estimation of Non-Normalized Statistical Models by Score Matching , 2005, J. Mach. Learn. Res..
[23] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[24] Matt J. Kusner,et al. Grammar Variational Autoencoder , 2017, ICML.
[25] Christopher King,et al. The CERT Guide to Coordinated Vulnerability Disclosure , 2017 .
[26] Lantao Yu,et al. SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient , 2016, AAAI.
[27] Jakob Uszkoreit,et al. Insertion Transformer: Flexible Sequence Generation via Insertion Operations , 2019, ICML.
[28] Hyrum S. Anderson,et al. The Malicious Use of Artificial Intelligence: Forecasting, Prevention, and Mitigation , 2018, ArXiv.
[29] Yoshua Bengio,et al. Estimating or Propagating Gradients Through Stochastic Neurons for Conditional Computation , 2013, ArXiv.
[30] Weinan Zhang,et al. GraphAF: a Flow-based Autoregressive Model for Molecular Graph Generation , 2020, ICLR.
[31] Rishabh Singh,et al. Learn&Fuzz: Machine learning for input fuzzing , 2017, 2017 32nd IEEE/ACM International Conference on Automated Software Engineering (ASE).
[32] Michael I. Jordan,et al. Graphical Models, Exponential Families, and Variational Inference , 2008, Found. Trends Mach. Learn..
[33] Karol Gregor,et al. Neural Variational Inference and Learning in Belief Networks , 2014, ICML.
[34] Pushmeet Kohli,et al. RobustFill: Neural Program Learning under Noisy I/O , 2017, ICML.
[35] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[36] Yuandong Tian,et al. Learning to Perform Local Rewriting for Combinatorial Optimization , 2019, NeurIPS.
[37] Rémi Munos,et al. Learning to Search with MCTSnets , 2018, ICML.
[38] Yang Lu,et al. Sparse and deep generalizations of the FRAME model , 2018 .
[39] Arthur Gretton,et al. KALE: When Energy-Based Learning Meets Adversarial Training , 2020, ArXiv.
[40] Jascha Sohl-Dickstein,et al. REBAR: Low-variance, unbiased gradient estimates for discrete latent variable models , 2017, NIPS.
[41] Ah Chung Tsoi,et al. The Graph Neural Network Model , 2009, IEEE Transactions on Neural Networks.
[42] Changhan Wang,et al. Levenshtein Transformer , 2019, NeurIPS.
[43] Sergey Levine,et al. MuProp: Unbiased Backpropagation for Stochastic Neural Networks , 2015, ICLR.
[44] Motoki Abe,et al. GraphNVP: An Invertible Flow Model for Generating Molecular Graphs , 2019, ArXiv.
[45] Myle Ott,et al. Residual Energy-Based Models for Text Generation , 2020, ICLR.
[46] David Duvenaud,et al. FFJORD: Free-form Continuous Dynamics for Scalable Reversible Generative Models , 2018, ICLR.
[47] Bernhard Schölkopf,et al. A Kernel Two-Sample Test , 2012, J. Mach. Learn. Res..
[48] Max Welling,et al. To Relieve Your Headache of Training an MRF, Take AdVIL , 2020, ICLR.
[49] Le Song,et al. Exponential Family Estimation via Adversarial Dynamics Embedding , 2019, NeurIPS.
[50] Lantao Yu,et al. Training Deep Energy-Based Models with f-Divergence Minimization , 2020, ICML.
[51] Sepp Hochreiter,et al. Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs) , 2015, ICLR.
[52] Rishabh Singh,et al. Learning Transferable Graph Exploration , 2019, NeurIPS.
[53] Sergey Levine,et al. The Mirage of Action-Dependent Baselines in Reinforcement Learning , 2018, ICML.
[54] Aapo Hyvärinen,et al. Noise-contrastive estimation: A new estimation principle for unnormalized statistical models , 2010, AISTATS.
[55] John Langford,et al. Learning to Search Better than Your Teacher , 2015, ICML.
[56] He He,et al. Learning to Search in Branch and Bound Algorithms , 2014, NIPS.
[57] Erik Nijkamp,et al. Learning Non-Convergent Non-Persistent Short-Run MCMC Toward Energy-Based Model , 2019, NeurIPS.
[58] Ben Taskar,et al. Max-Margin Markov Networks , 2003, NIPS.
[59] Mingyuan Zhou,et al. ARSM: Augment-REINFORCE-Swap-Merge Estimator for Gradient Backpropagation Through Categorical Variables , 2019, ICML.
[60] Radford M. Neal. MCMC Using Hamiltonian Dynamics , 2011, 1206.1901.
[61] Masahiro Ono,et al. Learning to Search via Retrospective Imitation , 2018, 1804.00846.
[62] David Duvenaud,et al. Backpropagation through the Void: Optimizing control variates for black-box gradient estimation , 2017, ICLR.
[63] Kaushalya Madhawa,et al. GraphNVP: an Invertible Flow-based Model for Generating Molecular Graphs , 2019 .
[64] Yoshua Bengio,et al. Deep Directed Generative Models with Energy-Based Probability Estimation , 2016, ArXiv.
[65] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[66] Stefano Ermon,et al. Neural Variational Inference and Learning in Undirected Graphical Models , 2017, NIPS.
[67] Nicola De Cao,et al. MolGAN: An implicit generative model for small molecular graphs , 2018, ArXiv.
[68] Arthur Gretton,et al. Generalized Energy Based Models , 2020, ICLR.
[69] J. Besag. Statistical Analysis of Non-Lattice Data , 1975 .
[70] Yee Whye Teh,et al. Bayesian Learning via Stochastic Gradient Langevin Dynamics , 2011, ICML.
[71] Andrew McCallum,et al. End-to-End Learning for Structured Prediction Energy Networks , 2017, ICML.
[72] Geoffrey E. Hinton. Training Products of Experts by Minimizing Contrastive Divergence , 2002, Neural Computation.
[73] Liang Lu,et al. Top-down Tree Long Short-Term Memory Networks , 2015, NAACL.
[74] Nenghai Yu,et al. Deliberation Networks: Sequence Generation Beyond One-Pass Decoding , 2017, NIPS.