暂无分享,去创建一个
Alexander Lerchner | Loic Matthey | Nicholas Watters | Christopher P. Burgess | Irina Higgins | Sunny Duan | I. Higgins | L. Matthey | Alexander Lerchner | Nicholas Watters | Sunny Duan
[1] Joshua B. Tenenbaum,et al. Building machines that learn and think like people , 2016, Behavioral and Brain Sciences.
[2] Nikolaus Kriegeskorte,et al. Representational Similarity Analysis – Connecting the Branches of Systems Neuroscience , 2008, Frontiers in systems neuroscience.
[3] Liwei Wang,et al. Towards Understanding Learning Representations: To What Extent Do Different Neural Networks Learn the Same Representation , 2018, NeurIPS.
[4] John Shawe-Taylor,et al. Canonical Correlation Analysis: An Overview with Application to Learning Methods , 2004, Neural Computation.
[5] Demis Hassabis,et al. A general reinforcement learning algorithm that masters chess, shogi, and Go through self-play , 2018, Science.
[6] Roger B. Grosse,et al. Isolating Sources of Disentanglement in Variational Autoencoders , 2018, NeurIPS.
[7] Christopher Burgess,et al. beta-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework , 2016, ICLR 2016.
[8] Christopher K. I. Williams,et al. A Framework for the Quantitative Evaluation of Disentangled Representations , 2018, ICLR.
[9] Yee Whye Teh,et al. Disentangling Disentanglement in Variational Autoencoders , 2018, ICML.
[10] Guillaume Desjardins,et al. Understanding disentangling in β-VAE , 2018, ArXiv.
[11] Pedro M. Domingos,et al. Deep Symmetry Networks , 2014, NIPS.
[12] Stefan Bauer,et al. On the Fairness of Disentangled Representations , 2019, NeurIPS.
[13] Daan Wierstra,et al. Stochastic Backpropagation and Approximate Inference in Deep Generative Models , 2014, ICML.
[14] Shane Legg,et al. IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures , 2018, ICML.
[15] Matthew Botvinick,et al. MONet: Unsupervised Scene Decomposition and Representation , 2019, ArXiv.
[16] Christopher Burgess,et al. DARLA: Improving Zero-Shot Transfer in Reinforcement Learning , 2017, ICML.
[17] Sergey Levine,et al. Visual Reinforcement Learning with Imagined Goals , 2018, NeurIPS.
[18] David Pfau,et al. Towards a Definition of Disentangled Representations , 2018, ArXiv.
[19] Tim Verbelen,et al. Improving Generalization for Abstract Reasoning Tasks Using Disentangled Feature Representations , 2018, NIPS 2018.
[20] Murray Shanahan,et al. SCAN: Learning Hierarchical Compositional Visual Concepts , 2017, ICLR.
[21] Guillaume Lample,et al. Phrase-Based & Neural Unsupervised Machine Translation , 2018, EMNLP.
[22] Hod Lipson,et al. Convergent Learning: Do different neural networks learn the same representations? , 2015, FE@NIPS.
[23] Pierre-Yves Oudeyer,et al. Curiosity Driven Exploration of Learned Disentangled Goal Spaces , 2018, CoRL.
[24] Tom Schaul,et al. Rainbow: Combining Improvements in Deep Reinforcement Learning , 2017, AAAI.
[25] Kevin Leyton-Brown,et al. Auto-WEKA: combined selection and hyperparameter optimization of classification algorithms , 2012, KDD.
[26] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[27] Pieter Abbeel,et al. InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets , 2016, NIPS.
[28] Andriy Mnih,et al. Disentangling by Factorising , 2018, ICML.
[29] Gary Marcus,et al. Deep Learning: A Critical Appraisal , 2018, ArXiv.
[30] Michael C. Mozer,et al. Learning Deep Disentangled Embeddings with the F-Statistic Loss , 2018, NeurIPS.
[31] Sjoerd van Steenkiste,et al. Are Disentangled Representations Helpful for Abstract Visual Reasoning? , 2019, NeurIPS.
[32] Pascal Vincent,et al. Representation Learning: A Review and New Perspectives , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[33] Tom Eccles,et al. Life-Long Disentangled Representation Learning with Cross-Domain Latent Homologies , 2018, NeurIPS.
[34] Stefan Bauer,et al. Interventional Robustness of Deep Latent Variable Models , 2018, ArXiv.
[35] Kevin Leyton-Brown,et al. Sequential Model-Based Optimization for General Algorithm Configuration , 2011, LION.
[36] Yoshua Bengio,et al. Algorithms for Hyper-Parameter Optimization , 2011, NIPS.
[37] Murray Shanahan,et al. Towards Deep Symbolic Reinforcement Learning , 2016, ArXiv.
[38] Max Jaderberg,et al. Population Based Training of Neural Networks , 2017, ArXiv.
[39] Jasper Snoek,et al. Practical Bayesian Optimization of Machine Learning Algorithms , 2012, NIPS.
[40] Max Welling,et al. Group Equivariant Convolutional Networks , 2016, ICML.
[41] Yuting Zhang,et al. Learning to Disentangle Factors of Variation with Manifold Interaction , 2014, ICML.
[42] Pascal Vincent,et al. Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion , 2010, J. Mach. Learn. Res..
[43] Jascha Sohl-Dickstein,et al. SVCCA: Singular Vector Canonical Correlation Analysis for Deep Learning Dynamics and Interpretability , 2017, NIPS.
[44] Samy Bengio,et al. Insights on representational similarity in neural networks with canonical correlation , 2018, NeurIPS.
[45] J. Urgen Schmidhuber,et al. Learning Factorial Codes by Predictability Minimization , 1992 .
[46] Pierre Comon,et al. Independent component analysis, A new concept? , 1994, Signal Process..
[47] Abhishek Kumar,et al. Variational Inference of Disentangled Latent Concepts from Unlabeled Observations , 2017, ICLR.
[48] Georg Martius,et al. Variational Autoencoders Pursue PCA Directions (by Accident) , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[49] Alexander Lerchner,et al. COBRA: Data-Efficient Model-Based RL through Unsupervised Object Discovery and Curiosity-Driven Exploration , 2019, ArXiv.
[50] Bernhard Schölkopf,et al. Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations , 2018, ICML.
[51] Heiga Zen,et al. WaveNet: A Generative Model for Raw Audio , 2016, SSW.
[52] Gang Sun,et al. Squeeze-and-Excitation Networks , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.