暂无分享,去创建一个
Makoto Yamada | Ruslan Salakhutdinov | Louis-Philippe Morency | Han Zhao | Yao-Hung Hubert Tsai | R. Salakhutdinov | Louis-Philippe Morency | Han Zhao | M. Yamada
[1] Kaiming He,et al. Momentum Contrast for Unsupervised Visual Representation Learning , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[2] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[3] Gerlof Bouma,et al. Normalized (pointwise) mutual information in collocation extraction , 2009 .
[4] Kurt Hornik,et al. Multilayer feedforward networks are universal approximators , 1989, Neural Networks.
[5] Jianfeng Gao,et al. A Diversity-Promoting Objective Function for Neural Conversation Models , 2015, NAACL.
[6] James R. Glass,et al. Speech2Vec: A Sequence-to-Sequence Framework for Learning Word Embeddings from Speech , 2018, INTERSPEECH.
[7] Ruslan Salakhutdinov,et al. Learning Factorized Multimodal Representations , 2018, ICLR.
[8] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[9] A. Kraskov,et al. Estimating mutual information. , 2003, Physical review. E, Statistical, nonlinear, and soft matter physics.
[10] Ali Razavi,et al. Data-Efficient Image Recognition with Contrastive Predictive Coding , 2019, ICML.
[11] R Devon Hjelm,et al. Learning Representations by Maximizing Mutual Information Across Views , 2019, NeurIPS.
[12] Alexander A. Alemi,et al. On Variational Bounds of Mutual Information , 2019, ICML.
[13] Pieter Abbeel,et al. InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets , 2016, NIPS.
[14] Martin J. Wainwright,et al. Estimating Divergence Functionals and the Likelihood Ratio by Convex Risk Minimization , 2008, IEEE Transactions on Information Theory.
[15] Omer Levy,et al. Neural Word Embedding as Implicit Matrix Factorization , 2014, NIPS.
[16] S. Varadhan,et al. Asymptotic evaluation of certain Markov process expectations for large time , 1975 .
[17] Geoffrey E. Hinton,et al. A Simple Framework for Contrastive Learning of Visual Representations , 2020, ICML.
[18] Yoshua Bengio,et al. Learning deep representations by mutual information estimation and maximization , 2018, ICLR.
[19] Fuhui Long,et al. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy , 2003, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[20] Stefano Ermon,et al. Understanding the Limitations of Variational Mutual Information Estimators , 2020, ICLR.
[21] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[22] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[23] Phillip Isola,et al. Contrastive Multiview Coding , 2019, ECCV.
[24] Kenneth Ward Church,et al. Word Association Norms, Mutual Information, and Lexicography , 1989, ACL.
[25] Xing-Ming Zhao,et al. Inferring gene regulatory networks from gene expression data by path consistency algorithm based on conditional mutual information , 2012, Bioinform..
[26] Michael Tschannen,et al. On Mutual Information Maximization for Representation Learning , 2019, ICLR.
[27] Lei Yu,et al. A Mutual Information Maximization Perspective of Language Representation Learning , 2019, ICLR.
[28] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[29] Aaron C. Courville,et al. MINE: Mutual Information Neural Estimation , 2018, ArXiv.
[30] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[31] Jitendra Malik,et al. Learning to See by Moving , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[32] Yuan Yu,et al. TensorFlow: A system for large-scale machine learning , 2016, OSDI.
[33] M. Kawanabe,et al. Direct importance estimation for covariate shift adaptation , 2008 .
[34] Masashi Sugiyama,et al. Density-ratio matching under the Bregman divergence: a unified framework of density-ratio estimation , 2012 .
[35] Ruslan Salakhutdinov,et al. Self-supervised Learning from a Multi-view Perspective , 2020, ICLR.
[36] Peter L. Bartlett,et al. The Sample Complexity of Pattern Classification with Neural Networks: The Size of the Weights is More Important than the Size of the Network , 1998, IEEE Trans. Inf. Theory.
[37] Mikhail Khodak,et al. A Theoretical Analysis of Contrastive Unsupervised Representation Learning , 2019, ICML.
[38] Takafumi Kanamori,et al. A Least-squares Approach to Direct Importance Estimation , 2009, J. Mach. Learn. Res..
[39] Christoph Molnar,et al. Interpretable Machine Learning , 2020 .
[40] Sergey Levine,et al. Wasserstein Dependency Measure for Representation Learning , 2019, NeurIPS.
[41] Takafumi Kanamori,et al. Density Ratio Estimation in Machine Learning , 2012 .
[42] Jeffrey Dean,et al. Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.
[43] Kristen Grauman,et al. Learning Image Representations Tied to Ego-Motion , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[44] Peter L. Bartlett,et al. Neural Network Learning - Theoretical Foundations , 1999 .
[45] M. C. Jones,et al. Robust and efficient estimation by minimising a density power divergence , 1998 .
[46] David Barber,et al. The IM algorithm: a variational approach to Information Maximization , 2003, NIPS 2003.
[47] Oriol Vinyals,et al. Representation Learning with Contrastive Predictive Coding , 2018, ArXiv.
[48] L. Bregman. The relaxation method of finding the common point of convex sets and its application to the solution of problems in convex programming , 1967 .
[49] Le Song,et al. Learning to Explain: An Information-Theoretic Perspective on Model Interpretation , 2018, ICML.
[50] Andrew Zisserman,et al. Look, Listen and Learn , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[51] Alexander Kolesnikov,et al. Revisiting Self-Supervised Visual Representation Learning , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[52] Kentaro Inui,et al. Pointwise HSIC: A Linear-Time Kernelized Co-occurrence Norm for Sparse Linguistic Expressions , 2018, EMNLP.
[53] Sebastian Nowozin,et al. f-GAN: Training Generative Neural Samplers using Variational Divergence Minimization , 2016, NIPS.