暂无分享,去创建一个
[1] Cordelia Schmid,et al. What makes for good views for contrastive learning , 2020, NeurIPS.
[2] Yoshua Bengio,et al. Learning deep representations by mutual information estimation and maximization , 2018, ICLR.
[3] Ali Razavi,et al. Data-Efficient Image Recognition with Contrastive Predictive Coding , 2019, ICML.
[4] Alexander Kolesnikov,et al. Revisiting Self-Supervised Visual Representation Learning , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[5] Moustapha Cissé,et al. Parseval Networks: Improving Robustness to Adversarial Examples , 2017, ICML.
[6] Masashi Sugiyama,et al. Lipschitz-Margin Training: Scalable Certification of Perturbation Invariance for Deep Neural Networks , 2018, NeurIPS.
[7] Daan Wierstra,et al. Towards Conceptual Compression , 2016, NIPS.
[8] Li Li,et al. Tensor Field Networks: Rotation- and Translation-Equivariant Neural Networks for 3D Point Clouds , 2018, ArXiv.
[9] Phillip Isola,et al. Contrastive Multiview Coding , 2019, ECCV.
[10] Kaiming He,et al. Momentum Contrast for Unsupervised Visual Representation Learning , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[11] Hu Chen,et al. Learning Invariant Representation for Unsupervised Image Restoration , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[12] Oriol Vinyals,et al. Representation Learning with Contrastive Predictive Coding , 2018, ArXiv.
[13] R Devon Hjelm,et al. Learning Representations by Maximizing Mutual Information Across Views , 2019, NeurIPS.
[14] Geoffrey E. Hinton,et al. A Simple Framework for Contrastive Learning of Visual Representations , 2020, ICML.
[15] Hongseok Yang,et al. On Nesting Monte Carlo Estimators , 2017, ICML.
[16] Premkumar Natarajan,et al. Invariant Representations through Adversarial Forgetting , 2019, AAAI.
[17] Sergey Levine,et al. Wasserstein Dependency Measure for Representation Learning , 2019, NeurIPS.
[18] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[19] O. Catoni. PAC-BAYESIAN SUPERVISED CLASSIFICATION: The Thermodynamics of Statistical Learning , 2007, 0712.0248.
[20] Yang You,et al. Large Batch Training of Convolutional Networks , 2017, 1708.03888.
[21] Song-Chun Zhu,et al. Deeper Interpretability of Deep Networks , 2018, ArXiv.
[22] Max Welling,et al. Spherical CNNs , 2018, ICLR.
[23] Aaron C. Courville,et al. Improved Training of Wasserstein GANs , 2017, NIPS.
[24] Alexander J. Smola,et al. Deep Sets , 2017, 1703.06114.
[25] Phillip Isola,et al. Understanding Contrastive Representation Learning through Alignment and Uniformity on the Hypersphere , 2020, ICML.
[26] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[27] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[28] Harris Drucker,et al. Improving generalization performance using double backpropagation , 1992, IEEE Trans. Neural Networks.
[29] Mark van der Wilk,et al. On the Benefits of Invariance in Neural Networks , 2020, ArXiv.
[30] Kaiming He,et al. Improved Baselines with Momentum Contrastive Learning , 2020, ArXiv.
[31] Geoffrey E. Hinton,et al. Big Self-Supervised Models are Strong Semi-Supervised Learners , 2020, NeurIPS.
[32] Yang You,et al. Scaling SGD Batch Size to 32K for ImageNet Training , 2017, ArXiv.
[33] Yann LeCun,et al. Dimensionality Reduction by Learning an Invariant Mapping , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).
[34] Max Welling,et al. Group Equivariant Convolutional Networks , 2016, ICML.
[35] Guillermo Sapiro,et al. Robust Large Margin Deep Neural Networks , 2016, IEEE Transactions on Signal Processing.
[36] Alexander A. Alemi,et al. On Variational Bounds of Mutual Information , 2019, ICML.