暂无分享,去创建一个
Armand Joulin | Mahmoud Assran | Nicolas Ballas | Piotr Bojanowski | Ishan Misra | Mathilde Caron | Michael Rabbat | Michael G. Rabbat | Nicolas Ballas | Armand Joulin | Piotr Bojanowski | Ishan Misra | Mahmoud Assran | Mathilde Caron
[1] Kaiming He,et al. Momentum Contrast for Unsupervised Visual Representation Learning , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[2] Julien Mairal,et al. Unsupervised Learning of Visual Features by Contrasting Cluster Assignments , 2020, NeurIPS.
[3] Avrim Blum,et al. The Bottleneck , 2021, Monopsony Capitalism.
[4] Geoffrey E. Hinton,et al. Big Self-Supervised Models are Strong Semi-Supervised Learners , 2020, NeurIPS.
[5] Quoc V. Le,et al. AutoAugment: Learning Augmentation Strategies From Data , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[6] J. Piaget. Biology and knowledge;: An essay on the relations between organic regulations and cognitive processes , 1971 .
[7] Richard S. Zemel,et al. Prototypical Networks for Few-shot Learning , 2017, NIPS.
[8] Yang You,et al. Large Batch Training of Convolutional Networks , 2017, 1708.03888.
[9] Quoc V. Le,et al. Unsupervised Data Augmentation , 2019, ArXiv.
[10] Harri Valpola,et al. Weight-averaged consistency targets improve semi-supervised deep learning results , 2017, ArXiv.
[11] Seong Joon Oh,et al. CutMix: Regularization Strategy to Train Strong Classifiers With Localizable Features , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[12] Andrea Vedaldi,et al. Self-labelling via simultaneous clustering and representation learning , 2020, ICLR.
[13] Quoc V. Le,et al. Self-Training With Noisy Student Improves ImageNet Classification , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[14] Daisuke Kihara,et al. EnAET: Self-Trained Ensemble AutoEncoding Transformations for Semi-Supervised Learning , 2019, ArXiv.
[15] Alexander Kolesnikov,et al. S4L: Self-Supervised Semi-Supervised Learning , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[16] Sergey Levine,et al. Unsupervised Learning via Meta-Learning , 2018, ICLR.
[17] Yoshua Bengio,et al. Interpolation Consistency Training for Semi-Supervised Learning , 2019, IJCAI.
[18] David Yarowsky,et al. Unsupervised Word Sense Disambiguation Rivaling Supervised Methods , 1995, ACL.
[19] Geoffrey E. Hinton,et al. Learning a Nonlinear Embedding by Preserving Class Neighbourhood Structure , 2007, AISTATS.
[20] Joshua B. Tenenbaum,et al. Building machines that learn and think like people , 2016, Behavioral and Brain Sciences.
[21] Dong-Hyun Lee,et al. Pseudo-Label : The Simple and Efficient Semi-Supervised Learning Method for Deep Neural Networks , 2013 .
[22] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[23] Ellen Riloff,et al. Automatically Generating Extraction Patterns from Untagged Text , 1996, AAAI/IAAI, Vol. 2.
[24] Kaiming He,et al. Improved Baselines with Momentum Contrastive Learning , 2020, ArXiv.
[25] Quoc V. Le,et al. Rethinking Pre-training and Self-training , 2020, NeurIPS.
[26] Quoc V. Le,et al. Meta Pseudo Labels , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[27] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[28] David Berthelot,et al. ReMixMatch: Semi-Supervised Learning with Distribution Alignment and Augmentation Anchoring , 2019, ArXiv.
[29] Mahmoud Assran,et al. Recovering Petaflops in Contrastive Semi-Supervised Learning of Visual Representations , 2020, ArXiv.
[30] Ce Liu,et al. Supervised Contrastive Learning , 2020, NeurIPS.
[31] Frank Hutter,et al. SGDR: Stochastic Gradient Descent with Warm Restarts , 2016, ICLR.
[32] Francis R. Bach,et al. A convex relaxation for weakly supervised classifiers , 2012, ICML.
[33] Margaret A. Boden. Artificial intelligence and Piagetian theory , 2004, Synthese.
[34] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[35] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[36] Michal Valko,et al. Bootstrap Your Own Latent: A New Approach to Self-Supervised Learning , 2020, NeurIPS.
[37] Xiaojin Zhu,et al. Semi-Supervised Learning , 2010, Encyclopedia of Machine Learning.
[38] Hongyi Zhang,et al. mixup: Beyond Empirical Risk Minimization , 2017, ICLR.
[39] Jacob Jackson,et al. Semi-Supervised Learning by Label Gradient Alignment , 2019, ArXiv.
[40] J. Bruner. INDIVIDUAL AND COLLECTIVE PROBLEMS IN THE STUDY OF THINKING , 1960 .
[41] Oriol Vinyals,et al. Matching Networks for One Shot Learning , 2016, NIPS.
[42] Yoshua Bengio,et al. Entropy Regularization , 2006, Semi-Supervised Learning.
[43] H. J. Scudder,et al. Probability of error of some adaptive pattern-recognition machines , 1965, IEEE Trans. Inf. Theory.
[44] David Berthelot,et al. MixMatch: A Holistic Approach to Semi-Supervised Learning , 2019, NeurIPS.
[45] Iasonas Kokkinos,et al. MultiGrain: a unified image embedding for classes and instances , 2019, ArXiv.
[46] Shin Ishii,et al. Virtual Adversarial Training: A Regularization Method for Supervised and Semi-Supervised Learning , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[47] Xinlei Chen,et al. Exploring Simple Siamese Representation Learning , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[48] Laurens van der Maaten,et al. Self-Supervised Learning of Pretext-Invariant Representations , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[49] Geoffrey E. Hinton,et al. A Simple Framework for Contrastive Learning of Visual Representations , 2020, ICML.
[50] David Berthelot,et al. FixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence , 2020, NeurIPS.