暂无分享,去创建一个
[1] Kaiming He,et al. Momentum Contrast for Unsupervised Visual Representation Learning , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[2] John Duchi,et al. Understanding and Mitigating the Tradeoff Between Robustness and Accuracy , 2020, ICML.
[3] Avrim Blum,et al. The Bottleneck , 2021, Monopsony Capitalism.
[4] Roi Livni,et al. Effective Semisupervised Learning on Manifolds , 2017, COLT.
[5] Koby Crammer,et al. A theory of learning from different domains , 2010, Machine Learning.
[6] Ruth Urner,et al. Probabilistic Lipschitzness A niceness assumption for deterministic labels , 2013 .
[7] DeLiang Wang,et al. Unsupervised Learning: Foundations of Neural Computation , 2001, AI Mag..
[8] Yi Zhang,et al. Stronger generalization bounds for deep nets via a compression approach , 2018, ICML.
[9] Laurens van der Maaten,et al. Self-Supervised Learning of Pretext-Invariant Representations , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[10] Quoc V. Le,et al. Self-Training With Noisy Student Improves ImageNet Classification , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[11] Nikos Komodakis,et al. Unsupervised Representation Learning by Predicting Image Rotations , 2018, ICLR.
[12] Geoffrey French,et al. Self-ensembling for visual domain adaptation , 2017, ICLR.
[13] Shai Ben-David,et al. Does Unlabeled Data Provably Help? Worst-case Analysis of the Sample Complexity of Semi-Supervised Learning , 2008, COLT.
[14] Bo Wang,et al. Deep Co-Training for Semi-Supervised Image Recognition , 2018, ECCV.
[15] Xiaojin Zhu,et al. Semi-Supervised Learning , 2010, Encyclopedia of Machine Learning.
[16] J. Cheeger. A lower bound for the smallest eigenvalue of the Laplacian , 1969 .
[17] Kan Chen,et al. Billion-scale semi-supervised learning for image classification , 2019, ArXiv.
[18] Ran El-Yaniv,et al. Error Bounds for Transductive Learning via Compression and Clustering , 2003, NIPS.
[19] Alexei A. Efros,et al. Unsupervised Visual Representation Learning by Context Prediction , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[20] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[21] Max Welling,et al. Semi-supervised Learning with Deep Generative Models , 2014, NIPS.
[22] Prasad Raghavendra,et al. Graph expansion and the unique games conjecture , 2010, STOC '10.
[23] S. Bobkov. Isoperimetric and Analytic Inequalities for Log-Concave Probability Measures , 1999 .
[24] Yuchen Zhang,et al. Bridging Theory and Algorithm for Domain Adaptation , 2019, ICML.
[25] Quoc V. Le,et al. Unsupervised Data Augmentation for Consistency Training , 2019, NeurIPS.
[26] Geoffrey E. Hinton,et al. A Simple Framework for Contrastive Learning of Visual Representations , 2020, ICML.
[27] Harri Valpola,et al. Weight-averaged consistency targets improve semi-supervised deep learning results , 2017, ArXiv.
[28] Tatsuya Harada,et al. Asymmetric Tri-training for Unsupervised Domain Adaptation , 2017, ICML.
[29] Alexei A. Efros,et al. Context Encoders: Feature Learning by Inpainting , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[30] Maria-Florina Balcan,et al. Co-Training and Expansion: Towards Bridging Theory and Practice , 2004, NIPS.
[31] Philippe Rigollet,et al. Generalization Error Bounds in Semi-supervised Classification Under the Cluster Assumption , 2006, J. Mach. Learn. Res..
[32] J. Zico Kolter,et al. Deterministic PAC-Bayesian generalization bounds for deep networks via generalizing noise-resilience , 2019, ICLR.
[33] Michal Valko,et al. Bootstrap Your Own Latent: A New Approach to Self-Supervised Learning , 2020, NeurIPS.
[34] Mikhail Khodak,et al. A Theoretical Analysis of Contrastive Unsupervised Representation Learning , 2019, ICML.
[35] Maria-Florina Balcan,et al. A discriminative model for semi-supervised learning , 2010, J. ACM.
[36] Jeff Donahue,et al. Large Scale GAN Training for High Fidelity Natural Image Synthesis , 2018, ICLR.
[37] Ludwig Schmidt,et al. Unlabeled Data Improves Adversarial Robustness , 2019, NeurIPS.
[38] François Laviolette,et al. Domain-Adversarial Training of Neural Networks , 2015, J. Mach. Learn. Res..
[39] Kaiming He,et al. Improved Baselines with Momentum Contrastive Learning , 2020, ArXiv.
[40] Hossein Mobahi,et al. Self-Distillation Amplifies Regularization in Hilbert Space , 2020, NeurIPS.
[41] Yoshua Bengio,et al. Semi-supervised Learning by Entropy Minimization , 2004, CAP.
[42] Shin Ishii,et al. Virtual Adversarial Training: A Regularization Method for Supervised and Semi-Supervised Learning , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[43] Matthias Seeger,et al. Learning from Labeled and Unlabeled Data , 2010, Encyclopedia of Machine Learning.
[44] Stefano Ermon,et al. A DIRT-T Approach to Unsupervised Domain Adaptation , 2018, ICLR.
[45] Yue Wu,et al. Demystifying Self-Supervised Learning: An Information-Theoretical Framework , 2020, ArXiv.
[46] Paolo Favaro,et al. Unsupervised Learning of Visual Representations by Solving Jigsaw Puzzles , 2016, ECCV.
[47] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[48] Miklós Simonovits,et al. Isoperimetric problems for convex bodies and a localization lemma , 1995, Discret. Comput. Geom..
[49] Akshay Krishnamurthy,et al. Contrastive estimation reveals topic posterior information to linear models , 2020, J. Mach. Learn. Res..
[50] Dong-Hyun Lee,et al. Pseudo-Label : The Simple and Efficient Semi-Supervised Learning Method for Deep Neural Networks , 2013 .
[51] Colin Wei,et al. Data-dependent Sample Complexity of Deep Neural Networks via Lipschitz Augmentation , 2019, NeurIPS.
[52] Samet Oymak,et al. Statistical and Algorithmic Insights for Semi-supervised Learning with Self-training , 2020, ArXiv.
[53] Yuchen Zhang,et al. A Hitting Time Analysis of Stochastic Gradient Langevin Dynamics , 2017, COLT.
[54] Colin Wei,et al. Self-training Avoids Using Spurious Features Under Domain Shift , 2020, NeurIPS.
[55] Taesung Park,et al. CyCADA: Cycle-Consistent Adversarial Domain Adaptation , 2017, ICML.
[56] Robert D. Nowak,et al. Unlabeled data: Now it helps, now it doesn't , 2008, NIPS.
[57] Oriol Vinyals,et al. Representation Learning with Contrastive Predictive Coding , 2018, ArXiv.
[58] Masashi Sugiyama,et al. Learning Discrete Representations via Information Maximizing Self-Augmented Training , 2017, ICML.
[59] Colin Wei,et al. Improved Sample Complexities for Deep Networks and Robust Classification via an All-Layer Margin , 2019, ArXiv.
[60] Yannis Avrithis,et al. Label Propagation for Deep Semi-Supervised Learning , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[61] Victor S. Lempitsky,et al. Unsupervised Domain Adaptation by Backpropagation , 2014, ICML.
[62] Jason D. Lee,et al. Predicting What You Already Know Helps: Provable Self-Supervised Learning , 2020, ArXiv.
[63] S. Bobkov. An isoperimetric inequality on the discrete cube, and an elementary proof of the isoperimetric inequality in Gauss space , 1997 .
[64] Timo Aila,et al. Temporal Ensembling for Semi-Supervised Learning , 2016, ICLR.
[65] Chen Sun,et al. What makes for good views for contrastive learning , 2020, NeurIPS.
[66] Sanjoy Dasgupta,et al. PAC Generalization Bounds for Co-training , 2001, NIPS.
[67] U. Feige,et al. Spectral Graph Theory , 2015 .
[68] Trevor Darrell,et al. Adversarial Discriminative Domain Adaptation , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[69] Santosh S. Vempala,et al. The geometry of logconcave functions and sampling algorithms , 2007, Random Struct. Algorithms.
[70] Alexander Zien,et al. Semi-Supervised Learning , 2006 .
[71] Xiaofeng Liu,et al. Confidence Regularized Self-Training , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[72] N. Linial,et al. Expander Graphs and their Applications , 2006 .
[73] David Berthelot,et al. FixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence , 2020, NeurIPS.
[74] Max Welling,et al. Semi-Supervised Classification with Graph Convolutional Networks , 2016, ICLR.
[75] Trevor Darrell,et al. Deep Domain Confusion: Maximizing for Domain Invariance , 2014, CVPR 2014.
[76] David Yarowsky,et al. Unsupervised Word Sense Disambiguation Rivaling Supervised Methods , 1995, ACL.
[77] Tengyu Ma,et al. Understanding Self-Training for Gradual Domain Adaptation , 2020, ICML.
[78] Philip S. Yu,et al. Transfer Feature Learning with Joint Distribution Adaptation , 2013, 2013 IEEE International Conference on Computer Vision.
[79] Vladimir N. Vapnik,et al. The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.