暂无分享,去创建一个
[1] Ji Xu,et al. On the Optimal Weighted $\ell_2$ Regularization in Overparameterized Linear Regression , 2020, NeurIPS.
[2] Andrea Montanari,et al. The generalization error of max-margin linear classifiers: High-dimensional asymptotics in the overparametrized regime , 2019 .
[3] Rich Caruana,et al. Multitask Learning , 1998, Encyclopedia of Machine Learning and Data Mining.
[4] Roman Vershynin,et al. Introduction to the non-asymptotic analysis of random matrices , 2010, Compressed Sensing.
[5] Philipp Koehn,et al. Findings of the 2014 Workshop on Statistical Machine Translation , 2014, WMT@ACL.
[6] Oriol Vinyals,et al. Matching Networks for One Shot Learning , 2016, NIPS.
[7] Christos Thrampoulidis,et al. Provable Benefits of Overparameterization in Model Compression: From Double Descent to Pruning Neural Networks , 2020, AAAI.
[8] Andrea Montanari,et al. Surprises in High-Dimensional Ridgeless Least Squares Interpolation , 2019, Annals of statistics.
[9] Samet Oymak,et al. Generalization Guarantees for Neural Networks via Harnessing the Low-rank Structure of the Jacobian , 2019, ArXiv.
[10] Zhao Song,et al. Learning mixtures of linear regressions in subexponential time via Fourier moments , 2019, STOC.
[11] Mikhail Belkin,et al. Classification vs regression in overparameterized regimes: Does the loss function matter? , 2020, J. Mach. Learn. Res..
[12] Fei-Fei Li,et al. ImageNet: A large-scale hierarchical image database , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.
[13] Andrea Montanari,et al. The Generalization Error of Random Features Regression: Precise Asymptotics and the Double Descent Curve , 2019, Communications on Pure and Applied Mathematics.
[14] Weihao Kong,et al. Robust Meta-learning for Mixed Linear Regression with Small Batches , 2020, NeurIPS.
[15] Santosh S. Vempala,et al. Efficient Representations for Lifelong Learning and Autoencoding , 2014, COLT.
[16] Claudio Gentile,et al. Linear Algorithms for Online Multitask Classification , 2010, COLT.
[17] Mikhail Belkin,et al. Reconciling modern machine-learning practice and the classical bias–variance trade-off , 2018, Proceedings of the National Academy of Sciences.
[18] W. Kahan,et al. The Rotation of Eigenvectors by a Perturbation. III , 1970 .
[19] Jonathan Baxter,et al. A Model of Inductive Bias Learning , 2000, J. Artif. Intell. Res..
[20] Tengyu Ma,et al. Optimal Regularization Can Mitigate Double Descent , 2021, ICLR.
[21] S. Geer,et al. Oracle Inequalities and Optimal Inference under Group Sparsity , 2010, 1007.1771.
[22] Wei Hu,et al. Provable Benefits of Representation Learning in Linear Bandits , 2020, ArXiv.
[23] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[24] Sham M. Kakade,et al. Few-Shot Learning via Learning the Representation, Provably , 2020, ICLR.
[25] Toniann Pitassi,et al. Theoretical bounds on estimation error for meta-learning , 2020, ArXiv.
[26] Philip M. Long,et al. Benign overfitting in linear regression , 2019, Proceedings of the National Academy of Sciences.
[27] 俊一 甘利. 5分で分かる!? 有名論文ナナメ読み:Jacot, Arthor, Gabriel, Franck and Hongler, Clement : Neural Tangent Kernel : Convergence and Generalization in Neural Networks , 2020 .
[28] Massimiliano Pontil,et al. The Benefit of Multitask Representation Learning , 2015, J. Mach. Learn. Res..
[29] Jaehoon Lee,et al. Wide neural networks of any depth evolve as linear models under gradient descent , 2019, NeurIPS.
[30] Francis Bach,et al. On Lazy Training in Differentiable Programming , 2018, NeurIPS.
[31] Simon S. Du,et al. Impact of Representation Learning in Linear Bandits , 2020, ICLR.
[32] Yuanzhi Li,et al. Learning Mixtures of Linear Regressions with Nearly Optimal Complexity , 2018, COLT.
[33] Inderjit S. Dhillon,et al. Mixed Linear Regression with Multiple Components , 2016, NIPS.
[34] Barnabás Póczos,et al. Gradient Descent Provably Optimizes Over-parameterized Neural Networks , 2018, ICLR.
[35] Weihao Kong,et al. Meta-learning for mixed linear regression , 2020, ICML.
[36] Michael I. Jordan,et al. Provable Meta-Learning of Linear Representations , 2020, ICML.
[37] Chandler Davis. The rotation of eigenvectors by a perturbation , 1963 .
[38] Samet Oymak,et al. Sample Efficient Subspace-Based Representations for Nonlinear Meta-Learning , 2021, ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[39] Anant Sahai,et al. Harmless interpolation of noisy data in regression , 2019, 2019 IEEE International Symposium on Information Theory (ISIT).
[40] Massimiliano Pontil,et al. Convex multi-task feature learning , 2008, Machine Learning.
[41] Christos Thrampoulidis,et al. LASSO with Non-linear Measurements is Equivalent to One With Linear Measurements , 2015, NIPS.