Fast Adaptation with Linearized Neural Networks
暂无分享,去创建一个
Andrew Gordon Wilson | Andreas Damianou | Wesley J. Maddox | Pablo Garcia Moreno | Pablo G. Moreno | Shuai Tang | A. Damianou | Shuai Tang | A. Wilson
[1] Kilian Q. Weinberger,et al. Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[2] David J. C. MacKay,et al. Information Theory, Inference, and Learning Algorithms , 2004, IEEE Transactions on Information Theory.
[3] Carlos Guestrin,et al. Adversarial Fisher Vectors for Unsupervised Representation Learning , 2019, NeurIPS.
[4] Ruosong Wang,et al. Enhanced Convolutional Neural Tangent Kernels , 2019, ArXiv.
[5] Kilian Q. Weinberger,et al. On Calibration of Modern Neural Networks , 2017, ICML.
[6] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[7] Andrew Gordon Wilson,et al. GPyTorch: Blackbox Matrix-Matrix Gaussian Process Inference with GPU Acceleration , 2018, NeurIPS.
[8] David J. C. MacKay,et al. Bayesian Interpolation , 1992, Neural Computation.
[9] Christopher K. I. Williams,et al. Gaussian Processes for Machine Learning (Adaptive Computation and Machine Learning) , 2005 .
[10] Alexander Immer,et al. Improving predictions of Bayesian neural networks via local linearization , 2020, ArXiv.
[11] Ruosong Wang,et al. On Exact Computation with an Infinitely Wide Neural Net , 2019, NeurIPS.
[12] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[13] Hod Lipson,et al. Convergent Learning: Do different neural networks learn the same representations? , 2015, FE@NIPS.
[14] Matthias W. Seeger,et al. Covariance Kernels from Bayesian Generative Models , 2001, NIPS.
[15] Yoshua Bengio,et al. How transferable are features in deep neural networks? , 2014, NIPS.
[16] Dale Schuurmans,et al. Holographic Feature Representations of Deep Networks , 2017, UAI.
[17] 俊一 甘利. 5分で分かる!? 有名論文ナナメ読み:Jacot, Arthor, Gabriel, Franck and Hongler, Clement : Neural Tangent Kernel : Convergence and Generalization in Neural Networks , 2020 .
[18] Andrew Gordon Wilson,et al. Constant-Time Predictive Distributions for Gaussian Processes , 2018, ICML.
[19] David Haussler,et al. Exploiting Generative Models in Discriminative Classifiers , 1998, NIPS.
[20] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[21] Stefan Carlsson,et al. CNN Features Off-the-Shelf: An Astounding Baseline for Recognition , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition Workshops.
[22] Yoshua Bengio,et al. Deep Learning of Representations for Unsupervised and Transfer Learning , 2011, ICML Unsupervised and Transfer Learning.
[23] Daniel R. Jiang,et al. BoTorch: Programmable Bayesian Optimization in PyTorch , 2019, ArXiv.
[24] Jaehoon Lee,et al. Wide neural networks of any depth evolve as linear models under gradient descent , 2019, NeurIPS.
[25] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[26] Neil D. Lawrence,et al. Metrics for Probabilistic Geometries , 2014, UAI.
[27] Yoshua Bengio,et al. Bayesian Model-Agnostic Meta-Learning , 2018, NeurIPS.
[28] Gene H. Golub,et al. The differentiation of pseudo-inverses and non-linear least squares problems whose variables separate , 1972, Milestones in Matrix Computation.
[29] Andrew Gordon Wilson,et al. Deep Kernel Learning , 2015, AISTATS.
[30] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[31] Matthias W. Seeger,et al. Scalable Hyperparameter Transfer Learning , 2018, NeurIPS.
[32] Neil D. Lawrence,et al. Kernels for Vector-Valued Functions: a Review , 2011, Found. Trends Mach. Learn..
[33] Mohammad Emtiyaz Khan,et al. Continual Deep Learning by Functional Regularisation of Memorable Past , 2020, NeurIPS.
[34] Trevor Darrell,et al. Adversarial Feature Learning , 2016, ICLR.
[35] Mohammad Emtiyaz Khan,et al. Approximate Inference Turns Deep Networks into Gaussian Processes , 2019, NeurIPS.
[36] Alexandre Lacoste,et al. Adaptive Deep Kernel Learning , 2019, ArXiv.
[37] Yousef Saad,et al. Iterative methods for sparse linear systems , 2003 .
[38] Carl E. Rasmussen,et al. Gaussian processes for machine learning , 2005, Adaptive computation and machine learning.
[39] Andrew Y. Ng,et al. Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .
[40] Chong Wang,et al. Stochastic variational inference , 2012, J. Mach. Learn. Res..
[41] Alessandra Tosi,et al. Visualization and interpretability in probabilistic dimensionality reduction models , 2014 .
[42] Subhransu Maji,et al. Task2Vec: Task Embedding for Meta-Learning , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[43] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[44] Greg Yang,et al. Scaling Limits of Wide Neural Networks with Weight Sharing: Gaussian Process Behavior, Gradient Independence, and Neural Tangent Kernel Derivation , 2019, ArXiv.
[45] Yingyu Liang,et al. Gradients as Features for Deep Representation Learning , 2020, ICLR.
[46] Elliot J. Crowley,et al. Deep Kernel Transfer in Gaussian Processes for Few-shot Learning , 2019, ArXiv.