暂无分享,去创建一个
Stefano Soatto | Avinash Ravichandran | Alessandro Achille | Marzia Polito | Aditya Golatkar | Stefano Soatto | A. Achille | Avinash Ravichandran | M. Polito | Aditya Golatkar
[1] Daniel S. Kermany,et al. Identifying Medical Diagnoses and Treatable Diseases by Image-Based Deep Learning , 2018, Cell.
[2] George R. Thoma,et al. Pre-trained convolutional neural networks as feature extractors toward improved malaria parasite detection in thin blood smear images , 2018, PeerJ.
[3] Hermann Ney,et al. Cross-entropy vs. squared error training: a theoretical and experimental comparison , 2013, INTERSPEECH.
[4] Joachim Denzler,et al. Deep Learning on Small Datasets without Pre-Training using Cosine Loss , 2019, 2020 IEEE Winter Conference on Applications of Computer Vision (WACV).
[5] Roger B. Grosse,et al. Optimizing Neural Networks with Kronecker-factored Approximate Curvature , 2015, ICML.
[6] L. V. D. Maaten,et al. Certified Data Removal from Machine Learning Models , 2019, ICML.
[7] Honglak Lee,et al. What's in a Loss Function for Image Classification? , 2020, ArXiv.
[8] Pietro Perona,et al. Caltech-UCSD Birds 200 , 2010 .
[9] Subhransu Maji,et al. Improved Bilinear Pooling with CNNs , 2017, BMVC.
[10] Cynthia Dwork,et al. Differential Privacy: A Survey of Results , 2008, TAMC.
[11] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[12] Michael L. Littman,et al. Online Linear Regression and Its Application to Model-Based Reinforcement Learning , 2007, NIPS.
[13] Yurii Nesterov,et al. Introductory Lectures on Convex Optimization - A Basic Course , 2014, Applied Optimization.
[14] Stefano Soatto,et al. Time Matters in Regularizing Deep Networks: Weight Decay and Data Augmentation Affect Early Learning Dynamics, Matter Little Near Convergence , 2019, NeurIPS.
[15] Xiaogang Wang,et al. Deeply learned face representations are sparse, selective, and robust , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[16] Stefano Soatto,et al. Eternal Sunshine of the Spotless Net: Selective Forgetting in Deep Networks , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[17] Andrew Zisserman,et al. A Visual Vocabulary for Flower Classification , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).
[18] Huan Wang,et al. Taylorized Training: Towards Better Approximation of Neural Network Training at Finite Width , 2020, ArXiv.
[19] 俊一 甘利. 5分で分かる!? 有名論文ナナメ読み:Jacot, Arthor, Gabriel, Franck and Hongler, Clement : Neural Tangent Kernel : Convergence and Generalization in Neural Networks , 2020 .
[20] Stefano Soatto,et al. Estimating informativeness of samples with Smooth Unique Information , 2021, ICLR.
[21] Andrew L. Maas. Rectifier Nonlinearities Improve Neural Network Acoustic Models , 2013 .
[22] Pascal Vincent,et al. Fast Approximate Natural Gradient Descent in a Kronecker-factored Eigenbasis , 2018, NeurIPS.
[23] Jaehoon Lee,et al. Wide neural networks of any depth evolve as linear models under gradient descent , 2019, NeurIPS.
[24] Yingyu Liang,et al. Gradients as Features for Deep Representation Learning , 2020, ICLR.
[25] Stefano Soatto,et al. Rethinking the Hyperparameters for Fine-tuning , 2020, ICLR.
[26] C. V. Jawahar,et al. Cats and dogs , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.
[27] Subhransu Maji,et al. Task2Vec: Task Embedding for Meta-Learning , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[28] I. Csiszár. Why least squares and maximum entropy? An axiomatic approach to inference for linear inverse problems , 1991 .
[29] Geoffrey E. Hinton,et al. Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.
[30] Subhransu Maji,et al. Bilinear CNN Models for Fine-Grained Visual Recognition , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[31] Mikhail Belkin,et al. Evaluation of Neural Architectures Trained with Square Loss vs Cross-Entropy in Classification Tasks , 2020, ICLR.
[32] Stefano Soatto,et al. Towards Backward-Compatible Representation Learning , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[33] Samyadeep Basu,et al. Influence Functions in Deep Learning Are Fragile , 2020, ICLR.
[35] Tal Hassner,et al. LEEP: A New Measure to Evaluate Transferability of Learned Representations , 2020, ICML.
[36] Yang Song,et al. Large Scale Fine-Grained Categorization and Domain-Specific Transfer Learning , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[37] Razvan Pascanu,et al. Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.
[38] Roger B. Grosse,et al. A Kronecker-factored approximate Fisher matrix for convolution layers , 2016, ICML.
[39] Percy Liang,et al. Understanding Black-box Predictions via Influence Functions , 2017, ICML.
[40] Jonathan Krause,et al. 3D Object Representations for Fine-Grained Categorization , 2013, 2013 IEEE International Conference on Computer Vision Workshops.
[41] Michael I. Jordan,et al. A Swiss Army Infinitesimal Jackknife , 2018, AISTATS.
[42] Antonio Torralba,et al. Recognizing indoor scenes , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.
[43] Subhransu Maji,et al. Fine-Grained Visual Classification of Aircraft , 2013, ArXiv.
[44] Xuhong Li,et al. Explicit Inductive Bias for Transfer Learning with Convolutional Networks , 2018, ICML.