暂无分享,去创建一个
Yann Dauphin | Aaron C. Courville | Aaron Courville | Andrea Frome | Sara Hooker | Yann Dauphin | Andrea Frome | Sara Hooker | Gregory Clark
[1] Welch Bl. THE GENERALIZATION OF ‘STUDENT'S’ PROBLEM WHEN SEVERAL DIFFERENT POPULATION VARLANCES ARE INVOLVED , 1947 .
[2] D. Darling,et al. A Test of Goodness of Fit , 1954 .
[3] Ralph B. D'Agostino,et al. Goodness-of-Fit-Techniques , 2020 .
[4] R. D'Agostino,et al. Goodness-of-Fit-Techniques , 1987 .
[5] Yann LeCun,et al. Optimal Brain Damage , 1989, NIPS.
[6] David E. Rumelhart,et al. Generalization by Weight-Elimination with Application to Forecasting , 1990, NIPS.
[7] Babak Hassibi,et al. Second Order Derivatives for Network Pruning: Optimal Brain Surgeon , 1992, NIPS.
[8] Jianping Zhang,et al. Selecting Typical Instances in Instance-Based Learning , 1992, ML.
[9] Geoffrey E. Hinton,et al. Simplifying Neural Networks by Soft Weight-Sharing , 1992, Neural Computation.
[10] Gregory J. Wolff,et al. Optimal Brain Surgeon and general network pruning , 1993, IEEE International Conference on Neural Networks.
[11] P. Goldman-Rakic,et al. Synaptic development of the cerebral cortex: implications for learning, memory, and mental illness. , 1994, Progress in brain research.
[12] Nikko Ström,et al. Sparse connection and pruning in large dynamic artificial neural networks , 1997, EUROSPEECH.
[13] Rich Caruana,et al. Case-Based Explanation for Artificial Neural Nets , 2000, ANNIMAB.
[14] B. J. Casey,et al. Structural and functional brain development and its relation to cognitive development , 2000, Biological Psychology.
[15] Suzanne E. Welcome,et al. Longitudinal Mapping of Cortical Thickness and Brain Growth in Normal Children , 2022 .
[16] Peter L. Bartlett,et al. Classification with a Reject Option using a Hinge Loss , 2008, J. Mach. Learn. Res..
[17] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[18] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[19] Vincent Vanhoucke,et al. Improving the speed of neural networks on CPUs , 2011 .
[20] C. Huber-Carol. Goodness-of-Fit Tests and Model Validity , 2012 .
[21] Yoshua Bengio,et al. Training deep neural networks with low precision multiplications , 2014 .
[22] Pushmeet Kohli,et al. Memory Bounded Deep Convolutional Networks , 2014, ArXiv.
[23] Dragomir Anguelov,et al. Capturing Long-Tail Distributions of Object Subcategories , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.
[24] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[25] Song Han,et al. Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.
[26] Jason Yosinski,et al. Deep neural networks are easily fooled: High confidence predictions for unrecognizable images , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[27] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[28] Xiaogang Wang,et al. Deep Learning Face Attributes in the Wild , 2014, 2015 IEEE International Conference on Computer Vision (ICCV).
[29] Pritish Narayanan,et al. Deep Learning with Limited Numerical Precision , 2015, ICML.
[30] Mehryar Mohri,et al. Boosting with Abstention , 2016, NIPS.
[31] Maya R. Gupta,et al. Satisfying Real-world Goals with Dataset Constraints , 2016, NIPS.
[32] Rohit Prabhavalkar,et al. On the Efficient Representation and Execution of Deep Acoustic Models , 2016, INTERSPEECH.
[33] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[34] Christopher D. Manning,et al. Compression of Neural Machine Translation Models via Pruning , 2016, CoNLL.
[35] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[36] Joel Emer,et al. Eyeriss: a spatial architecture for energy-efficient dataflow for convolutional neural networks , 2016, CARN.
[37] Oluwasanmi Koyejo,et al. Examples are not enough, learn to criticize! Criticism for Interpretability , 2016, NIPS.
[38] Yiran Chen,et al. Learning Structured Sparsity in Deep Neural Networks , 2016, NIPS.
[39] Gu-Yeon Wei,et al. Minerva: Enabling Low-Power, Highly-Accurate Deep Neural Network Accelerators , 2016, 2016 ACM/IEEE 43rd Annual International Symposium on Computer Architecture (ISCA).
[40] Forrest N. Iandola,et al. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <1MB model size , 2016, ArXiv.
[41] Eugenio Culurciello,et al. An Analysis of Deep Neural Network Models for Practical Applications , 2016, ArXiv.
[42] Yurong Chen,et al. Dynamic Network Surgery for Efficient DNNs , 2016, NIPS.
[43] Erich Elsen,et al. Exploring Sparsity in Recurrent Neural Networks , 2017, ICLR.
[44] Bo Chen,et al. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications , 2017, ArXiv.
[45] Alexandra Chouldechova,et al. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments , 2016, Big Data.
[46] Max Welling,et al. Soft Weight-Sharing for Neural Network Compression , 2017, ICLR.
[47] Kevin Gimpel,et al. A Baseline for Detecting Misclassified and Out-of-Distribution Examples in Neural Networks , 2016, ICLR.
[48] Charles Blundell,et al. Simple and Scalable Predictive Uncertainty Estimation using Deep Ensembles , 2016, NIPS.
[49] Zhiqiang Shen,et al. Learning Efficient Convolutional Networks through Network Slimming , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[50] Ran El-Yaniv,et al. Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations , 2016, J. Mach. Learn. Res..
[51] Siegfried Wahl,et al. Leveraging uncertainty information from deep neural networks for disease detection , 2016, Scientific Reports.
[52] Kilian Q. Weinberger,et al. On Calibration of Modern Neural Networks , 2017, ICML.
[53] Saurabh Goyal,et al. Resource-efficient Machine Learning in 2 KB RAM for the Internet of Things , 2017, ICML.
[54] Alex Kendall,et al. What Uncertainties Do We Need in Bayesian Deep Learning for Computer Vision? , 2017, NIPS.
[55] Timnit Gebru,et al. Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification , 2018, FAT.
[56] Changshui Zhang,et al. Sparse DNNs with Improved Adversarial Robustness , 2018, NeurIPS.
[57] Hao Wu,et al. Mixed Precision Training , 2017, ICLR.
[58] Moustapha Cissé,et al. ConvNets and ImageNet Beyond Accuracy: Understanding Mistakes and Uncovering Biases , 2017, ECCV.
[59] Joost van de Weijer,et al. Metric Learning for Novelty and Anomaly Detection , 2018, BMVC.
[60] Suyog Gupta,et al. To prune, or not to prune: exploring the efficacy of pruning for model compression , 2017, ICLR.
[61] Claudio Gentile,et al. Online Learning with Abstention , 2017, ICML.
[62] R. Srikant,et al. Enhancing The Reliability of Out-of-distribution Image Detection in Neural Networks , 2017, ICLR.
[63] Max Welling,et al. Learning Sparse Neural Networks through L0 Regularization , 2017, ICLR.
[64] Bo Chen,et al. Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[65] David B. Paradice,et al. 3D deep learning for detecting pulmonary nodules in CT scans , 2018, J. Am. Medical Informatics Assoc..
[66] Nicholas D. Lane,et al. The Deep (Learning) Transformation of Mobile and Embedded Computing , 2018, Computer.
[67] Lucas Theis,et al. Faster gaze prediction with dense networks and Fisher pruning , 2018, ArXiv.
[68] Elad Eban,et al. MorphNet: Fast & Simple Resource-Constrained Structure Learning of Deep Networks , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[69] Erich Elsen,et al. Efficient Neural Audio Synthesis , 2018, ICML.
[70] Lubomir M. Hadjiiski,et al. Evolutionary pruning of transfer learned deep convolutional neural network for breast cancer diagnosis in digital breast tomosynthesis , 2018, Physics in medicine and biology.
[71] Marcus A. Badgeley,et al. Variable generalization performance of a deep learning model to detect pneumonia in chest radiographs: A cross-sectional study , 2018, PLoS medicine.
[72] Erich Elsen,et al. The State of Sparsity in Deep Neural Networks , 2019, ArXiv.
[73] Thomas G. Dietterich,et al. Benchmarking Neural Network Robustness to Common Corruptions and Perturbations , 2018, ICLR.
[74] D. Erhan,et al. A Benchmark for Interpretability Methods in Deep Neural Networks , 2018, NeurIPS.
[75] Prateek Mittal,et al. Towards Compact and Robust Deep Neural Networks , 2019, ArXiv.
[76] Ryan P. Adams,et al. Non-vacuous Generalization Bounds at the ImageNet Scale: a PAC-Bayesian Compression Approach , 2018, ICLR.
[77] Sherri Rose,et al. Fair regression for health care spending , 2019, Biometrics.
[78] Philip H. S. Torr,et al. SNIP: Single-shot Network Pruning based on Connection Sensitivity , 2018, ICLR.
[79] Andrew McCallum,et al. Energy and Policy Considerations for Deep Learning in NLP , 2019, ACL.
[80] Charu C. Aggarwal,et al. Efficient Data Representation by Selecting Prototypes with Importance Weights , 2017, 2019 IEEE International Conference on Data Mining (ICDM).
[81] Jan Skoglund,et al. LPCNET: Improving Neural Speech Synthesis through Linear Prediction , 2018, ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[82] Vitaly Feldman,et al. Does learning require memorization? a short tale about a long tail , 2019, STOC.
[83] Dan Klein,et al. Train Large, Then Compress: Rethinking Model Size for Efficient Training and Inference of Transformers , 2020, ArXiv.
[84] P. S. Castro,et al. Rigging the Lottery: Making All Tickets Winners , 2019, ICML.
[85] Vitaly Feldman,et al. What Neural Networks Memorize and Why: Discovering the Long Tail via Influence Estimation , 2020, NeurIPS.
[86] Uday Bondhugula,et al. MLIR: A Compiler Infrastructure for the End of Moore's Law , 2020, ArXiv.
[87] Jose Javier Gonzalez Ortiz,et al. What is the State of Neural Network Pruning? , 2020, MLSys.
[88] Dawn Song,et al. Natural Adversarial Examples , 2019, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[89] Benjamin Rosman,et al. Keep the Gradients Flowing: Using Gradient Flow to Study Sparse Network Optimization , 2021, ArXiv.