暂无分享,去创建一个
Quanshi Zhang | Shufan Wang | Liyao Xiang | Nanyang Ye | Ningyi Liao | Quanshi Zhang | Nanyang Ye | Liyao Xiang | Shu-Fan Wang | Ningyi Liao
[1] Pascal Frossard,et al. Analysis of classifiers’ robustness to adversarial perturbations , 2015, Machine Learning.
[2] Max Welling,et al. Learning Sparse Neural Networks through L0 Regularization , 2017, ICLR.
[3] Yiran Chen,et al. Learning Structured Sparsity in Deep Neural Networks , 2016, NIPS.
[4] Dina Katabi,et al. ME-Net: Towards Effective Adversarial Robustness with Matrix Estimation , 2019, ICML.
[5] Rui Xu,et al. When NAS Meets Robustness: In Search of Robust Architectures Against Adversarial Attacks , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[6] Samy Bengio,et al. Adversarial Machine Learning at Scale , 2016, ICLR.
[7] Luyu Wang,et al. Adversarial Robustness of Pruned Neural Networks , 2018 .
[8] Michael Carbin,et al. The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks , 2018, ICLR.
[9] Kilian Q. Weinberger,et al. Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[10] Stanley Osher,et al. ResNets Ensemble via the Feynman-Kac Formalism to Improve Natural and Robust Accuracies , 2018, NeurIPS.
[11] Song Han,et al. Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.
[12] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[13] Yann LeCun,et al. Optimal Brain Damage , 1989, NIPS.
[14] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[15] Yiren Zhao,et al. To compress or not to compress: Understanding the Interactions between Adversarial Attacks and Neural Network Compression , 2018, SysML.
[16] Jianxin Wu,et al. ThiNet: A Filter Level Pruning Method for Deep Neural Network Compression , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[17] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[18] Kamyar Azizzadenesheli,et al. Stochastic Activation Pruning for Robust Adversarial Defense , 2018, ICLR.
[19] Hanan Samet,et al. Pruning Filters for Efficient ConvNets , 2016, ICLR.
[20] Micah Goldblum,et al. Adversarially Robust Distillation , 2019, AAAI.
[21] Antonio Criminisi,et al. Measuring Neural Net Robustness with Constraints , 2016, NIPS.
[22] Cho-Jui Hsieh,et al. A Convex Relaxation Barrier to Tight Robustness Verification of Neural Networks , 2019, NeurIPS.
[23] Zhiqiang Shen,et al. Learning Efficient Convolutional Networks through Network Slimming , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[24] Li Yang,et al. Robust Sparse Regularization: Simultaneously Optimizing Neural Network Robustness and Compactness , 2019, ArXiv.
[25] Federico Zaiter,et al. The Search for Sparse, Robust Neural Networks , 2019, ArXiv.
[26] Jinwoo Shin,et al. Adversarial Neural Pruning with Latent Vulnerability Suppression , 2020, ICML.
[27] Suman Jana,et al. Towards Practical Lottery Ticket Hypothesis for Adversarial Training , 2020, ArXiv.
[28] Babak Hassibi,et al. Second Order Derivatives for Network Pruning: Optimal Brain Surgeon , 1992, NIPS.
[29] Matthias Hein,et al. Formal Guarantees on the Robustness of a Classifier against Adversarial Manipulation , 2017, NIPS.
[30] Shie Mannor,et al. Robustness and generalization , 2010, Machine Learning.
[31] Mingjie Sun,et al. Rethinking the Value of Network Pruning , 2018, ICLR.
[32] Zhanxing Zhu,et al. Interpreting Adversarially Trained Convolutional Neural Networks , 2019, ICML.
[33] Changshui Zhang,et al. Sparse DNNs with Improved Adversarial Robustness , 2018, NeurIPS.
[34] Prateek Mittal,et al. Towards Compact and Robust Deep Neural Networks , 2019, ArXiv.
[35] Song Han,et al. Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.
[36] James Zijun Wang,et al. Rethinking the Smaller-Norm-Less-Informative Assumption in Channel Pruning of Convolution Layers , 2018, ICLR.
[37] Andrea L. Bertozzi,et al. Sparsity Meets Robustness: Channel Pruning for the Feynman-Kac Formalism Principled Robust Deep Neural Nets , 2020, LOD.
[38] Hao Cheng,et al. Adversarial Robustness vs. Model Compression, or Both? , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[39] Michael I. Jordan,et al. Theoretically Principled Trade-off between Robustness and Accuracy , 2019, ICML.
[40] Zhangyang Wang,et al. Adversarially Trained Model Compression: When Robustness Meets Efficiency , 2019, ArXiv.