暂无分享,去创建一个
Marc Fischer | Martin T. Vechev | Miklós Z. Horváth | Mark Niklas Müller | Marc Fischer | Miklós Z. Horváth
[1] Hiroshi Inoue. Adaptive Ensemble Prediction for Deep Neural Networks based on Confidence Level , 2019, AISTATS.
[2] Cho-Jui Hsieh,et al. Automatic Perturbation Analysis for Scalable Certified Robustness and Beyond , 2020, NeurIPS.
[3] Aleksander Madry,et al. On Adaptive Attacks to Adversarial Example Defenses , 2020, NeurIPS.
[4] Russ Tedrake,et al. Evaluating Robustness of Neural Networks with Mixed Integer Programming , 2017, ICLR.
[5] Binghui Wang,et al. Certified Robustness of Community Detection against Adversarial Structural Perturbation via Randomized Smoothing , 2020, WWW.
[6] Mykel J. Kochenderfer,et al. Reluplex: An Efficient SMT Solver for Verifying Deep Neural Networks , 2017, CAV.
[7] Ilya P. Razenshteyn,et al. Randomized Smoothing of All Shapes and Sizes , 2020, ICML.
[8] Timothy A. Mann,et al. On the Effectiveness of Interval Bound Propagation for Training Verifiably Robust Models , 2018, ArXiv.
[9] Alexander Levine,et al. Robustness Certificates for Sparse Adversarial Attacks by Randomized Ablation , 2019, AAAI.
[10] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[11] Junfeng Yang,et al. Efficient Formal Safety Analysis of Neural Networks , 2018, NeurIPS.
[12] Aditi Raghunathan,et al. Certified Defenses against Adversarial Examples , 2018, ICLR.
[13] Kagan Tumer,et al. Analysis of decision boundaries in linearly combined neural classifiers , 1996, Pattern Recognit..
[14] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[15] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[16] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[17] Abdul Wasay,et al. More or Less: When and How to Build Convolutional Neural Network Ensembles , 2021, ICLR.
[18] Rüdiger Ehlers,et al. Formal Verification of Piece-Wise Linear Feed-Forward Neural Networks , 2017, ATVA.
[19] J. Zico Kolter,et al. Provable defenses against adversarial examples via the convex outer adversarial polytope , 2017, ICML.
[20] Suman Jana,et al. Certified Robustness to Adversarial Examples with Differential Privacy , 2018, 2019 IEEE Symposium on Security and Privacy (SP).
[21] Matthew Mirman,et al. Differentiable Abstract Interpretation for Provably Robust Neural Networks , 2018, ICML.
[22] J. Zico Kolter,et al. Certified Adversarial Robustness via Randomized Smoothing , 2019, ICML.
[23] Samy Bengio,et al. Adversarial Machine Learning at Scale , 2016, ICLR.
[24] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[25] Mislav Balunovic,et al. Adversarial Training and Provable Defenses: Bridging the Gap , 2020, ICLR.
[26] Kagan Tumer,et al. Error Correlation and Error Reduction in Ensemble Classifiers , 1996, Connect. Sci..
[27] Bin Dong,et al. Enhancing Certified Robustness of Smoothed Classifiers via Weighted Model Ensembling , 2020, ArXiv.
[28] Mingjie Sun,et al. Denoised Smoothing: A Provable Defense for Pretrained Classifiers , 2020, NeurIPS.
[29] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[30] Jinwoo Shin,et al. Consistency Regularization for Certified Robustness of Smoothed Classifiers , 2020, NeurIPS.
[31] Cho-Jui Hsieh,et al. Efficient Neural Network Robustness Certification with General Activation Functions , 2018, NeurIPS.
[32] Josef Kittler,et al. Multiple Classifier Systems: First International Workshop, MCS 2000 Cagliari, Italy, June 21-23, 2000 Proceedings , 2000 .
[33] Maximilian Baader,et al. Certified Defense to Image Transformations via Randomized Smoothing , 2020, NeurIPS.
[34] Maya R. Gupta,et al. Quit When You Can: Efficient Evaluation of Ensembles with Ordering Optimization , 2018, ArXiv.
[35] Timon Gehr,et al. An abstract domain for certifying neural networks , 2019, Proc. ACM Program. Lang..
[36] Fabio Roli,et al. Evasion Attacks against Machine Learning at Test Time , 2013, ECML/PKDD.
[37] Alexander Levine,et al. (De)Randomized Smoothing for Certifiable Defense against Patch Attacks , 2020, NeurIPS.
[38] Martin T. Vechev,et al. PRIMA: Precise and General Neural Network Certification via Multi-Neuron Convex Relaxations , 2021 .
[39] Pushmeet Kohli,et al. A Framework for robustness Certification of Smoothed Classifiers using F-Divergences , 2020, ICLR.
[40] Greg Yang,et al. Provably Robust Deep Learning via Adversarially Trained Smoothed Classifiers , 2019, NeurIPS.
[41] Tommi S. Jaakkola,et al. Tight Certificates of Adversarial Robustness for Randomly Smoothed Classifiers , 2019, NeurIPS.
[42] Swarat Chaudhuri,et al. AI2: Safety and Robustness Certification of Neural Networks with Abstract Interpretation , 2018, 2018 IEEE Symposium on Security and Privacy (SP).
[43] Stephan Günnemann,et al. Efficient Robustness Certificates for Discrete Data: Sparsity-Aware Randomized Smoothing for Graphs, Images and More , 2020, ICML.
[44] Aditi Raghunathan,et al. Semidefinite relaxations for certifying robustness to adversarial examples , 2018, NeurIPS.
[45] Pradeep Ravikumar,et al. MACER: Attack-free and Scalable Robust Training via Maximizing Certified Radius , 2020, ICLR.
[46] Kilian Q. Weinberger,et al. Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[47] Matthew Mirman,et al. Fast and Effective Robustness Certification , 2018, NeurIPS.
[48] Ian Goodfellow,et al. Enabling certification of verification-agnostic networks via memory-efficient semidefinite programming , 2020, NeurIPS.
[49] Luiz Eduardo Soares de Oliveira,et al. Decoupling Direction and Norm for Efficient Gradient-Based L2 Adversarial Attacks and Defenses , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[50] Xingyuan Chen,et al. Dynamic Defense Approach for Adversarial Robustness in Deep Neural Networks via Stochastic Ensemble Smoothed Model , 2021, ArXiv.
[51] Víctor Soto,et al. An urn model for majority voting in classification ensembles , 2016, NIPS.
[52] E. S. Pearson,et al. THE USE OF CONFIDENCE OR FIDUCIAL LIMITS ILLUSTRATED IN THE CASE OF THE BINOMIAL , 1934 .
[53] Oleksandr Makeyev,et al. Neural network with ensembles , 2010, The 2010 International Joint Conference on Neural Networks (IJCNN).
[54] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[55] Inderjit S. Dhillon,et al. Towards Fast Computation of Certified Robustness for ReLU Networks , 2018, ICML.
[56] Jiri Matas,et al. On Combining Classifiers , 1998, IEEE Trans. Pattern Anal. Mach. Intell..
[57] Neil Zhenqiang Gong,et al. PointGuard: Provably Robust 3D Point Cloud Classification , 2021, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).