暂无分享,去创建一个
[1] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[2] Junfeng Yang,et al. DeepXplore: Automated Whitebox Testing of Deep Learning Systems , 2017, SOSP.
[3] Jianjun Zhao,et al. DeepStellar: model-based quantitative analysis of stateful deep learning systems , 2019, ESEC/SIGSOFT FSE.
[4] H. B. Mann,et al. On a Test of Whether one of Two Random Variables is Stochastically Larger than the Other , 1947 .
[5] Mykel J. Kochenderfer,et al. Reluplex: An Efficient SMT Solver for Verifying Deep Neural Networks , 2017, CAV.
[6] Roland Vollgraf,et al. Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.
[7] Daniel Kroening,et al. Concolic Testing for Deep Neural Networks , 2018, 2018 33rd IEEE/ACM International Conference on Automated Software Engineering (ASE).
[8] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[9] Sarfraz Khurshid,et al. DeepRoad: GAN-based Metamorphic Autonomous Driving System Testing , 2018, ArXiv.
[10] Matthew Mirman,et al. Differentiable Abstract Interpretation for Provably Robust Neural Networks , 2018, ICML.
[11] Aleksander Madry,et al. Exploring the Landscape of Spatial Robustness , 2017, ICML.
[12] Lin Tan,et al. CRADLE: Cross-Backend Validation to Detect and Localize Bugs in Deep Learning Libraries , 2019, 2019 IEEE/ACM 41st International Conference on Software Engineering (ICSE).
[13] Kilian Q. Weinberger,et al. On Calibration of Modern Neural Networks , 2017, ICML.
[14] Zoubin Ghahramani,et al. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning , 2015, ICML.
[15] Hyun Oh Song,et al. Parsimonious Black-Box Adversarial Attacks via Efficient Combinatorial Optimization , 2019, ICML.
[16] Geoffrey E. Hinton,et al. Visualizing Data using t-SNE , 2008 .
[17] Aleksander Madry,et al. Robustness May Be at Odds with Accuracy , 2018, ICLR.
[18] Somesh Jha,et al. Attribution-Based Confidence Metric For Deep Neural Networks , 2019, NeurIPS.
[19] J. Zico Kolter,et al. Scaling provable adversarial defenses , 2018, NeurIPS.
[20] Ryan R. Curtin,et al. Detecting Adversarial Samples from Artifacts , 2017, ArXiv.
[21] Sanjit A. Seshia,et al. Formal Specification for Deep Neural Networks , 2018, ATVA.
[22] Hridesh Rajan,et al. A comprehensive study on deep learning bug characteristics , 2019, ESEC/SIGSOFT FSE.
[23] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[24] Pinjia He,et al. Structure-Invariant Testing for Machine Translation , 2019, 2020 IEEE/ACM 42nd International Conference on Software Engineering (ICSE).
[25] Samy Bengio,et al. Adversarial examples in the physical world , 2016, ICLR.
[26] Xiaoxing Ma,et al. Boosting operational DNN testing efficiency through conditioning , 2019, ESEC/SIGSOFT FSE.
[27] Xiaogang Wang,et al. Residual Attention Network for Image Classification , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[28] Aleksander Madry,et al. Adversarial Examples Are Not Bugs, They Are Features , 2019, NeurIPS.
[29] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[30] Mingyan Liu,et al. Spatially Transformed Adversarial Examples , 2018, ICLR.
[31] Paolo Tonella,et al. Misbehaviour Prediction for Autonomous Driving Systems , 2019, 2020 IEEE/ACM 42nd International Conference on Software Engineering (ICSE).
[32] Aleksander Madry,et al. A Rotation and a Translation Suffice: Fooling CNNs with Simple Transformations , 2017, ArXiv.
[33] Lei Ma,et al. DeepGauge: Comprehensive and Multi-Granularity Testing Criteria for Gauging the Robustness of Deep Learning Systems , 2018, ArXiv.
[34] Andrew Y. Ng,et al. Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .
[35] S. Sawilowsky. New Effect Size Rules of Thumb , 2009 .
[36] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[37] Uri Shaham,et al. Understanding Adversarial Training: Increasing Local Stability of Neural Nets through Robust Optimization , 2015, ArXiv.
[38] Mislav Balunovic,et al. Certifying Geometric Robustness of Neural Networks , 2019, NeurIPS.
[39] Maya R. Gupta,et al. To Trust Or Not To Trust A Classifier , 2018, NeurIPS.
[40] Elliot Meyerson,et al. Quantifying Point-Prediction Uncertainty in Neural Networks via Residual Estimation with an I/O Kernel , 2020, ICLR.
[41] Wei Li,et al. DeepBillboard: Systematic Physical-World Testing of Autonomous Driving Systems , 2018, 2020 IEEE/ACM 42nd International Conference on Software Engineering (ICSE).
[42] Wen-Chuan Lee,et al. MODE: automated neural network model debugging via state differential analysis and input selection , 2018, ESEC/SIGSOFT FSE.
[43] Simos Gerasimou,et al. Importance-Driven Deep Learning System Testing , 2020, 2020 IEEE/ACM 42nd International Conference on Software Engineering: Companion Proceedings (ICSE-Companion).
[44] Xin Zhang,et al. End to End Learning for Self-Driving Cars , 2016, ArXiv.
[45] Min Wu,et al. Safety Verification of Deep Neural Networks , 2016, CAV.
[46] Koushik Sen,et al. CUTE: a concolic unit testing engine for C , 2005, ESEC/FSE-13.
[47] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[48] Junfeng Yang,et al. Towards Practical Verification of Machine Learning: The Case of Computer Vision Systems , 2017, ArXiv.
[49] Hao Zhang,et al. Apricot: A Weight-Adaptation Approach to Fixing Deep Learning Models , 2019, 2019 34th IEEE/ACM International Conference on Automated Software Engineering (ASE).
[50] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[51] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[52] Jacob Cohen. Statistical Power Analysis for the Behavioral Sciences , 1969, The SAGE Encyclopedia of Research Design.
[53] Yizheng Chen,et al. MixTrain: Scalable Training of Formally Robust Neural Networks , 2018, ArXiv.
[54] Luca Rigazio,et al. Towards Deep Neural Network Architectures Robust to Adversarial Examples , 2014, ICLR.
[55] W. Bossert,et al. The Measurement of Diversity , 2001 .
[56] Ananthram Swami,et al. Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks , 2015, 2016 IEEE Symposium on Security and Privacy (SP).
[57] Mingyan Liu,et al. Generating Adversarial Examples with Adversarial Networks , 2018, IJCAI.
[58] Rüdiger Ehlers,et al. Formal Verification of Piece-Wise Linear Feed-Forward Neural Networks , 2017, ATVA.
[59] Kevin Smith,et al. Bayesian Uncertainty Estimation for Batch Normalized Deep Networks , 2018, ICML.
[60] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[61] Baishakhi Ray,et al. Metric Learning for Adversarial Robustness , 2019, NeurIPS.
[62] Harald C. Gall,et al. Software Engineering for Machine Learning: A Case Study , 2019, 2019 IEEE/ACM 41st International Conference on Software Engineering: Software Engineering in Practice (ICSE-SEIP).
[63] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[64] James Bailey,et al. Characterizing Adversarial Subspaces Using Local Intrinsic Dimensionality , 2018, ICLR.
[65] Vicente Ordonez,et al. Testing DNN Image Classifiers for Confusion & Bias Errors , 2020, 2020 IEEE/ACM 42nd International Conference on Software Engineering: Companion Proceedings (ICSE-Companion).
[66] Junfeng Yang,et al. Formal Security Analysis of Neural Networks using Symbolic Intervals , 2018, USENIX Security Symposium.
[67] Suman Jana,et al. DeepTest: Automated Testing of Deep-Neural-Network-Driven Autonomous Cars , 2017, 2018 IEEE/ACM 40th International Conference on Software Engineering (ICSE).
[68] Andrew Gordon Wilson,et al. Simple Black-box Adversarial Attacks , 2019, ICML.
[69] Junfeng Yang,et al. Efficient Formal Safety Analysis of Neural Networks , 2018, NeurIPS.
[70] Sameer Singh,et al. Generating Natural Adversarial Examples , 2017, ICLR.
[71] Fanny Yang,et al. Invariance-inducing regularization using worst-case transformations suffices to boost accuracy and spatial robustness , 2019, NeurIPS.
[72] Dan Boneh,et al. Ensemble Adversarial Training: Attacks and Defenses , 2017, ICLR.
[73] Jan Hendrik Metzen,et al. On Detecting Adversarial Perturbations , 2017, ICLR.
[74] Shin Yoo,et al. Guiding Deep Learning System Testing Using Surprise Adequacy , 2018, 2019 IEEE/ACM 41st International Conference on Software Engineering (ICSE).
[75] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[76] Ananthram Swami,et al. The Limitations of Deep Learning in Adversarial Settings , 2015, 2016 IEEE European Symposium on Security and Privacy (EuroS&P).
[77] Xiang Gao,et al. Fuzz Testing based Data Augmentation to Improve Robustness of Deep Neural Networks , 2020, 2020 IEEE/ACM 42nd International Conference on Software Engineering (ICSE).