暂无分享,去创建一个
[1] Beilun Wang,et al. DeepMask: Masking DNN Models for robustness against adversarial samples , 2017, ArXiv.
[2] Yin Yang,et al. Functional Mechanism: Regression Analysis under Differential Privacy , 2012, Proc. VLDB Endow..
[3] Catuscia Palamidessi,et al. Broadening the Scope of Differential Privacy Using Metrics , 2013, Privacy Enhancing Technologies.
[4] Cynthia Dwork,et al. Calibrating Noise to Sensitivity in Private Data Analysis , 2006, TCC.
[5] Yanjun Qi,et al. Feature Squeezing: Detecting Adversarial Examples in Deep Neural Networks , 2017, NDSS.
[6] Jihun Hamm,et al. Crowd-ML: A library for privacy-preserving machine learning on smart devices , 2017, 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[7] Radha Poovendran,et al. Blocking Transferability of Adversarial Examples in Black-Box Learning Systems , 2017, ArXiv.
[8] Samy Bengio,et al. Adversarial Machine Learning at Scale , 2016, ICLR.
[9] Jan Hendrik Metzen,et al. On Detecting Adversarial Perturbations , 2017, ICLR.
[10] Moustapha Cissé,et al. Parseval Networks: Improving Robustness to Adversarial Examples , 2017, ICML.
[11] Dawn Xiaodong Song,et al. Delving into Transferable Adversarial Examples and Black-box Attacks , 2016, ICLR.
[12] Aaron Roth,et al. The Algorithmic Foundations of Differential Privacy , 2014, Found. Trends Theor. Comput. Sci..
[13] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[14] Xiaolin Hu,et al. Discovering Adversarial Examples with Momentum , 2017, ArXiv.
[15] Christian Gagné,et al. Robustness to Adversarial Examples through an Ensemble of Specialists , 2017, ICLR.
[16] Suman Jana,et al. Certified Robustness to Adversarial Examples with Differential Privacy , 2018, 2019 IEEE Symposium on Security and Privacy (SP).
[17] Somesh Jha,et al. Model Inversion Attacks that Exploit Confidence Information and Basic Countermeasures , 2015, CCS.
[18] Patrick D. McDaniel,et al. Extending Defensive Distillation , 2017, ArXiv.
[19] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[20] Patrick D. McDaniel,et al. On the (Statistical) Detection of Adversarial Examples , 2017, ArXiv.
[21] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[22] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[23] Xintao Wu,et al. Regression Model Fitting under Differential Privacy and Model Inversion Attack , 2015, IJCAI.
[24] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[25] Lap-Pui Chau,et al. Margin maximization for robust classification using deep learning , 2017, 2017 International Joint Conference on Neural Networks (IJCNN).
[26] Ananthram Swami,et al. The Limitations of Deep Learning in Adversarial Settings , 2015, 2016 IEEE European Symposium on Security and Privacy (EuroS&P).
[27] Dejing Dou,et al. Preserving differential privacy in convolutional deep belief networks , 2017, Machine Learning.
[28] Dejing Dou,et al. Differential Privacy Preservation for Deep Auto-Encoders: an Application of Human Behavior Prediction , 2016, AAAI.
[29] Kenneth O. Stanley,et al. Mitigating fooling with competitive overcomplete output layer neural networks , 2017, 2017 International Joint Conference on Neural Networks (IJCNN).
[30] Dan Boneh,et al. Ensemble Adversarial Training: Attacks and Defenses , 2017, ICLR.
[31] Michael P. Wellman,et al. Towards the Science of Security and Privacy in Machine Learning , 2016, ArXiv.
[32] Ananthram Swami,et al. Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks , 2015, 2016 IEEE Symposium on Security and Privacy (SP).
[33] Aditi Raghunathan,et al. Certified Defenses against Adversarial Examples , 2018, ICLR.
[34] Vitaly Shmatikov,et al. Membership Inference Attacks Against Machine Learning Models , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[35] Vitaly Shmatikov,et al. Privacy-preserving deep learning , 2015, 2015 53rd Annual Allerton Conference on Communication, Control, and Computing (Allerton).
[36] Yan Zhang,et al. Privacy-Preserving Collaborative Deep Learning with Irregular Participants. , 2018, 1812.10113.
[37] Dejing Dou,et al. Adaptive Laplace Mechanism: Differential Privacy Preservation in Deep Learning , 2017, 2017 IEEE International Conference on Data Mining (ICDM).
[38] Luca Rigazio,et al. Towards Deep Neural Network Architectures Robust to Adversarial Examples , 2014, ICLR.
[39] J. Zico Kolter,et al. Provable defenses against adversarial examples via the convex outer adversarial polytope , 2017, ICML.
[40] Samy Bengio,et al. Adversarial examples in the physical world , 2016, ICLR.