暂无分享,去创建一个
Hisashi Kashima | Rafael Pinot | Florian Yger | Jamal Atif | Cédric Gouy-Pailler | Alexandre Araujo | Laurent Meunier | H. Kashima | F. Yger | C. Gouy-Pailler | J. Atif | Laurent Meunier | Rafael Pinot | Alexandre Araujo
[1] Igor Vajda,et al. Note on discrimination information and variation (Corresp.) , 1970, IEEE Trans. Inf. Theory.
[2] Seyed-Mohsen Moosavi-Dezfooli,et al. Universal Adversarial Perturbations , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[3] Ananthram Swami,et al. The Limitations of Deep Learning in Adversarial Settings , 2015, 2016 IEEE European Symposium on Security and Privacy (EuroS&P).
[4] Logan Engstrom,et al. Synthesizing Robust Adversarial Examples , 2017, ICML.
[5] Sergey Ioffe,et al. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning , 2016, AAAI.
[6] Suman Jana,et al. Certified Robustness to Adversarial Examples with Differential Privacy , 2018, 2019 IEEE Symposium on Security and Privacy (SP).
[7] Deliang Fan,et al. Parametric Noise Injection: Trainable Randomness to Improve Deep Neural Network Robustness Against Adversarial Attack , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[8] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[9] Rama Chellappa,et al. Defense-GAN: Protecting Classifiers Against Adversarial Attacks Using Generative Models , 2018, ICLR.
[10] C. Villani. Optimal Transport: Old and New , 2008 .
[11] Gustavo L. Gilardoni. On Pinsker's and Vajda's Type Inequalities for Csiszár's $f$ -Divergences , 2006, IEEE Transactions on Information Theory.
[12] Cho-Jui Hsieh,et al. Towards Robust Neural Networks via Random Self-ensemble , 2017, ECCV.
[13] A. Rényi. On Measures of Entropy and Information , 1961 .
[14] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[15] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[16] Moustapha Cissé,et al. Parseval Networks: Improving Robustness to Adversarial Examples , 2017, ICML.
[17] Thomas M. Cover,et al. Elements of Information Theory , 2005 .
[18] Alison L Gibbs,et al. On Choosing and Bounding Probability Metrics , 2002, math/0209021.
[19] Moustapha Cissé,et al. Countering Adversarial Images using Input Transformations , 2018, ICLR.
[20] Seyed-Mohsen Moosavi-Dezfooli,et al. DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[21] Kamyar Azizzadenesheli,et al. Stochastic Activation Pruning for Robust Adversarial Defense , 2018, ICLR.
[22] Jinfeng Yi,et al. EAD: Elastic-Net Attacks to Deep Neural Networks via Adversarial Examples , 2017, AAAI.
[23] Bernhard Schölkopf,et al. Adversarial Vulnerability of Neural Networks Increases With Input Dimension , 2018, ArXiv.
[24] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[25] Ilya P. Razenshteyn,et al. Adversarial examples from computational constraints , 2018, ICML.
[26] Ananthram Swami,et al. Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks , 2015, 2016 IEEE Symposium on Security and Privacy (SP).
[27] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[28] B. Ripley,et al. Robust Statistics , 2018, Encyclopedia of Mathematical Geosciences.
[29] Stefano Soatto,et al. Empirical Study of the Topology and Geometry of Deep Networks , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[30] B. Kosko,et al. Adaptive stochastic resonance , 1998, Proc. IEEE.
[31] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[32] Aleksander Madry,et al. On Evaluating Adversarial Robustness , 2019, ArXiv.
[33] O. Krafft. A note on exponential bounds for binomial probabilities , 1969 .
[34] Renato Renner,et al. An intuitive proof of the data processing inequality , 2011, Quantum Inf. Comput..
[35] Alan L. Yuille,et al. Mitigating adversarial effects through randomization , 2017, ICLR.
[36] Luis Perez,et al. The Effectiveness of Data Augmentation in Image Classification using Deep Learning , 2017, ArXiv.
[37] François Chapeau-Blondeau,et al. Noise-enhanced performance for an optimal Bayesian estimator , 2004, IEEE Transactions on Signal Processing.
[38] Saeed Mahloujifar,et al. Adversarial Risk and Robustness: General Definitions and Implications for the Uniform Distribution , 2018, NeurIPS.
[39] Pierre-Olivier Amblard,et al. Stochastic resonance in discrete time nonlinear AR(1) models , 1999, IEEE Trans. Signal Process..
[40] Bernhard Pfahringer,et al. Regularisation of neural networks by enforcing Lipschitz continuity , 2018, Machine Learning.
[41] Hao Chen,et al. MagNet: A Two-Pronged Defense against Adversarial Examples , 2017, CCS.
[42] Aaron Roth,et al. The Algorithmic Foundations of Differential Privacy , 2014, Found. Trends Theor. Comput. Sci..
[43] Seyed-Mohsen Moosavi-Dezfooli,et al. Robustness of classifiers: from adversarial to random noise , 2016, NIPS.
[44] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[45] Samy Bengio,et al. Adversarial examples in the physical world , 2016, ICLR.
[46] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[47] Motoaki Kawanabe,et al. Machine Learning in Non-Stationary Environments - Introduction to Covariate Shift Adaptation , 2012, Adaptive computation and machine learning.