暂无分享,去创建一个
[1] Bistra N. Dilkina,et al. Combinatorial Attacks on Binarized Neural Networks , 2019, ICLR.
[2] Yoshua Bengio,et al. Hierarchical Multiscale Recurrent Neural Networks , 2016, ICLR.
[3] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[4] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[5] Logan Engstrom,et al. Synthesizing Robust Adversarial Examples , 2017, ICML.
[6] Ananthram Swami,et al. The Limitations of Deep Learning in Adversarial Settings , 2015, 2016 IEEE European Symposium on Security and Privacy (EuroS&P).
[7] Yuval Elovici,et al. User Authentication Based on Mouse Dynamics Using Deep Neural Networks: A Comprehensive Study , 2020, IEEE Transactions on Information Forensics and Security.
[8] Max Welling,et al. Rotation Equivariant CNNs for Digital Pathology , 2018, MICCAI.
[9] Mingyuan Zhou,et al. ARM: Augment-REINFORCE-Merge Gradient for Stochastic Binary Networks , 2018, ICLR.
[10] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[11] Matthew Cook,et al. Unsupervised learning of digit recognition using spike-timing-dependent plasticity , 2015, Front. Comput. Neurosci..
[12] J Gautrais,et al. Rate coding versus temporal order coding: a theoretical approach. , 1998, Bio Systems.
[13] Luca Antiga,et al. Automatic differentiation in PyTorch , 2017 .
[14] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[15] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[16] Terrence J. Sejnowski,et al. Gradient Descent for Spiking Neural Networks , 2017, NeurIPS.
[17] Seyed-Mohsen Moosavi-Dezfooli,et al. DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[18] Medhat A. Moussa,et al. Attacking Binarized Neural Networks , 2017, ICLR.
[19] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[20] Tapani Raiko,et al. Techniques for Learning Binary Stochastic Feedforward Neural Networks , 2014, ICLR.
[21] Narasimhan Sundararajan,et al. A novel method for extracting interpretable knowledge from a spiking neural classifier with time-varying synaptic weights , 2019, ArXiv.
[22] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[23] Priyadarshini Panda,et al. A Comprehensive Analysis on Adversarial Robustness of Spiking Neural Networks , 2019, 2019 International Joint Conference on Neural Networks (IJCNN).
[24] Ananthram Swami,et al. Practical Black-Box Attacks against Machine Learning , 2016, AsiaCCS.
[25] Dawn Xiaodong Song,et al. Adversarial Example Defenses: Ensembles of Weak Defenses are not Strong , 2017, ArXiv.
[26] Kaushik Roy,et al. Enabling Spike-Based Backpropagation for Training Deep Neural Network Architectures , 2019, Frontiers in Neuroscience.
[27] Kaushik Roy,et al. Going Deeper in Spiking Neural Networks: VGG and Residual Architectures , 2018, Front. Neurosci..
[28] Ran El-Yaniv,et al. Binarized Neural Networks , 2016, ArXiv.
[29] Sridhar Adepu,et al. Anomaly Detection in Cyber Physical Systems Using Recurrent Neural Networks , 2017, 2017 IEEE 18th International Symposium on High Assurance Systems Engineering (HASE).
[30] S. Thorpe,et al. STDP-based spiking deep convolutional neural networks for object recognition , 2018 .
[31] Matthias Bethge,et al. Decision-Based Adversarial Attacks: Reliable Attacks Against Black-Box Machine Learning Models , 2017, ICLR.
[32] David Wagner,et al. Adversarial Examples Are Not Easily Detected: Bypassing Ten Detection Methods , 2017, AISec@CCS.
[33] Seyed-Mohsen Moosavi-Dezfooli,et al. SparseFool: A Few Pixels Make a Big Difference , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[34] Geoffrey E. Hinton,et al. Visualizing Data using t-SNE , 2008 .
[35] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[36] Sander M. Bohte,et al. Error-backpropagation in temporally encoded networks of spiking neurons , 2000, Neurocomputing.
[37] Jun Zhu,et al. Improving Black-box Adversarial Attacks with a Transfer-based Prior , 2019, NeurIPS.
[38] Dawn Xiaodong Song,et al. Delving into Transferable Adversarial Examples and Black-box Attacks , 2016, ICLR.
[39] Yoshua Bengio,et al. Estimating or Propagating Gradients Through Stochastic Neurons for Conditional Computation , 2013, ArXiv.
[40] Matthias Bethge,et al. Foolbox v0.8.0: A Python toolbox to benchmark the robustness of machine learning models , 2017, ArXiv.
[41] Abbas Nowzari-Dalini,et al. SpykeTorch: Efficient Simulation of Convolutional Spiking Neural Networks With at Most One Spike per Neuron , 2019, Front. Neurosci..
[42] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[43] Yuval Elovici,et al. Adversarial Attacks on Remote User Authentication Using Behavioural Mouse Dynamics , 2019, 2019 International Joint Conference on Neural Networks (IJCNN).
[44] Lior Rokach,et al. Low Resource Black-Box End-to-End Attack Against State of the Art API Call Based Malware Classifiers , 2018, ArXiv.
[45] Ananthram Swami,et al. Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks , 2015, 2016 IEEE Symposium on Security and Privacy (SP).
[46] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[47] W. Brendel,et al. Foolbox: A Python toolbox to benchmark the robustness of machine learning models , 2017 .
[48] Samy Bengio,et al. Adversarial examples in the physical world , 2016, ICLR.
[49] Roland Memisevic,et al. How far can we go without convolution: Improving fully-connected networks , 2015, ArXiv.
[50] Abbas Nowzari-Dalini,et al. Bio-inspired digit recognition using reward-modulated spike-timing-dependent plasticity in deep convolutional networks , 2019, Pattern Recognit..