暂无分享,去创建一个
Dawn Song | Shouling Ji | Bo Li | Youcheng Sun | Jingyi Wang | Jialuo Chen | Peng Cheng | Tinglan Peng | Xingjun Ma | D. Song | Xingjun Ma | S. Ji | Youcheng Sun | Jingyi Wang | Bo Li | Peng Cheng | Jialuo Chen | Tinglan Peng
[1] Yoav Shoham,et al. The Cost of Training NLP Models: A Concise Overview , 2020, ArXiv.
[2] Nicolas Papernot,et al. Entangled Watermarks as a Defense against Model Extraction , 2020, USENIX Security Symposium.
[3] Flemming Topsøe,et al. Jensen-Shannon divergence and Hilbert space embedding , 2004, International Symposium onInformation Theory, 2004. ISIT 2004. Proceedings..
[4] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[5] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[6] Jason Weston,et al. Natural Language Processing (Almost) from Scratch , 2011, J. Mach. Learn. Res..
[7] Seyed-Mohsen Moosavi-Dezfooli,et al. The Robustness of Deep Networks: A Geometrical Perspective , 2017, IEEE Signal Processing Magazine.
[8] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[9] Mingjie Sun,et al. Rethinking the Value of Network Pruning , 2018, ICLR.
[10] Ananthram Swami,et al. Practical Black-Box Attacks against Machine Learning , 2016, AsiaCCS.
[11] Aleksander Madry,et al. Robustness May Be at Odds with Accuracy , 2018, ICLR.
[12] Fan Zhang,et al. Stealing Machine Learning Models via Prediction APIs , 2016, USENIX Security Symposium.
[13] Brendan Dolan-Gavitt,et al. Fine-Pruning: Defending Against Backdooring Attacks on Deep Neural Networks , 2018, RAID.
[14] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[15] Tao Xiang,et al. The Hidden Vulnerability of Watermarking for Deep Neural Networks , 2020, ArXiv.
[16] Pratyush Maini,et al. Dataset Inference: Ownership Resolution in Machine Learning , 2021, ICLR.
[17] Florian Kerschbaum,et al. Deep Neural Network Fingerprinting by Conferrable Adversarial Examples , 2019, ICLR.
[18] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[19] Benny Pinkas,et al. Turning Your Weakness Into a Strength: Watermarking Deep Neural Networks by Backdooring , 2018, USENIX Security Symposium.
[20] Pete Warden,et al. Speech Commands: A Dataset for Limited-Vocabulary Speech Recognition , 2018, ArXiv.
[21] Michael I. Jordan,et al. Theoretically Principled Trade-off between Robustness and Accuracy , 2019, ICML.
[22] Siddharth Garg,et al. BadNets: Evaluating Backdooring Attacks on Deep Neural Networks , 2019, IEEE Access.
[23] Farinaz Koushanfar,et al. DeepSigns: An End-to-End Watermarking Framework for Ownership Protection of Deep Neural Networks , 2019, ASPLOS.
[24] Junfeng Yang,et al. DeepXplore: Automated Whitebox Testing of Deep Learning Systems , 2017, SOSP.
[25] Andrew Y. Ng,et al. Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .
[26] Alberto Ferreira de Souza,et al. Copycat CNN: Stealing Knowledge by Persuading Confession with Random Non-Labeled Data , 2018, 2018 International Joint Conference on Neural Networks (IJCNN).
[27] James Bailey,et al. Skip Connections Matter: On the Transferability of Adversarial Examples Generated with ResNets , 2020, ICLR.
[28] D. Wu,et al. ES Attack: Model Stealing Against Deep Neural Networks Without Data Hurdles , 2020, IEEE Transactions on Emerging Topics in Computational Intelligence.
[29] Jinyuan Jia,et al. IPGuard: Protecting the Intellectual Property of Deep Neural Networks via Fingerprinting the Classification Boundary , 2019, ArXiv.
[30] Yang Feng,et al. DeepGini: prioritizing massive tests to enhance the robustness of deep neural networks , 2020, ISSTA.
[31] Lixin Fan,et al. Rethinking Deep Neural Network Ownership Verification: Embedding Passports to Defeat Ambiguity Attacks , 2019, NeurIPS.
[32] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[33] Hod Lipson,et al. Understanding Neural Networks Through Deep Visualization , 2015, ArXiv.
[34] David Berthelot,et al. High Accuracy and High Fidelity Extraction of Neural Networks , 2020, USENIX Security Symposium.
[35] Kristina Lerman,et al. A Survey on Bias and Fairness in Machine Learning , 2019, ACM Comput. Surv..
[36] Michael Carbin,et al. Comparing Rewinding and Fine-tuning in Neural Network Pruning , 2019, ICLR.
[37] Hui Wu,et al. Protecting Intellectual Property of Deep Neural Networks with Watermarking , 2018, AsiaCCS.
[38] Shin'ichi Satoh,et al. Embedding Watermarks into Deep Neural Networks , 2017, ICMR.
[39] Ilya Mironov,et al. Cryptanalytic Extraction of Neural Network Models , 2020, CRYPTO.
[40] Ben Y. Zhao,et al. Neural Cleanse: Identifying and Mitigating Backdoor Attacks in Neural Networks , 2019, 2019 IEEE Symposium on Security and Privacy (SP).
[41] Geoffrey E. Hinton,et al. Speech recognition with deep recurrent neural networks , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.
[42] Juho Kim,et al. Kapre: On-GPU Audio Preprocessing Layers for a Quick Implementation of Deep Neural Network Models with Keras , 2017, ArXiv.
[43] Jianxiong Xiao,et al. DeepDriving: Learning Affordance for Direct Perception in Autonomous Driving , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[44] Samuel Marchal,et al. PRADA: Protecting Against DNN Model Stealing Attacks , 2018, 2019 IEEE European Symposium on Security and Privacy (EuroS&P).
[45] Tribhuvanesh Orekondy,et al. Knockoff Nets: Stealing Functionality of Black-Box Models , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[46] Erwan Le Merrer,et al. Adversarial frontier stitching for remote neural network watermarking , 2017, Neural Computing and Applications.