暂无分享,去创建一个
[1] Vladimir Braverman,et al. Adversarial Robustness of Streaming Algorithms through Importance Sampling , 2021, NeurIPS.
[2] Colin Raffel,et al. Imperceptible, Robust, and Targeted Adversarial Examples for Automatic Speech Recognition , 2019, ICML.
[3] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[4] Matthias Hein,et al. Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks , 2020, ICML.
[5] Carl Vondrick,et al. Adversarial Attacks are Reversible with Natural Supervision , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[6] Michael Felsberg,et al. The Visual Object Tracking VOT2017 Challenge Results , 2017, 2017 IEEE International Conference on Computer Vision Workshops (ICCVW).
[7] J. Zico Kolter,et al. Fast is better than free: Revisiting adversarial training , 2020, ICLR.
[8] Rundi Wu,et al. Listening to Sounds of Silence for Speech Denoising , 2020, NeurIPS.
[9] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[10] Ludwig Schmidt,et al. Unlabeled Data Improves Adversarial Robustness , 2019, NeurIPS.
[11] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[12] Baishakhi Ray,et al. Metric Learning for Adversarial Robustness , 2019, NeurIPS.
[13] Erich Elsen,et al. Deep Speech: Scaling up end-to-end speech recognition , 2014, ArXiv.
[14] Bo Chen,et al. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications , 2017, ArXiv.
[15] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[16] Ben Y. Zhao,et al. Wearable Microphone Jamming , 2020, CHI.
[17] Michael I. Jordan,et al. Theoretically Principled Trade-off between Robustness and Accuracy , 2019, ICML.
[18] Dorothea Kolossa,et al. Adversarial Attacks Against Automatic Speech Recognition Systems via Psychoacoustic Hiding , 2018, NDSS.
[19] Hiromu Yakura,et al. Robust Audio Adversarial Example for a Physical Attack , 2018, IJCAI.
[20] Joan Serra,et al. Adversarial Auto-Encoding for Packet Loss Concealment , 2021, 2021 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA).
[21] Christian Poellabauer,et al. Real-Time Adversarial Attacks , 2019, IJCAI.
[22] Chong Wang,et al. Deep Speech 2 : End-to-End Speech Recognition in English and Mandarin , 2015, ICML.
[23] Thomas Wolf,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[24] J. Zico Kolter,et al. Adversarial Music: Real World Audio Adversary Against Wake-word Detection System , 2019, NeurIPS.
[25] James Bailey,et al. Improving Adversarial Robustness Requires Revisiting Misclassified Examples , 2020, ICLR.
[26] Samy Bengio,et al. Adversarial examples in the physical world , 2016, ICLR.
[27] Baishakhi Ray,et al. Multitask Learning Strengthens Adversarial Robustness , 2020, ECCV.
[28] Seyed-Mohsen Moosavi-Dezfooli,et al. DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[29] Liangliang Cao,et al. Exploring Targeted Universal Adversarial Perturbations to End-to-end ASR Models , 2021, Interspeech.
[30] Jason Yosinski,et al. Deep neural networks are easily fooled: High confidence predictions for unrecognizable images , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[31] Nathanael Perraudin,et al. Adversarial Generation of Time-Frequency Features with application in audio synthesis , 2019, ICML.
[32] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[33] Alexei Baevski,et al. wav2vec 2.0: A Framework for Self-Supervised Learning of Speech Representations , 2020, NeurIPS.
[34] E. Poovammal,et al. Adversarial Attack by Inducing Drift in Streaming Data , 2021 .
[35] Jürgen Schmidhuber,et al. Connectionist temporal classification: labelling unsegmented sequence data with recurrent neural networks , 2006, ICML.
[36] Sanjeev Khudanpur,et al. Adversarial Attacks and Defenses for Speech Recognition Systems , 2021, ArXiv.
[37] Shlomo Zilberstein,et al. Using Anytime Algorithms in Intelligent Systems , 1996, AI Mag..
[38] Mingyan Liu,et al. Generating Adversarial Examples with Adversarial Networks , 2018, IJCAI.
[39] Sanjeev Khudanpur,et al. Librispeech: An ASR corpus based on public domain audio books , 2015, 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[40] David A. Wagner,et al. Audio Adversarial Examples: Targeted Attacks on Speech-to-Text , 2018, 2018 IEEE Security and Privacy Workshops (SPW).
[41] Aleksander Madry,et al. Robustness May Be at Odds with Accuracy , 2018, ICLR.
[42] Deva Ramanan,et al. Towards Streaming Perception , 2020, ECCV.
[43] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).