BioGD: Bio-inspired robust gradient descent

Recent research in machine learning pointed to the core problem of state-of-the-art models which impedes their widespread adoption in different domains. The models’ inability to differentiate between noise and subtle, yet significant variation in data leads to their vulnerability to adversarial perturbations that cause wrong predictions with high confidence. The study is aimed at identifying whether the algorithms inspired by biological evolution may achieve better results in cases where brittle robustness properties are highly sensitive to the slight noise. To answer this question, we introduce the new robust gradient descent inspired by the stability and adaptability of biological systems to unknown and changing environments. The proposed optimization technique involves an open-ended adaptation process with regard to two hyperparameters inherited from the generalized Verhulst population growth equation. The hyperparameters increase robustness to adversarial noise by penalizing the degree to which hardly visible changes in gradients impact prediction. The empirical evidence on synthetic and experimental datasets confirmed the viability of the bio-inspired gradient descent and suggested promising directions for future research. The code used for computational experiments is provided in a repository at https://github.com/yukinoi/bio_gradient_descent.

[1]  Animal Aggregations. , 1931, Nature.

[2]  Sarah Webb Deep learning for biology , 2018, Nature.

[3]  T. Malthus An essay on the principle of population, as it affects the future improvement of society , 2006 .

[4]  Chenchen Liu,et al.  Towards Robust Training of Neural Networks by Regularizing Adversarial Gradients , 2018, ArXiv.

[5]  Aleksander Madry,et al.  Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.

[6]  Ruben Martinez-Cantin,et al.  Practical Bayesian optimization in the presence of outliers , 2017, AISTATS.

[7]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[8]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[9]  P. Verhulst Notice sur la loi que la population pursuit dans son accroissement , 1838 .

[10]  David A. Wagner,et al.  Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).

[11]  Fabio Roli,et al.  Wild Patterns: Ten Years After the Rise of Adversarial Machine Learning , 2017, Pattern Recognit..

[12]  Jonathon Shlens,et al.  Explaining and Harnessing Adversarial Examples , 2014, ICLR.

[13]  A. Shariff,et al.  Psychological roadblocks to the adoption of self-driving vehicles , 2017, Nature Human Behaviour.

[14]  Dániel Varga,et al.  Gradient Regularization Improves Accuracy of Discriminative Models , 2017, ArXiv.

[15]  B. Jang,et al.  Prediction of Pseudoprogression versus Progression using Machine Learning Algorithm in Glioblastoma , 2018, Scientific Reports.

[16]  W. Brendel,et al.  Foolbox: A Python toolbox to benchmark the robustness of machine learning models , 2017 .

[17]  D. Floreano,et al.  Evolution of Adaptive Behaviour in Robots by Means of Darwinian Selection , 2010, PLoS biology.

[18]  Germinal Cocho,et al.  Cancer growth and metastasis as a metaphor of Go gaming: An Ising model approach , 2018, PloS one.

[19]  Seyed-Mohsen Moosavi-Dezfooli,et al.  Universal Adversarial Perturbations , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[20]  Ananthram Swami,et al.  The Limitations of Deep Learning in Adversarial Settings , 2015, 2016 IEEE European Symposium on Security and Privacy (EuroS&P).

[21]  Yingzhen Li,et al.  Are Generative Classifiers More Robust to Adversarial Attacks? , 2018, ICML.

[22]  Martin Wattenberg,et al.  SmoothGrad: removing noise by adding noise , 2017, ArXiv.

[23]  Xingquan Zhu,et al.  Class Noise vs. Attribute Noise: A Quantitative Study , 2003, Artificial Intelligence Review.

[24]  Seyed-Mahdi Khaligh-Razavi,et al.  How Can Selection of Biologically Inspired Features Improve the Performance of a Robust Object Recognition Model? , 2012, PloS one.

[25]  Amy Maxmen Deep learning sharpens views of cells and genes , 2018, Nature.

[26]  Heinz von Foerster,et al.  Doomsday: Friday, 13 November, A.D. 2026 , 1960 .

[27]  Joan Bruna,et al.  Intriguing properties of neural networks , 2013, ICLR.

[28]  Fabio Roli,et al.  Wild Patterns: Ten Years After the Rise of Adversarial Machine Learning , 2018, CCS.

[29]  Antoine Cully,et al.  Robots that can adapt like animals , 2014, Nature.

[30]  Moustapha Cissé,et al.  Countering Adversarial Images using Input Transformations , 2018, ICLR.

[31]  Seyed-Mohsen Moosavi-Dezfooli,et al.  DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[32]  Yoshua Bengio,et al.  Random Search for Hyper-Parameter Optimization , 2012, J. Mach. Learn. Res..

[33]  Harris Drucker,et al.  Improving generalization performance using double backpropagation , 1992, IEEE Trans. Neural Networks.

[34]  The case for and against autonomous weapon systems , 2017, Nature Human Behaviour.

[35]  P. Fischer,et al.  Bioinspired microrobots , 2018, Nature Reviews Materials.

[36]  F. Iida,et al.  Morphological Evolution of Physical Robots through Model-Free Phenotype Development , 2015, PloS one.

[37]  Peter Stone,et al.  Scalable training of artificial neural networks with adaptive sparse connectivity inspired by network science , 2017, Nature Communications.

[38]  David A. Wagner,et al.  Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.

[39]  Y. van de Peer,et al.  Improving the Adaptability of Simulated Evolutionary Swarm Robots in Dynamically Changing Environments , 2014, PloS one.

[40]  L. Floridi,et al.  Regulate artificial intelligence to avert cyber arms race , 2018, Nature.

[41]  Robert J. Wood,et al.  Science, technology and the future of small autonomous drones , 2015, Nature.

[42]  Iyad Rahwan,et al.  Cooperating with machines , 2017, Nature Communications.

[43]  Valentina Zantedeschi,et al.  Efficient Defenses Against Adversarial Attacks , 2017, AISec@CCS.

[44]  Samy Bengio,et al.  Adversarial examples in the physical world , 2016, ICLR.

[45]  F. Wilcoxon Individual Comparisons by Ranking Methods , 1945 .

[46]  Benjamin Edwards,et al.  Detecting Backdoor Attacks on Deep Neural Networks by Activation Clustering , 2018, SafeAI@AAAI.

[47]  Alexandros G. Dimakis,et al.  The Robust Manifold Defense: Adversarial Training using Generative Models , 2017, ArXiv.

[48]  Steve Oudot,et al.  Statistical Analysis and Parameter Selection for Mapper , 2017, J. Mach. Learn. Res..

[49]  R. Schoen Dynamic Population Models , 2006 .

[50]  Andrew Janowczyk,et al.  A deep-learning classifier identifies patients with clinical heart failure using whole-slide images of H&E tissue , 2018, PloS one.

[51]  Dario Floreano,et al.  Bio-Inspired Artificial Intelligence: Theories, Methods, and Technologies , 2008 .

[52]  P. Verhulst,et al.  Notice sur la loi que la population suit dans son accroissement. Correspondance Mathematique et Physique Publiee par A , 1838 .

[53]  Yanjun Qi,et al.  Feature Squeezing: Detecting Adversarial Examples in Deep Neural Networks , 2017, NDSS.

[54]  Lulu Qian,et al.  Scaling up molecular pattern recognition with DNA-based winner-take-all neural networks , 2018, Nature.

[55]  Andrew Slavin Ross,et al.  Improving the Adversarial Robustness and Interpretability of Deep Neural Networks by Regularizing their Input Gradients , 2017, AAAI.

[56]  Shin Ishii,et al.  Distributional Smoothing with Virtual Adversarial Training , 2015, ICLR 2016.

[57]  Nathan Srebro,et al.  The Implicit Bias of Gradient Descent on Separable Data , 2017, J. Mach. Learn. Res..