Synthetic-Neuroscore: Using a neuro-AI interface for evaluating generative adversarial networks

Generative adversarial networks (GANs) are increasingly attracting attention in the computer vision, natural language processing, speech synthesis and similar domains. Arguably the most striking results have been in the area of image synthesis. However, evaluating the performance of GANs is still an open and challenging problem. Existing evaluation metrics primarily measure the dissimilarity between real and generated images using automated statistical methods. They often require large sample sizes for evaluation and do not directly reflect human perception of image quality. In this work, we describe an evaluation metric we call Neuroscore, for evaluating the performance of GANs, that more directly reflects psychoperceptual image quality through the utilization of brain signals. Our results show that Neuroscore has superior performance to the current evaluation metrics in that: (1) It is more consistent with human judgment; (2) The evaluation process needs much smaller numbers of samples; and (3) It is able to rank the quality of images on a per GAN basis. A convolutional neural network (CNN) based neuro-AI interface is proposed to predict Neuroscore from GAN-generated images directly without the need for neural responses. Importantly, we show that including neural responses during the training phase of the network can significantly improve the prediction capability of the proposed model. Materials related to this work are provided at this https URL.

[1]  Wojciech Zaremba,et al.  Improved Techniques for Training GANs , 2016, NIPS.

[2]  Li Fei-Fei,et al.  Distinct contributions of functional and deep neural network features to representational similarity of scenes in human brain and behavior , 2018, eLife.

[3]  Surya Ganguli,et al.  Deep Learning Models of the Retinal Response to Natural Scenes , 2017, NIPS.

[4]  Sergey Ioffe,et al.  Rethinking the Inception Architecture for Computer Vision , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[5]  Léon Bottou,et al.  Wasserstein GAN , 2017, ArXiv.

[6]  Sepp Hochreiter,et al.  GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium , 2017, NIPS.

[7]  Kilian Q. Weinberger,et al.  An empirical study on evaluation metrics of generative adversarial networks , 2018, ArXiv.

[8]  Stephan Waldert,et al.  Invasive vs. Non-Invasive Neuronal Signals for Brain-Machine Interfaces: Will One Prevail? , 2016, Front. Neurosci..

[9]  Radoslaw Martin Cichy,et al.  Deep Neural Networks as Scientific Models , 2019, Trends in Cognitive Sciences.

[10]  Kai Xu,et al.  Reduced-Rank Linear Dynamical Systems , 2018, AAAI.

[11]  Antonio Torralba,et al.  Comparison of deep neural networks to spatio-temporal cortical dynamics of human visual object recognition reveals hierarchical correspondence , 2016, Scientific Reports.

[12]  David Berthelot,et al.  BEGAN: Boundary Equilibrium Generative Adversarial Networks , 2017, ArXiv.

[13]  Evan M. Palmer,et al.  Reaction time distributions constrain models of visual search , 2010, Vision Research.

[14]  J. DiCarlo,et al.  Using goal-driven deep learning models to understand sensory cortex , 2016, Nature Neuroscience.

[15]  J. Polich Updating P300: An integrative theory of P3a and P3b , 2007, Clinical Neurophysiology.

[16]  Klaus-Robert Müller,et al.  The LDA beamformer: Optimal estimation of ERP source time series using linear discriminant analysis , 2016, NeuroImage.

[17]  Nikolaus Kriegeskorte,et al.  Deep neural networks: a new framework for modelling biological vision and brain information processing , 2015, bioRxiv.

[18]  Alan F. Smeaton,et al.  Exploring EEG for Object Detection and Retrieval , 2015, ICMR.

[19]  Richard S. Zemel,et al.  Generative Moment Matching Networks , 2015, ICML.

[20]  Ki-Young Jung,et al.  Influence of task difficulty on the features of event-related potential during visual oddball task , 2008, Neuroscience Letters.

[21]  Alan F. Smeaton,et al.  A review of feature extraction and classification algorithms for image RSVP based BCI , 2018 .

[22]  Bernhard Schölkopf,et al.  A Kernel Two-Sample Test , 2012, J. Mach. Learn. Res..

[23]  Alexei A. Efros,et al.  Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[24]  Aaron P. Batista,et al.  Deep learning reaches the motor system , 2018, Nature Methods.

[25]  M. Botsch,et al.  Differential effects of face-realism and emotion on event-related brain potentials and their implications for the uncanny valley theory , 2017, Scientific Reports.

[26]  Anthony J. Ries,et al.  A Novel Method for Single-Trial Classification in the Face of Temporal Variability , 2013, HCI.

[27]  Percy Liang,et al.  Understanding Black-box Predictions via Influence Functions , 2017, ICML.

[28]  Mark Sandler,et al.  MobileNetV2: Inverted Residuals and Linear Bottlenecks , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[29]  K. Upton,et al.  A modern approach , 1995 .

[30]  P. Sajda,et al.  Cortically coupled computer vision for rapid image search , 2006, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[31]  Christian Ledig,et al.  Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[32]  M. Carrillo-de-la-Peña,et al.  The effect of motivational instructions on P300 amplitude , 2000, Neurophysiologie Clinique/Clinical Neurophysiology.

[33]  Tomas E. Ward,et al.  Generative Adversarial Networks: A Survey and Taxonomy , 2019, ArXiv.

[34]  Timothée Masquelier,et al.  Deep Networks Can Resemble Human Feed-forward Vision in Invariant Object Recognition , 2015, Scientific Reports.

[35]  Chris Donahue,et al.  Synthesizing Audio with Generative Adversarial Networks , 2018, ArXiv.

[36]  Yifang Wang,et al.  An event-related potential comparison of facial expression processing between cartoon and real faces , 2019, PloS one.

[37]  Cathal Gurrin,et al.  Experiences and Insights from the Collection of a Novel Multimedia EEG Dataset , 2020, MMM.

[38]  Alan F. Smeaton,et al.  An investigation of triggering approaches for the rapid serial visual presentation paradigm in brain computer interfacing , 2016, 2016 27th Irish Signals and Systems Conference (ISSC).

[39]  Soumith Chintala,et al.  Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks , 2015, ICLR.

[40]  Andrea Cester,et al.  Flexible and Organic Neural Interfaces: A Review , 2017 .

[41]  Philippe Kahane,et al.  Activations of deep convolutional neural networks are aligned with gamma band activity of human visual cortex , 2017, Communications Biology.

[42]  Ali Borji,et al.  Pros and Cons of GAN Evaluation Measures , 2018, Comput. Vis. Image Underst..

[43]  Rishi Sharma,et al.  A Note on the Inception Score , 2018, ArXiv.

[44]  Li Fei-Fei,et al.  ImageNet: A large-scale hierarchical image database , 2009, CVPR.

[45]  John Duncan,et al.  A neural basis for visual search in inferior temporal cortex , 1993, Nature.

[46]  Alan F. Smeaton,et al.  Use of Neural Signals to Evaluate the Quality of Generative Adversarial Network Performance in Facial Image Generation , 2018, Cognitive Computation.

[47]  P. Sajda,et al.  Human Scalp Potentials Reflect a Mixture of Decision-Related Signals during Perceptual Choices , 2014, The Journal of Neuroscience.

[48]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[49]  Xiaogang Wang,et al.  Deep Learning Face Attributes in the Wild , 2014, 2015 IEEE International Conference on Computer Vision (ICCV).

[50]  Alan F. Smeaton,et al.  Spatial Filtering Pipeline Evaluation of Cortically Coupled Computer Vision System for Rapid Serial Visual Presentation , 2018, Brain-Computer Interfaces.

[51]  D. Senkowski,et al.  Effects of task difficulty on evoked gamma activity and ERPs in a visual discrimination task , 2002, Clinical Neurophysiology.

[52]  Thomas S. Huang,et al.  Generative Image Inpainting with Contextual Attention , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[53]  A Mouraux,et al.  Across-trial averaging of event-related EEG responses and beyond. , 2008, Magnetic resonance imaging.

[54]  Alan F. Smeaton,et al.  Overview of NTCIR-13 NAILS Task , 2017, NTCIR.

[55]  Tomas E. Ward,et al.  Quick and Easy Time Series Generation with Established Image-based GANs , 2019, ArXiv.

[56]  Paul Sajda,et al.  Relating Deep Neural Network Representations to EEG-fMRI Spatiotemporal Dynamics in a Perceptual Decision-Making Task , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).

[57]  Guanrong Chen,et al.  Evaluating the Small-World-Ness of a Sampled Network: Functional Connectivity of Entorhinal-Hippocampal Circuitry , 2016, Scientific Reports.

[58]  Timo Aila,et al.  A Style-Based Generator Architecture for Generative Adversarial Networks , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[59]  Ha Hong,et al.  Performance-optimized hierarchical models predict neural responses in higher visual cortex , 2014, Proceedings of the National Academy of Sciences.

[60]  T. Schubert,et al.  Comparison of the Working Memory Load in N-Back and Working Memory Span Tasks by Means of EEG Frequency Band Power and P300 Amplitude , 2017, Front. Hum. Neurosci..