Selective Particle Attention: Visual Feature-Based Attention in Deep Reinforcement Learning

The human brain uses selective attention to filter perceptual input so that only the components that are useful for behaviour are processed using its limited computational resources. We focus on one particular form of visual attention known as feature-based attention, which is concerned with identifying features of the visual input that are important for the current task regardless of their spatial location. Visual feature-based attention has been proposed to improve the efficiency of Reinforcement Learning (RL) by reducing the dimensionality of state representations and guiding learning towards relevant features. Despite achieving human level performance in complex perceptual-motor tasks, Deep RL algorithms have been consistently criticised for their poor efficiency and lack of flexibility. Visual feature-based attention therefore represents one option for addressing these criticisms. Nevertheless, it is still an open question how the brain is able to learn which features to attend to during RL. To help answer this question we propose a novel algorithm, termed Selective Particle Attention (SPA), which imbues a Deep RL agent with the ability to perform selective feature-based attention. SPA learns which combinations of features to attend to based on their bottom-up saliency and how accurately they predict future reward. We evaluate SPA on a multiple choice task and a 2D video game that both involve raw pixel input and dynamic changes to the task structure. We show various benefits of SPA over approaches that naively attend to either all or random subsets of features. Our results demonstrate (1) how visual feature-based attention in Deep RL models can improve their learning efficiency and ability to deal with sudden changes in task structure and (2) that particle filters may represent a viable computational account of how visual feature-based attention occurs in the brain.

[1]  Vaidehi S. Natu,et al.  The functional neuroanatomy of face perception: from brain measurements to deep neural networks , 2018, Interface Focus.

[2]  Joshua B. Tenenbaum,et al.  Building machines that learn and think like people , 2016, Behavioral and Brain Sciences.

[3]  Razvan Pascanu,et al.  Relational Deep Reinforcement Learning , 2018, ArXiv.

[4]  G. Schoenbaum,et al.  Neural Encoding in Ventral Striatum during Olfactory Discrimination Learning , 2003, Neuron.

[5]  Fabian Canas,et al.  Integrating reinforcement learning with models of representation learning , 2010 .

[6]  Alex Graves,et al.  Recurrent Models of Visual Attention , 2014, NIPS.

[7]  Anton van den Hengel,et al.  Reinforcement Learning with Attention that Works: A Self-Supervised Approach , 2019, ICONIP.

[8]  N. Mackintosh A Theory of Attention: Variations in the Associability of Stimuli with Reinforcement , 1975 .

[9]  Grace W. Lindsay Attention in Psychology, Neuroscience, and Machine Learning , 2020, Frontiers in Computational Neuroscience.

[10]  E. Miller,et al.  An integrative theory of prefrontal cortex function. , 2001, Annual review of neuroscience.

[11]  Bradley C. Love,et al.  The Costs and Benefits of Goal-Directed Attention in Deep Convolutional Neural Networks , 2020, Computational brain & behavior.

[12]  T. Maia Reinforcement learning, conditioning, and the brain: Successes and challenges , 2009, Cognitive, affective & behavioral neuroscience.

[13]  E. Lorch,et al.  Attentional inertia reduces distractibility during young children's TV viewing. , 1987, Child development.

[14]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[15]  M. Paradiso,et al.  Feature-specific effects of selective visual attention , 1995, Vision Research.

[16]  S. Monsell,et al.  Attentional inertia and delayed orienting of spatial attention in task-switching. , 2014, Journal of experimental psychology. Human perception and performance.

[17]  Robert C. Wilson,et al.  Reinforcement Learning in Multidimensional Environments Relies on Attention Mechanisms , 2015, The Journal of Neuroscience.

[18]  J. Wolfe,et al.  What attributes guide the deployment of visual attention and how do they do it? , 2004, Nature Reviews Neuroscience.

[19]  Robert C. Wilson,et al.  Inferring Relevance in a Changing World , 2012, Front. Hum. Neurosci..

[20]  J. DiCarlo,et al.  Using goal-driven deep learning models to understand sensory cortex , 2016, Nature Neuroscience.

[21]  Robert L. Goldstone,et al.  The development of features in object concepts , 1998, Behavioral and Brain Sciences.

[22]  S. Treue Visual attention: the where, what, how and why of saliency , 2003, Current Opinion in Neurobiology.

[23]  Yoshua Bengio,et al.  Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.

[24]  Daniel R. Anderson,et al.  Attentional Inertia and Recognition Memory in Adult Television Viewing , 1993 .

[25]  Koray Kavukcuoglu,et al.  Multiple Object Recognition with Visual Attention , 2014, ICLR.

[26]  R. Shepard,et al.  Learning and memorization of classifications. , 1961 .

[27]  Mikhail Pavlov,et al.  Deep Attention Recurrent Q-Network , 2015, ArXiv.

[28]  A. Cooper,et al.  Predictive Reward Signal of Dopamine Neurons , 2011 .

[29]  Christopher Burgess,et al.  beta-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework , 2016, ICLR 2016.

[30]  N. P. Bichot,et al.  A Source for Feature-Based Attention in the Prefrontal Cortex , 2015, Neuron.

[31]  Daniel R. Anderson,et al.  Attentional inertia in children's extended looking at television. , 2004, Advances in child development and behavior.

[32]  Georgia G. Gregoriou,et al.  Top-Down Control of Visual Attention by the Prefrontal Cortex. Functional Specialization and Long-Range Interactions , 2017, Front. Neurosci..

[33]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[34]  Yong Jeong,et al.  Exploring Feature Dimensions to Learn a New Policy in an Uninformed Reinforcement Learning Task , 2017, Scientific Reports.

[35]  Kenneth D Miller,et al.  How biological attention mechanisms improve task performance in a large-scale visual system model , 2017, bioRxiv.

[36]  Li Fei-Fei,et al.  ImageNet: A large-scale hierarchical image database , 2009, CVPR.

[37]  Alex Mott,et al.  Towards Interpretable Reinforcement Learning Using Attention Augmented Agents , 2019, NeurIPS.

[38]  G. Boynton,et al.  Global effects of feature-based attention in human visual cortex , 2002, Nature Neuroscience.

[39]  Marcel A. J. van Gerven,et al.  Deep Neural Networks Reveal a Gradient in the Complexity of Neural Representations across the Ventral Stream , 2014, The Journal of Neuroscience.

[40]  Daeyeol Lee,et al.  Feature-based learning improves adaptability without compromising precision , 2017, Nature Communications.

[41]  Ian C. Ballard,et al.  Holistic Reinforcement Learning: The Role of Structure and Attention , 2019, Trends in Cognitive Sciences.

[42]  Tao Mei,et al.  Look Closer to See Better: Recurrent Attention Convolutional Neural Network for Fine-Grained Image Recognition , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[43]  Peter Dayan,et al.  A Neural Substrate of Prediction and Reward , 1997, Science.

[44]  Fei-FeiLi,et al.  One-Shot Learning of Object Categories , 2006 .

[45]  Jonas Kubilius,et al.  Brain-Score: Which Artificial Neural Network for Object Recognition is most Brain-Like? , 2018, bioRxiv.

[46]  Alex Graves,et al.  Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.

[47]  Eytan Ruppin,et al.  Actor-critic models of the basal ganglia: new anatomical and computational perspectives , 2002, Neural Networks.

[48]  Yael Niv,et al.  A particle filtering account of selective attention during learning , 2019, 2019 Conference on Cognitive Computational Neuroscience.

[49]  O. Hikosaka Models of information processing in the basal Ganglia edited by James C. Houk, Joel L. Davis and David G. Beiser, The MIT Press, 1995. $60.00 (400 pp) ISBN 0 262 08234 9 , 1995, Trends in Neurosciences.

[50]  Aurelio Cortese,et al.  Attention or memory? Neurointerpretable agents in space and time , 2020, ArXiv.

[51]  Pietro Perona,et al.  One-shot learning of object categories , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[52]  Yuan Chang Leong,et al.  Dynamic Interaction between Reinforcement Learning and Attention in Multidimensional Environments , 2017, Neuron.

[53]  Stefan Treue,et al.  Feature-based attention influences motion processing gain in macaque visual cortex , 1999, Nature.

[54]  Yuxin Peng,et al.  The application of two-level attention models in deep convolutional neural network for fine-grained image classification , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).