Analyzing animal behavior via classifying each video frame using convolutional neural networks

High-throughput analysis of animal behavior requires software to analyze videos. Such software analyzes each frame individually, detecting animals’ body parts. But the image analysis rarely attempts to recognize “behavioral states”—e.g., actions or facial expressions—directly from the image instead of using the detected body parts. Here, we show that convolutional neural networks (CNNs)—a machine learning approach that recently became the leading technique for object recognition, human pose estimation, and human action recognition—were able to recognize directly from images whether Drosophila were “on” (standing or walking) or “off” (not in physical contact with) egg-laying substrates for each frame of our videos. We used multiple nets and image transformations to optimize accuracy for our classification task, achieving a surprisingly low error rate of just 0.072%. Classifying one of our 8 h videos took less than 3 h using a fast GPU. The approach enabled uncovering a novel egg-laying-induced behavior modification in Drosophila. Furthermore, it should be readily applicable to other behavior analysis tasks.

[1]  Ming Yang,et al.  DeepFace: Closing the Gap to Human-Level Performance in Face Verification , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[2]  Fei-Fei Li,et al.  Large-Scale Video Classification with Convolutional Neural Networks , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[3]  Trevor Darrell,et al.  DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition , 2013, ICML.

[4]  Karla R Kaun,et al.  Competing dopamine neurons drive oviposition choice for ethanol in Drosophila , 2013, Proceedings of the National Academy of Sciences.

[5]  Jonathan Tompson,et al.  Learning Human Pose Estimation Features with Convolutional Networks , 2013, ICLR.

[6]  Kristin Branson,et al.  JAABA: interactive machine learning for automatic annotation of animal behavior , 2013, Nature Methods.

[7]  Andrew Zisserman,et al.  Deep Convolutional Neural Networks for Efficient Pose Estimation in Gesture Videos , 2014, ACCV.

[8]  Ivan Laptev,et al.  Learning and Transferring Mid-level Image Representations Using Convolutional Neural Networks , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[9]  Jonathan Tompson,et al.  Joint Training of a Convolutional Network and a Graphical Model for Human Pose Estimation , 2014, NIPS.

[10]  David J. Anderson,et al.  Toward a Science of Computational Ethology , 2014, Neuron.

[11]  Trevor Hastie,et al.  An Introduction to Statistical Learning , 2013, Springer Texts in Statistics.

[12]  Larry Wasserman,et al.  All of Statistics , 2004 .

[13]  Michele Vendruscolo,et al.  The iFly tracking system for an automated locomotor and behavioural analysis of Drosophila melanogaster. , 2011, Integrative biology : quantitative biosciences from nano to macro.

[14]  Yann LeCun,et al.  Regularization of Neural Networks using DropConnect , 2013, ICML.

[15]  Rob Fergus,et al.  Visualizing and Understanding Convolutional Neural Networks , 2013 .

[16]  Larry Wasserman,et al.  All of Statistics: A Concise Course in Statistical Inference , 2004 .

[17]  Michael H Dickinson,et al.  Wing and body motion during flight initiation in Drosophila revealed by automated visual tracking , 2009, Journal of Experimental Biology.

[18]  Jitendra Malik,et al.  R-CNNs for Pose Estimation and Action Detection , 2014, ArXiv.

[19]  Luc Van Gool,et al.  The Pascal Visual Object Classes Challenge: A Retrospective , 2014, International Journal of Computer Vision.

[20]  Geoffrey E. Hinton,et al.  Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.

[21]  Ruo He,et al.  Behavioral and Circuit Basis of Sucrose Rejection by Drosophila Females in a Simple Decision-Making Task , 2015, The Journal of Neuroscience.

[22]  Jonathan Schor,et al.  Detecting Social Actions of Fruit Flies , 2014, ECCV.

[23]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[24]  P. Perona,et al.  utomated multi-day tracking of marked mice for the analysis of ocial behaviour , 2013 .

[25]  Andrew Zisserman,et al.  Two-Stream Convolutional Networks for Action Recognition in Videos , 2014, NIPS.

[26]  Pietro Perona,et al.  Automated monitoring and analysis of social behavior in Drosophila , 2009, Nature Methods.

[27]  Silke Sachse,et al.  Olfactory Preference for Egg Laying on Citrus Substrates in Drosophila , 2013, Current Biology.

[28]  Pietro Perona,et al.  High-throughput Ethomics in Large Groups of Drosophila , 2009, Nature Methods.

[29]  Christian Szegedy,et al.  DeepPose: Human Pose Estimation via Deep Neural Networks , 2013, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[30]  Li Fei-Fei,et al.  ImageNet: A large-scale hierarchical image database , 2009, CVPR.

[31]  Nitish Srivastava,et al.  Improving neural networks by preventing co-adaptation of feature detectors , 2012, ArXiv.

[32]  Richard Szeliski,et al.  Computer Vision - Algorithms and Applications , 2011, Texts in Computer Science.

[33]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.

[34]  Lawrence D. Jackel,et al.  Handwritten Digit Recognition with a Back-Propagation Network , 1989, NIPS.

[35]  A. Pérez-Escudero,et al.  idTracker: tracking individuals in a group by automatic identification of unmarked animals , 2014, Nature Methods.

[36]  A. Cressant,et al.  Computerized video analysis of social interactions in mice , 2012, Nature Methods.

[37]  Antoni B. Chan,et al.  3D Human Pose Estimation from Monocular Images with Deep Convolutional Neural Network , 2014, ACCV.

[38]  Jürgen Schmidhuber,et al.  Multi-column deep neural networks for image classification , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[39]  Ying Liu,et al.  Mechanosensitive neurons on the internal reproductive tract contribute to egg-laying-induced acetic acid attraction in Drosophila. , 2014, Cell reports.

[40]  E. Hafen,et al.  Drosophila Egg-Laying Site Selection as a System to Study Simple Decision-Making Processes , 2008, Science.

[41]  Greg J. Stephens,et al.  Automated Tracking of Animal Posture and Movement during Exploration and Sensory Orientation Behaviors , 2012, PloS one.

[42]  U. Heberlein,et al.  Oviposition preference for and positional avoidance of acetic acid provide a model for competing behavioral drives in Drosophila , 2009, Proceedings of the National Academy of Sciences.

[43]  Chung-Hui Yang,et al.  Egg-Laying Demand Induces Aversion of UV Light in Drosophila Females , 2014, Current Biology.

[44]  Michael S. Bernstein,et al.  ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.