SLEAP: Multi-animal pose tracking

The desire to understand how the brain generates and patterns behavior has driven rapid methodological innovation to quantify and model natural animal behavior. This has led to important advances in deep learning-based markerless pose estimation that have been enabled in part by the success of deep learning for computer vision applications. Here we present SLEAP (Social LEAP Estimates Animal Poses), a framework for multi-animal pose tracking via deep learning. This system is capable of simultaneously tracking any number of animals during social interactions and across a variety of experimental conditions. SLEAP implements several complementary approaches for dealing with the problems inherent in moving from single-to multi-animal pose tracking, including configurable neural network architectures, inference techniques, and tracking algorithms, enabling easy specialization and tuning for particular experimental conditions or performance requirements. We report results on multiple datasets of socially interacting animals (flies, bees, and mice) and describe how dataset-specific properties can be leveraged to determine the best configuration of SLEAP models. Using a high accuracy model (<2.8 px error on 95% of points), we were able to track two animals from full size 1024 × 1024 pixel frames at up to 320 FPS. The SLEAP framework comes with a sophisticated graphical user interface, multi-platform support, Colab-based GPU-free training and inference, and complete tutorials available, in addition to the datasets, at sleap.ai.

[1]  Gunnar Farnebäck,et al.  Two-Frame Motion Estimation Based on Polynomial Expansion , 2003, SCIA.

[2]  Yaser Sheikh,et al.  OpenPose: Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields , 2018, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[3]  Matthias Bethge,et al.  Pretraining boosts out-of-domain robustness for pose estimation , 2019, 2021 IEEE Winter Conference on Applications of Computer Vision (WACV).

[4]  Luc Van Gool,et al.  The Pascal Visual Object Classes (VOC) Challenge , 2010, International Journal of Computer Vision.

[5]  Varun Ramakrishna,et al.  Convolutional Pose Machines , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[6]  Danny Crookes,et al.  Detection and Tracking of Multiple Mice Using Part Proposal Networks , 2019, IEEE transactions on neural networks and learning systems.

[7]  Andre Araujo,et al.  Computing Receptive Fields of Convolutional Neural Networks , 2019, Distill.

[8]  Mikhail Kislin,et al.  Fast animal pose estimation using deep neural networks , 2018, Nature Methods.

[9]  Seng Bum Michael Yoo,et al.  OpenMonkeyStudio: Automated Markerless Pose Estimation in Freely Moving Macaques , 2020, bioRxiv.

[10]  Iasonas Kokkinos,et al.  Semantic Image Segmentation with Deep Convolutional Nets and Fully Connected CRFs , 2014, ICLR.

[11]  J. Altmann,et al.  Observational study of behavior: sampling methods. , 1974, Behaviour.

[12]  Thomas Brox,et al.  U-Net: Convolutional Networks for Biomedical Image Segmentation , 2015, MICCAI.

[13]  Paul Nuehrenberg,et al.  High-resolution animal tracking with integration of environmental information in aquatic systems , 2020, bioRxiv.

[14]  David J. Anderson,et al.  Computational Neuroethology: A Call to Action , 2019, Neuron.

[15]  Kevin M. Cury,et al.  DeepLabCut: markerless pose estimation of user-defined body parts with deep learning , 2018, Nature Neuroscience.

[16]  Bernt Schiele,et al.  PoseTrack: A Benchmark for Human Pose Estimation and Tracking , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[17]  Zhen He,et al.  Numerical Coordinate Regression with Convolutional Neural Networks , 2018, ArXiv.

[18]  Alexander Mathis,et al.  Deep learning tools for the measurement of animal behavior in neuroscience , 2019, Current Opinion in Neurobiology.

[19]  Dave Goulson,et al.  Bumblebees: Behaviour, Ecology, and Conservation , 2003 .

[20]  Bingni W. Brunton,et al.  Anipose: A toolkit for robust markerless 3D pose estimation , 2020, bioRxiv.

[21]  Bin Xiao,et al.  Bottom-up Higher-Resolution Networks for Multi-Person Pose Estimation , 2019, ArXiv.

[22]  Pietro Perona,et al.  Microsoft COCO: Common Objects in Context , 2014, ECCV.

[23]  Peyman Golshani,et al.  DeepBehavior: A Deep Learning Toolbox for Automated Analysis of Animal and Human Behavior Imaging Data , 2019, Front. Syst. Neurosci..

[24]  Pietro Perona,et al.  Benchmarking and Error Diagnosis in Multi-instance Pose Estimation , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[25]  Davide Modolo,et al.  Combining Detection and Tracking for Human Pose Estimation in Videos , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[26]  Christian L. Ebbesen,et al.  Automatic tracking of mouse social posture dynamics by 3D videography, deep learning and GPU-accelerated robust optimization , 2020 .

[27]  Rachel I. Wilson,et al.  Parallel Transformation of Tactile Signals in Central Circuits of Drosophila , 2016, Cell.

[28]  Eric Hervet,et al.  Applications for deep learning in ecology , 2018, bioRxiv.

[29]  Yichen Wei,et al.  Simple Baselines for Human Pose Estimation and Tracking , 2018, ECCV.

[30]  Gordon J. Berman,et al.  Measuring behavior across scales , 2017, BMC Biology.

[31]  Jacob M. Graving,et al.  DeepPoseKit, a software toolkit for fast and robust animal pose estimation using deep learning , 2019, bioRxiv.

[32]  Yaser Sheikh,et al.  Efficient Online Multi-Person 2D Pose Tracking With Recurrent Spatio-Temporal Affinity Fields , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).