Automatic recognition of lactating sow postures by refined two-stream RGB-D faster R-CNN

This paper proposes an end-to-end refined two-stream RGB-D Faster region convolutional neural network (R-CNN) algorithm, which fuses RGB-D image features in the feature extraction stage for recognising five postures of lactating sows (standing, sitting, sternal recumbency, ventral recumbency, and lateral recumbency) in scenes at a pig farm. Based on the Faster R-CNN algorithm, two CNNs were first used to extract the RGB image features and depth image features. Then, a proposed single RGB-D region proposal network was used to generate the regions of interest (ROIs) for the two types of image feature maps in RGB-D. Next, the features of the RGB-D ROIs were extracted and merged using a feature fusion layer. Finally, the fused features of the RGB-D ROIs were input into a Fast R-CNN to obtain the recognition results. A total of 12,600 pairs of RGB-D images of five postures were obtained by a Kinect v2.0 sensor and were randomly selected from the first 21 of 28 pens as the training set, and 5533 pairs were randomly selected from the remaining 7 pens as the test set. The proposed method was used to recognise the five postures of lactating sows. The recognition accuracy of the concatenation fusion method was the highest for the test set with average precisions for the five categories of lactating sow postures of 99.74%, 96.49%, 90.77%, 90.91%, and 99.45%, respectively. Compared with related methods (RGB-only method, depth-only method, RGB-D early fusion, and later fusion), our method attained the highest mean average precision.

[1]  Meng Wang,et al.  Robust Multiview Feature Learning for RGB-D Image Understanding , 2015, ACM Trans. Intell. Syst. Technol..

[2]  B. P. Shankar,et al.  Pre-Weaning Mortality in Pig-Causes and Management , 2009 .

[3]  Kaiming He,et al.  Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[4]  Jiwen Lu,et al.  Correlated and Individual Multi-Modal Deep Learning for RGB-D Object Recognition , 2016, ArXiv.

[5]  Jiandong Tian,et al.  RGBD Salient Object Detection via Deep Fusion , 2016, IEEE Transactions on Image Processing.

[6]  Ali Farhadi,et al.  You Only Look Once: Unified, Real-Time Object Detection , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[7]  Cordelia Schmid,et al.  Multi-region Two-Stream R-CNN for Action Detection , 2016, ECCV.

[8]  Shijian Lu,et al.  Discriminative Multi-modal Feature Fusion for RGBD Indoor Scene Recognition , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[9]  Tristan Perez,et al.  DeepFruits: A Fruit Detection System Using Deep Neural Networks , 2016, Sensors.

[10]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[11]  Yoshua Bengio,et al.  Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.

[12]  Zhengyou Zhang,et al.  A Flexible New Technique for Camera Calibration , 2000, IEEE Trans. Pattern Anal. Mach. Intell..

[13]  Kai Liu,et al.  Automatic recognition of lactating sow behaviors through depth image processing , 2016, Comput. Electron. Agric..

[14]  Claudia Bahr,et al.  Automatic monitoring of pig locomotion using image analysis , 2014 .

[15]  Geoffrey E. Hinton,et al.  Visualizing Data using t-SNE , 2008 .

[16]  Andrew Zisserman,et al.  Convolutional Two-Stream Network Fusion for Video Action Recognition , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[17]  Jian Sun,et al.  Object Detection Networks on Convolutional Feature Maps , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[18]  Luís A. Alexandre 3D Object Recognition Using Convolutional Neural Networks with Transfer Learning Between Input Channels , 2014, IAS.

[19]  Jian Sun,et al.  Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition , 2015, IEEE Trans. Pattern Anal. Mach. Intell..

[20]  Yueju Xue,et al.  Detection of passion fruits and maturity classification using Red-Green-Blue Depth images , 2018, Biosystems Engineering.

[21]  Bo Chen,et al.  MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications , 2017, ArXiv.

[22]  Derek Hoiem,et al.  Diagnosing Error in Object Detectors , 2012, ECCV.

[23]  Shuang Wang,et al.  RGB-D Hand-Held Object Recognition Based on Heterogeneous Feature Fusion , 2015, Journal of Computer Science and Technology.

[24]  Xunmu Zhu,et al.  Automatic recognition of lactating sow postures from depth images by deep learning detector , 2018, Comput. Electron. Agric..

[25]  Jiwen Lu,et al.  MMSS: Multi-modal Sharable and Specific Feature Learning for RGB-D Object Recognition , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[26]  Wolfram Burgard,et al.  Multimodal deep learning for robust RGB-D object recognition , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[27]  Tieniu Tan,et al.  A Light CNN for Deep Face Representation With Noisy Labels , 2015, IEEE Transactions on Information Forensics and Security.

[28]  Wei Liu,et al.  SSD: Single Shot MultiBox Detector , 2015, ECCV.

[29]  Aqing Yang,et al.  Automatic recognition of sow nursing behaviour using deep learning-based segmentation and spatial and temporal features , 2018, Biosystems Engineering.

[30]  Stefano Viazzi,et al.  Image feature extraction for classification of aggressive interactions among pigs , 2014 .

[31]  Rob Fergus,et al.  Visualizing and Understanding Convolutional Networks , 2013, ECCV.

[32]  Luc Van Gool,et al.  The Pascal Visual Object Classes (VOC) Challenge , 2010, International Journal of Computer Vision.

[33]  Lene Juul Pedersen,et al.  Lying down and rolling behaviour in sows in relation to piglet crushing , 2005 .

[34]  Uwe Richter,et al.  Using machine vision for investigation of changes in pig group lying patterns , 2015, Comput. Electron. Agric..

[35]  K. Stafford,et al.  Sow and piglet productivity and sow reproductive performance in farrowing pens with temporary crating or farrowing crates on a commercial New Zealand pig farm , 2015 .

[36]  Ross B. Girshick,et al.  Mask R-CNN , 2017, 1703.06870.

[37]  Daniel Berckmans,et al.  The automatic monitoring of pigs water use by cameras , 2013 .