Deep convolutional neural networks for image-based Convolvulus sepium detection in sugar beet fields

Convolvulus sepium (hedge bindweed) detection in sugar beet fields remains a challenging problem due to variation in appearance of plants, illumination changes, foliage occlusions, and different growth stages under field conditions. Current approaches for weed and crop recognition, segmentation and detection rely predominantly on conventional machine-learning techniques that require a large set of hand-crafted features for modelling. These might fail to generalize over different fields and environments. Here, we present an approach that develops a deep convolutional neural network (CNN) based on the tiny YOLOv3 architecture for C. sepium and sugar beet detection. We generated 2271 synthetic images, before combining these images with 452 field images to train the developed model. YOLO anchor box sizes were calculated from the training dataset using a k-means clustering approach. The resulting model was tested on 100 field images, showing that the combination of synthetic and original field images to train the developed model could improve the mean average precision (mAP) metric from 0.751 to 0.829 compared to using collected field images alone. We also compared the performance of the developed model with the YOLOv3 and Tiny YOLO models. The developed model achieved a better trade-off between accuracy and speed. Specifically, the average precisions (APs@IoU0.5) of C. sepium and sugar beet were 0.761 and 0.897 respectively with 6.48 ms inference time per image (800 × 1200) on a NVIDIA Titan X GPU environment. The developed model has the potential to be deployed on an embedded mobile platform like the Jetson TX for online weed detection and management due to its high-speed inference. It is recommendable to use synthetic images and empirical field images together in training stage to improve the performance of models.

[1]  Yong He,et al.  Recognising weeds in a maize crop using a random forest machine-learning algorithm and near-infrared snapshot mosaic hyperspectral imagery , 2018, Biosystems Engineering.

[2]  Janny L. Peters,et al.  Thrips Resistance Screening Is Coming of Age: Leaf Position and Ontogeny Are Important Determinants of Leaf-Based Resistance in Pepper , 2019, Front. Plant Sci..

[3]  Huan Zhang,et al.  Localization and Classification of Paddy Field Pests using a Saliency Map and Deep Convolutional Neural Network , 2016, Scientific Reports.

[4]  Trevor Darrell,et al.  Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation , 2013, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[5]  D. Peddle,et al.  Weed and crop discrimination using hyperspectral image data and reduced bandsets , 2014 .

[6]  Gerrit Polder,et al.  Potato Virus Y Detection in Seed Potatoes Using Deep Learning on Hyperspectral Images , 2019, Front. Plant Sci..

[7]  Cyrill Stachniss,et al.  Effective Vision‐based Classification for Separating Sugar Beets and Weeds for Precision Farming , 2017, J. Field Robotics.

[8]  Yoshua Bengio,et al.  How transferable are features in deep neural networks? , 2014, NIPS.

[9]  Roland Siegwart,et al.  weedNet: Dense Semantic Weed Classification Using Multispectral Images and MAV for Smart Farming , 2017, IEEE Robotics and Automation Letters.

[10]  A. T. Nieuwenhuizen,et al.  Performance evaluation of an automated detection and control system for volunteer potatoes in sugar beet fields , 2010 .

[11]  Shunping Xiao,et al.  Small Object Detection in Optical Remote Sensing Images via Modified Faster R-CNN , 2018 .

[12]  K. Neil Harker,et al.  Recent Weed Control, Weed Management, and Integrated Weed Management , 2013, Weed Technology.

[13]  Andreas Kamilaris,et al.  Deep learning in agriculture: A survey , 2018, Comput. Electron. Agric..

[14]  Gautam Shroff,et al.  Crop Planning using Stochastic Visual Optimization , 2017, 2017 IEEE Visualization in Data Science (VDS).

[15]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[16]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[17]  Luc Van Gool,et al.  The Pascal Visual Object Classes (VOC) Challenge , 2010, International Journal of Computer Vision.

[18]  Hayit Greenspan,et al.  Synthetic data augmentation using GAN for improved liver lesion classification , 2018, 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018).

[19]  Konstantinos P. Ferentinos,et al.  Deep learning models for plant disease detection and diagnosis , 2018, Comput. Electron. Agric..

[20]  Maryam Rahnemoonfar,et al.  Deep Count: Fruit Counting Based on Deep Simulated Learning , 2017, Sensors.

[21]  Haoxiang Wang,et al.  Visual features based boosted classification of weeds for real-time selective herbicide sprayer systems , 2018, Comput. Ind..

[22]  Joris IJsselmuiden,et al.  Synthetic bootstrapping of convolutional neural networks for semantic plant part segmentation , 2017, Comput. Electron. Agric..

[23]  Vasaka Visoottiviseth,et al.  Evaluating the power efficiency of deep learning inference on embedded GPU systems , 2017, 2017 2nd International Conference on Information Technology (INCIT).

[24]  Hans Jørgen Andersen,et al.  Detecting creeping thistle in sugar beet fields using vegetation indices , 2015, Comput. Electron. Agric..

[25]  Graham Brookes,et al.  Weed control changes and genetically modified herbicide tolerant crops in the USA 1996–2012 , 2014, GM crops & food.

[26]  P. Bonnet,et al.  Going deeper in the automated identification of Herbarium specimens , 2017, BMC Evolutionary Biology.

[27]  Aleksandra Pizurica,et al.  Fusion of pixel and object-based features for weed mapping using unmanned aerial vehicle imagery , 2018, Int. J. Appl. Earth Obs. Geoinformation.

[28]  Ali Farhadi,et al.  You Only Look Once: Unified, Real-Time Object Detection , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[29]  G. Meyer,et al.  Color indices for weed identification under various soil, residue, and lighting conditions , 1994 .

[30]  Cyrill Stachniss,et al.  Fully Convolutional Networks With Sequential Information for Robust Crop and Weed Detection in Precision Farming , 2018, IEEE Robotics and Automation Letters.

[31]  Ali Farhadi,et al.  YOLOv3: An Incremental Improvement , 2018, ArXiv.

[32]  Josef Soukup,et al.  Competitive relationships between sugar beet and weeds in dependence on time of weed control , 2018 .

[33]  Baskar Ganapathysubramanian,et al.  An explainable deep machine vision framework for plant stress phenotyping , 2018, Proceedings of the National Academy of Sciences.

[34]  Anders Krogh Mortensen,et al.  Pixel-wise classification of weeds and crops in images by using a Fully Convolutional neural network , 2016 .

[35]  C DeepikaH An Overview of You Only Look Once: Unified, Real-Time Object Detection , 2020 .

[36]  Hany Farid,et al.  Blind inverse gamma correction , 2001, IEEE Trans. Image Process..

[37]  David R. Shaw,et al.  Remote sensing and site-specific weed management , 2005 .

[38]  Hilde van der Togt,et al.  Publisher's Note , 2003, J. Netw. Comput. Appl..

[39]  N. Otsu A threshold selection method from gray level histograms , 1979 .

[40]  H. Beckie,et al.  The future for weed control and technology. , 2014, Pest management science.

[41]  Jochen Hemming,et al.  Data synthesis methods for semantic segmentation in agriculture: A Capsicum annuum dataset , 2018, Comput. Electron. Agric..

[42]  Wei Liu,et al.  SSD: Single Shot MultiBox Detector , 2015, ECCV.

[43]  Cyrill Stachniss,et al.  WeedMap: A large-scale semantic weed mapping framework using aerial multispectral imaging and deep neural network for precision farming , 2018, Remote. Sens..

[44]  Tony P. Pridmore,et al.  Deep machine learning provides state-of-the-art performance in image-based plant phenotyping , 2016, bioRxiv.

[45]  Cyrill Stachniss,et al.  UAV-based crop and weed classification for smart farming , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[46]  Joris IJsselmuiden,et al.  Transfer learning for the classification of sugar beet and volunteer potato under field conditions , 2018, Biosystems Engineering.

[47]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.