Leaf Segmentation under Loosely Controlled Conditions

Extracting accurately the shape of a leaf is a crucial step in image-based plant identification systems. The partial or total absence of textures on leaf surface and the high color variability of leaves belonging to same species make shape as the main recognition element. For such reason, leaf segmentation plays a decisive role in the leaf recognition process. Even though many general segmentation methods have been proposed in the last decades, leaf segmentation presents specific challenges. In particular, a pixel-level precision is required in order to highlight fine scale boundary structures and discriminate similar global shapes. Moreover, even if the input image can typically be taken in controlled conditions, where the leaf is the only visible object over a white background, the user taking the picture is not necessarily an expert and the conditions are often not ideal: the leaf exhibits specular reflections, casts shadows, the background is never exactly white and is usually non-uniform, and the image can be blurry. Recently, a solution for the problem at hand has been proposed in [1] where leaf segmentation is carried out by estimating the probability distribution of foreground and background pixels. However, several drawbacks appear in this formulation due to challenging leaves like pine needles, false positives detection related to shadows and false negatives detection related to specularities. Prior distributions and post-processing operations are employed to tackle such problem, with the risk of hurting the final leaf shape. In this paper we introduce a new solution by training a pixel-wise classifier [3] that learns filter responses associated to background and foreground regions in images of leaves. Our classifier is trained by selecting positive (leaf) and negative (non-leaf) feature samples that lie on the neighboorhod of the leaf boundary thus focusing learning only on "sensitive" pixels. Such classifier is then applied to each pixel location of a given unknown test image. This provides a score map that we then threshold using two different thresholds to detect pixels that belong to foreground and background with a high level of probability. With these pixels at hand we initialize an EM algorithm with a good initial estimate of foreground and background cluster parameters in the saturation-value color space, differently from [1] which has to initialize the EM segmentation with the same values for all the images. The other difference with [1] is that we can consider as unlabeled data only the pixels that are in the neighborhood of the detected leaf boundary. This allows to keep focusing on segmenting correctly the pixels around the leaf boundary, and in practice it is enough to get a good segmentation of the other pixels, which are easier to classify. For evaluation we use the Leafsnap Field image dataset publicly available online [1] where different leaves of different species are acquired against solid background and variable light conditions thus simulating typical images that a user could provide for plant recognition. To train our pixel-wise classifier we randomly select one image for each species and we manually produce segmentation and thicker contours to discriminate between positive and negative training samples placed in the neighborhood of boundary. Since segmentation ground truth is not available and its manual production for thousands of images would require an inestimable amount of time, we considered a subset of the original Field dataset. Our testing set is made of 300 images: 150 images for which the EM approach of [1] performs already well thus producing faithful segmentation in accordance with the leaf shape plus 150 more challenging images for which EM partially or totally fails. The general behavior of different methods can be qualitatively appreciated looking at Fig. 1 where results returned by Leafsnap, Leafsnap without post-processing (marked with *), GrabCut and our method are reported. As the reader can see comparing ground truth details with real segmentations, it is confirmed that post-processing hurts quality of (a) Leaf image (b) Ground truth

[1]  David W. Jacobs,et al.  Efficient segmentation of leaves in semi-controlled conditions , 2013, Machine Vision and Applications.

[2]  Sean White,et al.  First steps toward an electronic field guide for plants , 2006 .

[3]  Sean White,et al.  Searching the World's Herbaria: A System for Visual Identification of Plant Species , 2008, ECCV.

[4]  Berrin A. Yanikoglu,et al.  Automatic plant identification from photographs , 2014, Machine Vision and Applications.

[5]  Trevor Hastie,et al.  The Elements of Statistical Learning , 2001 .

[6]  Laure Tougne,et al.  Tree Leaves Extraction in Natural Images: Comparative Study of Preprocessing Tools and Segmentation Methods , 2015, IEEE Transactions on Image Processing.

[7]  Jianbo Shi,et al.  Spectral segmentation with multiscale graph decomposition , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[8]  Laure Tougne,et al.  ReVeS Participation - Tree Species Classification Using Random Forests and Botanical Features. , 2012 .

[9]  Berrin A. Yanikoglu,et al.  Sabanci-Okan System at ImageClef 2012: Combining Features and Classifiers for Plant Identification , 2012, CLEF.

[10]  Vincent Lepetit,et al.  Multiscale Centerline Detection by Learning a Scale-Space Distance Transform , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[11]  Itheri Yahiaoui,et al.  Interactive plant identification based on social image data , 2014, Ecol. Informatics.

[12]  Dorin Comaniciu,et al.  Mean Shift: A Robust Approach Toward Feature Space Analysis , 2002, IEEE Trans. Pattern Anal. Mach. Intell..

[13]  Daniel P. Huttenlocher,et al.  Efficient Graph-Based Image Segmentation , 2004, International Journal of Computer Vision.

[14]  Nozha Boujemaa,et al.  The ImageCLEF 2012 Plant Identification Task , 2012, CLEF.

[15]  Tolga Tasdizen,et al.  Image Segmentation with Cascaded Hierarchical Models and Logistic Disjunctive Normal Networks , 2013, 2013 IEEE International Conference on Computer Vision.

[16]  Arnab Bhattacharya,et al.  A Plant Identification System using Shape and Morphological Features on Segmented Leaflets: Team IITK, CLEF 2012 , 2012, CLEF.

[17]  P. Cochat,et al.  Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.

[18]  Jeff A. Bilmes,et al.  A gentle tutorial of the em algorithm and its application to parameter estimation for Gaussian mixture and hidden Markov models , 1998 .

[19]  W. Marsden I and J , 2012 .

[20]  Charless C. Fowlkes,et al.  Contour Detection and Hierarchical Image Segmentation , 2011, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[21]  Andrew Blake,et al.  "GrabCut" , 2004, ACM Trans. Graph..

[22]  Hervé Glotin,et al.  LifeCLEF 2014: Multimedia Life Species Identification Challenges , 2014, CLEF.

[23]  Laure Tougne,et al.  Understanding leaves in natural images - A model-based approach for tree species identification , 2013, Comput. Vis. Image Underst..

[24]  W. John Kress,et al.  Leafsnap: A Computer Vision System for Automatic Plant Species Identification , 2012, ECCV.

[25]  Odemir Martinez Bruno,et al.  IFSC/USP at ImageCLEF 2011: Plant Identication Task , 2011, CLEF.

[26]  Ronen Basri,et al.  Hierarchy and adaptivity in segmenting visual scenes , 2006, Nature.

[27]  Yuxuan Wang,et al.  A Leaf Recognition Algorithm for Plant Classification Using Probabilistic Neural Network , 2007, 2007 IEEE International Symposium on Signal Processing and Information Technology.

[28]  N. Otsu A threshold selection method from gray level histograms , 1979 .

[29]  Oskar Söderkvist,et al.  Computer Vision Classification of Leaves from Swedish Trees , 2001 .

[30]  Elaine G. Toms,et al.  Information Access Evaluation. Multilinguality, Multimodality, and Interaction , 2014 .

[31]  David Jones,et al.  Individual leaf extractions from young canopy images using Gustafson-Kessel clustering and a genetic algorithm , 2006 .