Automatic detection of informative frames from wireless capsule endoscopy images

Wireless capsule endoscopy (WCE) is a new clinical technology permitting visualization of the small bowel, the most difficult segment of the digestive tract. The major drawback of this technology is the excessive amount of time required for video diagnosis. We therefore propose a method for generating smaller videos by detecting informative frames from original WCE videos. This method isolates useless frames that are highly contaminated by turbid fluids, faecal materials and/or residual foods. These materials and fluids are presented in a wide range of colors, from brown to yellow, and/or have bubble-like texture patterns. The detection scheme therefore consists of two steps: isolating (Step-1) highly contaminated non-bubbled (HCN) frames and (Step-2) significantly bubbled (SB) frames. Two color representations, viz., local color moments in Ohta space and the HSV color histogram, are attempted to characterize HCN frames, which are isolated by a support vector machine (SVM) classifier in Step-1. The rest of the frames go to Step-2, where a Gauss Laguerre transform (GLT) based multiresolution texture feature is used to characterize the bubble structures in WCE frames. GLT uses Laguerre Gauss circular harmonic functions (LG-CHFs) to decompose WCE images into multiresolution components. An automatic method of segmentation was designed to extract bubbled regions from grayscale versions of the color images based on the local absolute energies of their CHF responses. The final informative frames were detected by using a threshold on the segmented regions. An automatic procedure for selecting features based on analyzing the consistency of the energy-contrast map is also proposed. Three experiments, two of which use 14,841 and 37,100 frames from three videos and the rest uses 66,582 frames from six videos, were conducted for justifying the proposed method. The two combinations of the proposed color and texture features showed excellent average detection accuracies (86.42% and 84.45%) with the final experiment, when compared with the same color features followed by conventional Gabor-based (78.18% and 76.29%) and discrete wavelet-based (65.43% and 63.83%) texture features. Although intra-video training-testing cases are typical choices for supervised classification in Step-1, combining a suitable number of training sets using a subset of the input videos was shown to be possible. This mixing not only reduced computation costs but also produced better detection accuracies by minimizing visual-selection errors, especially when processing large numbers of WCE videos.

[1]  G. Iddan,et al.  Wireless capsule endoscopy , 2003, Gut.

[2]  Patrick Haffner,et al.  Support vector machines for histogram-based image classification , 1999, IEEE Trans. Neural Networks.

[3]  Oscar Nestares,et al.  Efficient spatial-domain implementation of a multiscale image representation based on Gabor functions , 1998, J. Electronic Imaging.

[4]  Rafael C. González,et al.  Digital image processing using MATLAB , 2006 .

[5]  T. Kanade,et al.  Color information for region segmentation , 1980 .

[6]  Giovanni Jacovitti,et al.  Multiresolution circular harmonic decomposition , 2000, IEEE Trans. Signal Process..

[7]  G. Mercier,et al.  Support vector machines for hyperspectral image classification with spectral-based kernels , 2003, IGARSS 2003. 2003 IEEE International Geoscience and Remote Sensing Symposium. Proceedings (IEEE Cat. No.03CH37477).

[8]  Paul Smaglik,et al.  US climate report underlines local impacts of warming , 2000, Nature.

[9]  Keinosuke Fukunaga,et al.  Introduction to Statistical Pattern Recognition , 1972 .

[10]  Alessandro Neri,et al.  Keypoints Selection in the Gauss Laguerre Transformed Domain , 2006, BMVC.

[11]  Stéphane Mallat,et al.  A Theory for Multiresolution Signal Decomposition: The Wavelet Representation , 1989, IEEE Trans. Pattern Anal. Mach. Intell..

[12]  Michael J. Swain,et al.  Color indexing , 1991, International Journal of Computer Vision.

[13]  Christopher J. C. Burges,et al.  A Tutorial on Support Vector Machines for Pattern Recognition , 1998, Data Mining and Knowledge Discovery.

[14]  Vladimir Vapnik,et al.  Statistical learning theory , 1998 .

[15]  B. S. Manjunath,et al.  Color and texture descriptors , 2001, IEEE Trans. Circuits Syst. Video Technol..

[16]  Michael J. Swain,et al.  Indexing via color histograms , 1990, [1990] Proceedings Third International Conference on Computer Vision.

[17]  Hyoung Joong Kim,et al.  Image retrieval based on color coherence , 1999, Proceedings of IEEE. IEEE Region 10 Conference. TENCON 99. 'Multimedia Technology for Asia-Pacific Information Infrastructure' (Cat. No.99CH37030).

[18]  Duncan Bell,et al.  Stomach, intestine, and colon tissue discriminators for wireless capsule endoscopy images , 2005, SPIE Medical Imaging.

[19]  Fernando Vilariño,et al.  ROC curves and video analysis optimization in intestinal capsule endoscopy , 2006, Pattern Recognit. Lett..

[20]  N. Otsu A threshold selection method from gray level histograms , 1979 .

[21]  Fernando Vilariño,et al.  Automatic Detection of Intestinal Juices in Wireless Capsule Video Endoscopy , 2006, 18th International Conference on Pattern Recognition (ICPR'06).

[22]  Arivazhagan Selvaraj,et al.  Texture classification using wavelet transform , 2003, Pattern Recognit. Lett..

[23]  Sung Min Kim,et al.  Image Retrieval via Query-by-Layout Using MPEG-7 Visual Descriptors , 2007 .

[24]  Chih-Jen Lin,et al.  LIBSVM: A library for support vector machines , 2011, TIST.

[25]  Alessandro Neri,et al.  Maximum likelihood localization of 2-D patterns in the Gauss-Laguerre transform domain: theoretic framework and preliminary results , 2004, IEEE Transactions on Image Processing.

[26]  Markus A. Stricker,et al.  Similarity of color images , 1995, Electronic Imaging.

[27]  Jung-Hwan Oh,et al.  Informative frame classification for endoscopy video , 2007, Medical Image Anal..