Fast Extraction Method of High-Level Feature Using Random Forests from Imbalanced Training Data

We propose a method of using a random forest algorithm to quickly detect semantically high-level features such as specific objects. The random forest has a lower computation cost than that of the common algorithm such as a support vector machine (SVM). However, it cannot cope with training data that have a large bias in the number of negative and positive examples. We improve the conventional training algorithm to ensure sampling the data with equal probability from each class when creating bootstrap samples, which increases the classification accuracy. Experiments on the Caltech-101 dataset resulted in a recall of 64.3% and precision of 71.1%, which were comparable to those of conventional methods. The average time needed for training and for detection were reduced to one sixteenth and one twenty-seventh that of SVM, respectively.

[1]  Cordelia Schmid,et al.  Beyond Bags of Features: Spatial Pyramid Matching for Recognizing Natural Scene Categories , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).

[2]  Matti Pietikäinen,et al.  Multiresolution Gray-Scale and Rotation Invariant Texture Classification with Local Binary Patterns , 2002, IEEE Trans. Pattern Anal. Mach. Intell..

[3]  Gabriela Csurka,et al.  Visual categorization with bags of keypoints , 2002, eccv 2004.

[4]  Cordelia Schmid,et al.  Local Grayvalue Invariants for Image Retrieval , 1997, IEEE Trans. Pattern Anal. Mach. Intell..

[5]  David G. Lowe,et al.  Object recognition from local scale-invariant features , 1999, Proceedings of the Seventh IEEE International Conference on Computer Vision.

[6]  Duy-Dinh Le,et al.  NII-ISM, Japan at TRECVID 2007: High Level Feature Extraction , 2007, TRECVID.

[7]  Jitendra Malik,et al.  SVM-KNN: Discriminative Nearest Neighbor Classification for Visual Category Recognition , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).