Improving classication using a Condence Matrix based on weak classiers applied to OCR

This paper proposes a new feature representation method based on the construction of a Condence Matrix (CM). This representation consists of posterior probability values provided by several weak classiers, each one trained and used in dierent sets of features from the original sample. The CM allows the nal classier to abstract itself from discovering underlying groups of features. In this work the CM is applied to isolated character image recognition, for which several set of features can be extracted from each sample. Experimentation has shown that the use of the CM permits a signicant improvement in accuracy in most cases, while the others remain the same. The results were obtained after experimenting with four well-known corpora, using evolved meta-classiers with the k-Nearest Neighbor rule as weak classier and by applying statistical signicance tests.

[1]  Vladimir Vapnik,et al.  Statistical learning theory , 1998 .

[2]  Anil K. Jain,et al.  Artificial Neural Networks: A Tutorial , 1996, Computer.

[3]  Juan José Rodríguez Diez,et al.  Rotation Forest: A New Classifier Ensemble Method , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[4]  Arun Ross,et al.  Score normalization in multimodal biometric systems , 2005, Pattern Recognit..

[5]  J. Ross Quinlan,et al.  C4.5: Programs for Machine Learning , 1992 .

[6]  Robert P. W. Duin,et al.  The Dissimilarity Representation for Pattern Recognition - Foundations and Applications , 2005, Series in Machine Perception and Artificial Intelligence.

[7]  Masaki Nakagawa,et al.  A Compact On-line and Off-line Combined Recognizer , 2006 .

[8]  Ian H. Witten,et al.  The WEKA data mining software: an update , 2009, SKDD.

[9]  Ulrich Eckhardt,et al.  Shape descriptors for non-rigid shapes with a single closed contour , 2000, Proceedings IEEE Conference on Computer Vision and Pattern Recognition. CVPR 2000 (Cat. No.PR00662).

[10]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[11]  Dong Liu,et al.  Robust late fusion with rank minimization , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[12]  Ernest Valveny,et al.  Optimal Classifier Fusion in a Non-Bayesian Probabilistic Framework , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[13]  Jean Serra,et al.  Image Analysis and Mathematical Morphology , 1983 .

[14]  Alessandro Lameiras Koerich,et al.  Modeling segmentation cuts using support vector machines , 2006 .

[15]  Alexander K. Seewald,et al.  How to Make Stacking Better and Faster While Also Taking Care of an Unknown Weakness , 2002, International Conference on Machine Learning.

[16]  Rafael Llobet,et al.  Fast and Accurate Handwritten Character Recognition Using Approximate Nearest Neighbours Search on Large Databases , 2000, SSPR/SPR.

[17]  N. Otsu A threshold selection method from gray level histograms , 1979 .

[18]  Tony R. Martinez,et al.  Improved Heterogeneous Distance Functions , 1996, J. Artif. Intell. Res..

[19]  Jiri Matas,et al.  On Combining Classifiers , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[20]  Jonathan J. Hull,et al.  A Database for Handwritten Text Recognition Research , 1994, IEEE Trans. Pattern Anal. Mach. Intell..

[21]  Robert P. W. Duin,et al.  A Combine-Correct-Combine Scheme for Optimizing Dissimilarity-Based Classifiers , 2009, CIARP.

[22]  Frank Rosenblatt,et al.  PRINCIPLES OF NEURODYNAMICS. PERCEPTRONS AND THE THEORY OF BRAIN MECHANISMS , 1963 .

[23]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.