Finding Text Regions Using Localised Measures

We present a method based on statistical properties of local image neighbourhoods for the location of text in real-scene images. This has applications in robot vision, and desktop and wearable computing. The statistical measures we describe extract properties of the image which characterise text, invariant to a large degree to the orientation, scale or colour of the text in the scene. The measures are employed by a neural network to classify regions of an image as text or non-text. We thus avoid the use of different thresholds for the various situations we expect, including when text is too small to read, or when the text plane is not fronto-parallel to the camera. We briefly discuss applications and the possibility of recovery of the text for optical character recognition.

[1]  J. Barroso,et al.  Number plate reading using computer vision , 1997, ISIE '97 Proceeding of the IEEE International Symposium on Industrial Electronics.

[2]  Edward M. Riseman,et al.  Finding text in images , 1997, DL '97.

[3]  Qian Huang,et al.  Extracting characters of license plates from video sequences , 1998, Machine Vision and Applications.

[4]  Shu-Yuan Chen,et al.  Adaptive page segmentation for color technical journals' cover images , 1998, Image Vis. Comput..

[5]  Motoi Iwata,et al.  Segmentation of Page Images Using the Area Voronoi Diagram , 1998, Comput. Vis. Image Underst..

[6]  Majid Mirmehdi,et al.  Location and recovery of text on oriented surfaces , 1999, Electronic Imaging.

[7]  Stefano Messelodi,et al.  Automatic identification and skew estimation of text lines in real scene images , 1999, Pattern Recognition.

[8]  David S. Doermann,et al.  Automatic text detection and tracking in digital video , 2000, IEEE Trans. Image Process..