Translation Templates for Object Matching Across Predictable Pose Variation

Computer vision is the most promising technology for automated, passive tracking of multiple objects over large areas. Effecti ve inter-camera and intra-camera visual tracking can enable information about a vehicle or a pedestrian to be integrated from various sources. Unfortunately, tracking objects across multiple non-overlapping cameras requires reliable comparison of the objects’ appearance under widely-varying view an gles and resolutions. Fortunately, in most cases, an object of a particul ar type entering a scene at a particular position and direction will tend to be i n a very similar pose. This paper introduces Translation Templates (TTs). TTs exploit this regularity to learn a color-based matching metric for images from a pair of tracking source and sink points, without prior knowledge of object type or object pose. This model benefits from histogram-based aggre gation while still preserving spatial relationships between the two ima ges. The model can be learned directly from data and used to compare arbitrary types of objects observed from extremely different viewpoints, as long as the relationship between the viewpoints is preserved. This paper describes TTs, describes a method for efficient computation and for visualization of TT s, and presents experimental results from both an indoor pedestrian data set and an outdoor vehicle data set.

[1]  Chris Stauffer Learning a Probabilistic Similarity Function for Segmentation , 2004, 2004 Conference on Computer Vision and Pattern Recognition Workshop.

[2]  Leonidas J. Guibas,et al.  A metric for distributions with applications to image databases , 1998, Sixth International Conference on Computer Vision (IEEE Cat. No.98CH36271).

[3]  Jitendra Malik,et al.  Shape Matching and Object Recognition , 2006, Toward Category-Level Object Recognition.

[4]  Daniel P. Huttenlocher,et al.  A new Bayesian framework for object recognition , 1999, Proceedings. 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149).

[5]  Ying Shan,et al.  Unsupervised learning of discriminative edge measures for vehicle matching between non-overlapping cameras , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[6]  Mubarak Shah,et al.  Appearance modeling for tracking in multiple non-overlapping cameras , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[7]  William Rucklidge,et al.  Efficient Visual Recognition Using the Hausdorff Distance , 1996, Lecture Notes in Computer Science.

[8]  L. Davis,et al.  M2Tracker: A Multi-View Approach to Segmenting and Tracking People in a Cluttered Scene , 2003, International Journal of Computer Vision.

[9]  Raimondo Schettini,et al.  A survey on methods for colour image indexing and retrieval in image databases , 2002 .

[10]  Alan L. Yuille,et al.  Feature extraction from faces using deformable templates , 1989, Proceedings CVPR '89: IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[11]  W. J. Rucklidge E?cient Computation of the Minimum Hausdorfi Distance for Visual Recognition , 1994 .