MODEEP: a motion-based object detection and pose estimation method for airborne FLIR sequences

Abstract. In this paper, we present a method called MODEEP (Motion-based Object DEtection and Estimation of Pose) to detect independently moving objects (IMOs) in forward-looking infrared (FLIR) image sequences taken from an airborne, moving platform. Ego-motion effects are removed through a robust multi-scale affine image registration process. Thereafter, areas with residual motion indicate potential object activity. These areas are detected, refined and selected using a Bayesian classifier. The resulting regions are clustered into pairs such that each pair represents one object's front and rear end. Using motion and scene knowledge, we estimate object pose and establish a region of interest (ROI) for each pair. Edge elements within each ROI are used to segment the convex cover containing the IMO. We show detailed results on real, complex, cluttered and noisy sequences. Moreover, we outline the integration of our fast and robust system into a comprehensive automatic target recognition (ATR) and action classification system.

[1]  Andrew Zisserman,et al.  Concerning Bayesian Motion Segmentation, Model, Averaging, Matching and the Trifocal Tensor , 1998, ECCV.

[2]  Jake K. Aggarwal,et al.  On the computation of motion from sequences of images-A review , 1988, Proc. IEEE.

[3]  Edward H. Adelson,et al.  Spatio-temporal segmentation of video data , 1994, Electronic Imaging.

[4]  John F. Canny,et al.  A Computational Approach to Edge Detection , 1986, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[5]  Jake K. Aggarwal,et al.  Multisensory Computer Vision , 1992, Adv. Comput..

[6]  B. Bhanu,et al.  Estimating 3D Egomotion from Perspective Image Sequence , 1990, IEEE Trans. Pattern Anal. Mach. Intell..

[7]  V. Leitáo,et al.  Computer Graphics: Principles and Practice , 1995 .

[8]  P. Anandan,et al.  Hierarchical Model-Based Motion Estimation , 1992, ECCV.

[9]  Jake K. Aggarwal,et al.  A Focused Target Segmentation Paradigm , 1996, ECCV.

[10]  Richard O. Duda,et al.  Pattern classification and scene analysis , 1974, A Wiley-Interscience publication.

[11]  Irwin Edward Sobel,et al.  Camera Models and Machine Perception , 1970 .

[12]  Larry S. Davis,et al.  Detection of independently moving objects in passive video , 1995, Proceedings of the Intelligent Vehicles '95. Symposium.

[13]  Edward H. Adelson,et al.  The Laplacian Pyramid as a Compact Image Code , 1983, IEEE Trans. Commun..

[14]  Azriel Rosenfeld,et al.  Guest Editorial Introduction To The Special Issue On Automatic Target Detection And Recognition , 1997, IEEE Trans. Image Process..

[15]  Berthold K. P. Horn,et al.  Determining Optical Flow , 1981, Other Conferences.

[16]  Gilad Adiv,et al.  Determining Three-Dimensional Motion and Structure from Optical Flow Generated by Several Moving Objects , 1985, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[17]  B. Wandell Foundations of vision , 1995 .

[18]  P. Anandan,et al.  A Unified Approach to Moving Object Detection in 2D and 3D Scenes , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[19]  Michael J. Black,et al.  The Robust Estimation of Multiple Motions: Parametric and Piecewise-Smooth Flow Fields , 1996, Comput. Vis. Image Underst..

[20]  Mark E. Oxley,et al.  Neural networks for automatic target recognition , 1995, Neural Networks.

[21]  M. Carter Computer graphics: Principles and practice , 1997 .

[22]  Q.X. Wu,et al.  A Correlation-Relaxation-Labeling Framework for Computing Optical Flow - Template Matching from a New Perspective , 1995, IEEE Trans. Pattern Anal. Mach. Intell..

[23]  B. D. Guenther,et al.  Aided and automatic target recognition based upon sensory inputs from image forming systems , 1997 .