Video Annotation by Motion Interpretation Using Optical Flow Streams

Abstract A new approach to automatic annotation of video sequences by dominant camera motion interpretation is presented. Unlike other approaches, we separate the optical flow into two categories— singular and non-singular —which, as we show, is a more natural way of classification for the purpose of dominant camera motion interpretation. We show that identification of patterns created by such natural categories, which can be observed from the measured optical flow, can help focus the interpretation of dominant camera motion in video segments. For robust detection of such natural patterns, we propose the computation of optical flow streams (OFS) from the video data and analyze the OFS for the extraction of dominant motion content in the video segments. The advantage of the proposed approach is its robustness in the extraction of dominant motion content in a video segment. We demonstrate this on a variety of real video sequences by generating the automatic motion annotation of the video frames and comparing it with manual motion annotation.

[1]  Kanad K. Biswas,et al.  A cooperative integration of stereopsis and optic flow computation , 1994, Proceedings of 12th International Conference on Pattern Recognition.

[2]  Josef Kittler,et al.  A differential method for simultaneous estimation of rotation, change of scale and translation , 1990, Signal Process. Image Commun..

[3]  Jake K. Aggarwal,et al.  On the computation of motion from sequences of images-A review , 1988, Proc. IEEE.

[4]  Cornelia Fermüller Global 3D motion estimation , 1993, Proceedings of IEEE Conference on Computer Vision and Pattern Recognition.

[5]  Alex Pentland,et al.  Photobook: tools for content-based manipulation of image databases , 1994, Other Conferences.

[6]  Alberto Del Bimbo,et al.  Sequence retrieval by contents through spatio temporal indexing , 1993, Proceedings 1993 IEEE Symposium on Visual Languages.

[7]  John Chung-Mong Lee,et al.  A Robust Approach for Camera Break Detection in Color Video Sequence , 1994, MVA.

[8]  Mubarak Shah,et al.  Motion-based recognition a survey , 1995, Image Vis. Comput..

[9]  A. Verri,et al.  Mathematical properties of the two-dimensional motion field: from singular points to motion parameters , 1989 .

[10]  Mubarak Shah,et al.  Interpretation of Motion Trajectories using Focus of Expansion , 1992, IEEE Trans. Pattern Anal. Mach. Intell..

[11]  Christos Faloutsos,et al.  QBIC project: querying images by content, using color, texture, and shape , 1993, Electronic Imaging.

[12]  Ramesh C. Jain,et al.  Metadata in video databases , 1994, SGMD.

[13]  V. Michael Bove,et al.  Segmentation of frames in a video sequence using motion and other attributes , 1995, Electronic Imaging.

[14]  Hideo Hashimoto,et al.  Video indexing using motion vectors , 1992, Other Conferences.

[15]  H. C. Longuet-Higgins,et al.  The interpretation of a moving retinal image , 1980, Proceedings of the Royal Society of London. Series B. Biological Sciences.

[16]  A. Verri,et al.  Differential techniques for optical flow , 1990 .

[17]  Berthold K. P. Horn,et al.  "Determining optical flow": A Retrospective , 1993, Artif. Intell..

[18]  Svetha Venkatesh,et al.  Qualitative estimation of camera motion parameters from video sequences , 1997, Pattern Recognition.

[19]  Wei Xiong,et al.  Net comparison: a fast and effective method for classifying image sequences , 1995, Electronic Imaging.

[20]  Stephen W. Smoliar,et al.  Content based video indexing and retrieval , 1994, IEEE MultiMedia.

[21]  R. L. Baker,et al.  Global zoom/pan estimation and compensation for video compression , 1991, [Proceedings] ICASSP 91: 1991 International Conference on Acoustics, Speech, and Signal Processing.