Active Image Capturing and Dynamic Scene Visualization by Cooperative Distributed Vision

This paper addresses active image capturing and dynamic scene visualization by Cooperative Distributed Vision (CDV, in short). The concept of CDV was proposed by our five years project starting from 1996. From a practical point of view, the goal of CDV is summarized as follows: Embed in the real world a group of network-connected Observation Stations (real time video image processor with active camera(s)) and mobile robots with vision. And realize 1) wide-area dynamic scene understanding and 2) versatile scene visualization. Applications of CDV include real time wide-area surveillance, remote conference and lecturing systems, interactive 3D TV and intelligent TV studio, navigation of (non-intelligent) mobile robots and disabled people, cooperative mobile robots, and so on. In this paper, we first define the framework of CDV and give a brief retrospective view of the computer vision research to show the background of CDV. Then we present technical research results so far obtained: 1) fixed viewpoint pan-tilt-zoom camera for wide-area active imaging, 2) moving object detection and tracking for reactive image acquisition, 3) multi-viewpoints object imaging by cooperative observation stations, and 4) scenario-based cooperative camera-work planning for dynamic scene visualization. Prototype systems demonstrate the effectiveness and practical utilities of the proposed methods.

[1]  Yiannis Aloimonos,et al.  Purposive, qualitative, active vision , 1992, CVGIP Image Underst..

[2]  David Salesin,et al.  The virtual cinematographer: a paradigm for automatic real-time camera control and directing , 1996, SIGGRAPH.

[3]  Kenji Mase,et al.  Scripting Method Based on Temporal Intervals for Designing Interactive Systems (Special Issue on Next Generation Human Interface and Interaction) , 1998 .

[4]  David Salesin,et al.  Declarative Camera Control for Automatic Cinematography , 1996, AAAI/IAAI, Vol. 1.

[5]  D. Arijon,et al.  Grammar of Film Language , 1976 .

[6]  Paul S. Heckbert,et al.  Creating Raster Omnimax Images from Multiple Perspective Views Using the Elliptical Weighted Average Filter , 1986, IEEE Computer Graphics and Applications.

[7]  Jean-Marc Lavest,et al.  Implicit Reconstruction by Zooming , 1997, Comput. Vis. Image Underst..

[8]  Azriel Rosenfeld,et al.  Visual surveillance and monitoring , 1998 .

[9]  William Buxton,et al.  Beating the limitations of camera-monitor mediated telepresence with extra eyes , 1996, CHI.

[10]  James F. Allen Towards a General Theory of Action and Time , 1984, Artif. Intell..

[11]  Takashi Matsuyama Cooperative Distributed Vision: Dynamic Integration of Visual Perception, Action, and Communication , 1999, DAGM-Symposium.

[12]  Shree K. Nayar,et al.  Generation of Perspective and Panoramic Video from Omnidirectional Video , 1997 .

[13]  Shenchang Eric Chen,et al.  QuickTime VR: an image-based approach to virtual environment navigation , 1995, SIGGRAPH.

[14]  Anup Basu,et al.  Motion Tracking with an Active Camera , 1994, IEEE Trans. Pattern Anal. Mach. Intell..

[15]  Yasushi Yagi,et al.  Obstacle detection with omnidirectional image sensor HyperOmni Vision , 1995, Proceedings of 1995 IEEE International Conference on Robotics and Automation.

[16]  Takashi Matsuyama,et al.  Appearance sphere: background model for pan-tilt-zoom camera , 1996, Proceedings of 13th International Conference on Pattern Recognition.

[17]  Yasushi Yagi,et al.  Real-time generation of environmental map and obstacle avoidance using omnidirectional image sensor with conic mirror , 1991, Proceedings. 1991 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[18]  Ned Greene,et al.  Environment Mapping and Other Applications of World Projections , 1986, IEEE Computer Graphics and Applications.

[19]  Takashi Matsuyama,et al.  Depth measurement by the multi-focus camera , 1998, Proceedings. 1998 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No.98CB36231).