Best-next-view algorithm for three-dimensional scene reconstruction using range images

The primary focus of the research detailed in this paper is to develop an intelligent sensing module capable of automatically determining the optimal next sensor position and orientation during scene reconstruction. To facilitate a solution to this problem, we have assembled a system for reconstructing a 3D model of an object or scene from a sequence of range images. Candidates for the best-next-view position are determined by detecting and measuring occlusions to the range camera's view in an image. Ultimately, the candidate which will reveal the greatest amount of unknown scene information is selected as the best-next-view position. Our algorithm uses ray tracing to determine how much new information a given sensor perspective will reveal. We have tested our algorithm successfully on several synthetic range data streams, and found the system's results to be consistent with an intuitive human search. The models recovered by our system from range data compared well with the ideal models. Essentially, we have proven that range information of physical objects can be employed to automatically reconstruct a satisfactory dynamic 3D computer model at a minimal computational expense. This has obvious implications in the contexts of robot navigation, manufacturing, and hazardous materials handling. The algorithm we developed takes advantage of no a priori information in finding the best-next-view position.

[1]  Roger Y. Tsai,et al.  Viewpoint planning: the visibility constraint , 1989 .

[2]  Paul J. Besl,et al.  Surfaces in Range Image Understanding , 1988, Springer Series in Perception Engineering.

[3]  Richard A. Volz,et al.  Range image segmentation and surface parameter extraction for 3-D object recognition of industrial parts , 1987, Proceedings. 1987 IEEE International Conference on Robotics and Automation.

[4]  Anil K. Jain,et al.  On reliable curvature estimation , 1989, CVPR.

[5]  Aviv Bergman,et al.  Determining the camera and light source location for a visual task , 1989, Proceedings, 1989 International Conference on Robotics and Automation.

[6]  Mongi A. Abidi,et al.  Parallel range data processing: a real case study , 1992, Other Conferences.

[7]  John G. Proakis,et al.  Probability, random variables and stochastic processes , 1985, IEEE Trans. Acoust. Speech Signal Process..

[8]  Fumio Kishino,et al.  Active camera controlling for manipulation , 1991, Proceedings. 1991 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[9]  Mongi A. Abidi,et al.  Data fusion in robotics and machine intelligence , 1992 .

[10]  E. Dubois,et al.  Digital picture processing , 1985, Proceedings of the IEEE.

[11]  Thomas O. Binford,et al.  Survey of Model-Based Image Analysis Systems , 1982 .

[12]  C. Ian Connolly,et al.  The determination of next best views , 1985, Proceedings. 1985 IEEE International Conference on Robotics and Automation.

[13]  W. Brent Seales,et al.  Viewpoint from occluding contour , 1992, CVGIP Image Underst..

[14]  James F. Blinn,et al.  Where am I? What am I looking at? (cinematography) , 1988, IEEE Computer Graphics and Applications.

[15]  Ruzena Bajcsy,et al.  Occlusions as a Guide for Planning the Next View , 1993, IEEE Trans. Pattern Anal. Mach. Intell..

[16]  D. Ballard,et al.  Gaze Selection for Visual Search , 1994 .