State-space search for high-level control of machine vision

Computer vision is a task of information processing that can be modeled as a sequence of subtasks. A complete vision process can be constructed by synthesizing individual operators performing the subtasks. Previous work in computer vision has emphasized the development of individual operators for a specific subtask. However, the lack of knowledge about other levels of processing, while developing the operators for a specific level, makes the development of a robust operator and thus a robust system unlikely. To obtain vision problem-solving methods that are robust in the face of variations in image lighting, arrangements of objects, viewing parameters, etc., we can simply incorporate all possible sequences of image-processing operators, each of which deals with a specific situation of input images; then an adaptive control mechanism such as a state-space search procedure can be built into the methods. Such a procedure dynamically determines an optimal sequence of image-processing operators to classify an image or to put its parts into correspondence with a model or set of models. One critical problem in solving vision problems with a state-space search model is how to decide the costs of paths. This paper details the state-space search model of computer vision as well as the design of cost functions in terms of information distortions. A vision system, VISTAS, has been constructed under the state-space search model and its parallel version has been simulated.

[1]  Ze-Nian Li Comparisons of reasoning mechanisms for computer vision , 1988, Int. J. Approx. Reason..

[2]  Keki B. Irani,et al.  Parallel A^* and AO^* Algorithms: An Optimality Criterion and Performance Evaluation , 1986, ICPP.

[3]  Joan S. Weszka,et al.  A survey of threshold selection techniques , 1978 .

[4]  S. L. Tanimoto,et al.  Parallel coordination of image operators based on shared-memory architecture , 1990, [1990] Proceedings. 10th International Conference on Pattern Recognition.

[5]  A. Rosenfeld,et al.  Computer vision: basic principles , 1988, Proc. IEEE.

[6]  Tomás Lozano-Pérez,et al.  Tactile Recognition and Localization Using Object Models: The Case of Polyhedra on a Plane , 1984, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[7]  Steven L. Tanimoto Machine vision as state-space search , 1988 .

[8]  Dana H. Ballard,et al.  Computer Vision , 1982 .

[9]  Steven L. Tanimoto,et al.  Synthesis of vision algorithms based on state-space search , 1989 .

[10]  Claude L. Fennema,et al.  Scene Analysis Using Regions , 1970, Artif. Intell..

[11]  Allen Newell,et al.  SOAR: An Architecture for General Intelligence , 1987, Artif. Intell..

[12]  Alberto Martelli,et al.  An application of heuristic search methods to edge and contour detection , 1976, CACM.

[13]  W. Grimson,et al.  Model-Based Recognition and Localization from Sparse Range or Tactile Data , 1984 .

[14]  R. Haralick Edge and region analysis for digital image data , 1980 .

[15]  Richard O. Duda,et al.  Pattern classification and scene analysis , 1974, A Wiley-Interscience publication.

[16]  John E. Laird,et al.  A universal weak method , 1993 .

[17]  T. Garvey Perceptual strategies for purposive vision , 1975 .

[18]  Robert M. Haralick,et al.  Structural Descriptions and Inexact Matching , 1981, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[19]  Jerome A. Feldman,et al.  Decision Theory and Artificial Intelligence: I. A Semantics-Based Region Analyzer , 1974, Artif. Intell..

[20]  Steven Michael Rubin,et al.  The argos image understanding system. , 1978 .

[21]  Robert C. Bolles,et al.  Verification Vision for Programmable Assembly , 1977, IJCAI.