Ambiguity in Structure from Motion: Sphere versus Plane

If 3D rigid motion can be correctly estimated from image sequences, the structure of the scene can be correctly derived using the equations for image formation. However, an error in the estimation of 3D motion will result in the computation of a distorted version of the scene structure. Of computational interest are these regions in space where the distortions are such that the depths become negative, because in order for the scene to be visible it has to lie in front of the image, and thus the corresponding depth estimates have to be positive. The stability analysis for the structure from motion problem presented in this paper investigates the optimal relationship between the errors in the estimated translational and rotational parameters of a rigid motion that results in the estimation of a minimum number of negative depth values. The input used is the value of the flow along some direction, which is more general than optic flow or correspondence. For a planar retina it is shown that the optimal configuration is achieved when the projections of the translational and rotational errors on the image plane are perpendicular. Furthermore, the projection of the actual and the estimated translation lie on a line through the center. For a spherical retina, given a rotational error, the optimal translation is the correct one; given a translational error, the optimal rotational negative deptherror depends both in direction and value on the actual and estimated translation as well as the scene in view. The proofs, besides illuminating the confounding of translation and rotation in structure from motion, have an important application to ecological optics. The same analysis provides a computational explanation of why it is easier to estimate self-motion in the case of a spherical retina and why shape can be estimated easily in the case of a planar retina, thus suggesting that nature's design of compound eyes (or panoramic vision) for flying systems and camera-type eyes for primates (and other systems that perform manipulation) is optimal.

[1]  J. van Santen,et al.  Temporal covariance model of human motion perception. , 1984, Journal of the Optical Society of America. A, Optics and image science.

[2]  Thomas S. Huang,et al.  Theory of Reconstruction from Image Motion , 1992 .

[3]  Gilad Adiv,et al.  Inherent Ambiguities in Recovering 3-D Motion and Structure from a Noisy Flow Field , 1989, IEEE Trans. Pattern Anal. Mach. Intell..

[4]  Thomas S. Huang,et al.  Uniqueness and Estimation of Three-Dimensional Motion Parameters of Rigid Objects with Curved Surfaces , 1984, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[5]  Shahriar Negahdaripour,et al.  IEEE Transactions on Pattern Analysis and Machine Intelligence , 2004 .

[6]  H. Damasio,et al.  IEEE Transactions on Pattern Analysis and Machine Intelligence: Special Issue on Perceptual Organization in Computer Vision , 1998 .

[7]  Berthold K. P. Horn Relative orientation , 1987, International Journal of Computer Vision.

[8]  Andrew Blake,et al.  Robust estimation of egomotion from normal flow , 1994, International Journal of Computer Vision.

[9]  Minas E. Spetsakis,et al.  Models of statistical visual motion estimation , 1994 .

[10]  Yoshiaki Shirai,et al.  Three-Dimensional Computer Vision , 1987, Symbolic Computation.

[11]  Yiannis Aloimonos,et al.  What Is Computed by Structure from Motion Algorithms? , 1998, ECCV.

[12]  Richard I. Hartley,et al.  Projective Reconstruction and Invariants from Multiple Images , 1994, IEEE Trans. Pattern Anal. Mach. Intell..

[13]  Andrea J. van Doorn,et al.  Relief: pictorial and otherwise , 1995, Image Vis. Comput..

[14]  Berthold K. P. Horn,et al.  Passive navigation , 1982, Comput. Vis. Graph. Image Process..

[15]  Andrea J. van Doorn,et al.  Relief: pictorial and otherwise , 1995, Image Vis. Comput..

[16]  Soren W. Henriksen,et al.  Manual of photogrammetry , 1980 .

[17]  T. Poggio,et al.  Considerations on models of movement detection , 1973, Kybernetik.

[18]  O. Faugeras Three-dimensional computer vision: a geometric viewpoint , 1993 .

[19]  Yiannis Aloimonos,et al.  Estimating the heading direction using normal flow , 1994, International Journal of Computer Vision.

[20]  J HeegerDavid,et al.  Subspace methods for recovering rigid motion I , 1992 .

[21]  Loong Fah Cheong,et al.  Visual space distortion , 1997, Biological Cybernetics.

[22]  Olivier D. Faugeras,et al.  What can be seen in three dimensions with an uncalibrated stereo rig , 1992, ECCV.

[23]  Olivier D. Faugeras,et al.  On the geometry and algebra of the point and line correspondences between N images , 1995, Proceedings of IEEE International Conference on Computer Vision.

[24]  Yiannis Aloimonos,et al.  Directions of Motion Fields are Hardly Ever Ambiguous , 2004, International Journal of Computer Vision.

[25]  J J Koenderink,et al.  Affine structure from motion. , 1991, Journal of the Optical Society of America. A, Optics and image science.

[26]  S. Ullman,et al.  The interpretation of visual motion , 1977 .

[27]  Minas E. Spetsakis,et al.  Structure from motion using line correspondences , 1990, International Journal of Computer Vision.

[28]  Loong Fah Cheong,et al.  Effects of Errors in the Viewing Geometry on Shape Estimation , 1998, Comput. Vis. Image Underst..

[29]  J. Aloimonos,et al.  Optimal motion estimation , 1989, [1989] Proceedings. Workshop on Visual Motion.

[30]  Stephen J. Maybank,et al.  Algorithm for analysing optical flow based on the least-squares method , 1986, Image Vis. Comput..

[31]  H. C. Longuet-Higgins,et al.  The interpretation of a moving retinal image , 1980, Proceedings of the Royal Society of London. Series B. Biological Sciences.

[32]  Berthold K. P. Horn Robot vision , 1986, MIT electrical engineering and computer science series.

[33]  Kostas Daniilidis,et al.  Understanding noise sensitivity in structure from motion , 1996 .

[34]  H. C. Longuet-Higgins,et al.  A computer algorithm for reconstructing a scene from two projections , 1981, Nature.

[35]  Yiannis Aloimonos,et al.  On the Geometry of Visual Correspondence , 1997, International Journal of Computer Vision.

[36]  Berthold K. P. Horn,et al.  Direct methods for recovering motion , 1988, International Journal of Computer Vision.

[37]  Minas E. Spetsakis,et al.  Optimal Computing Of Structure From Motion Using Point Correspondences In Two Frames , 1988, [1988 Proceedings] Second International Conference on Computer Vision.

[38]  J. Aloimonos,et al.  Finding motion parameters from spherical motion fields (or the advantages of having eyes in the back of your head) , 1988, Biological Cybernetics.

[39]  Y. Aloimonos,et al.  Direct Perception of Three-Dimensional Motion from Patterns of Visual Motion , 1995, Science.

[40]  Berthold K. P. Horn Motion fields are hardly ever ambiguous , 1988, International Journal of Computer Vision.

[41]  Yiannis Aloimonos,et al.  Qualitative egomotion , 1995, International Journal of Computer Vision.

[42]  Andrea J. van Doorn,et al.  Two-plus-one-dimensional differential geometry , 1994, Pattern Recognition Letters.

[43]  Karl Johan Åström,et al.  Invariancy Methods for Points, Curves and Surfaces in Computational Vision , 1996 .

[44]  W. Reichardt,et al.  Autocorrelation, a principle for the evaluation of sensory information by the central nervous system , 1961 .

[45]  Allen M. Waxman,et al.  Contour Evolution, Neighborhood Deformation, and Global Image Flow: Planar Surfaces in Motion , 1985 .

[46]  Werner Reichardt,et al.  Evaluation of optical motion information by movement detectors , 1987, Journal of Comparative Physiology A.