Cue-based equivalence classes and incremental discrimination for multi-cue recognition of “interactionable” objects

There is a subset of objects for which interaction can provide numerous cues to those objects' identity. Robots are often in situations where they can take advantage being able to observe humans interacting with the objects. In this paper, we define this subset of ‘interactionable’ objects for which we use our Multiple-Cue Object Recognition algorithm (MCOR) to take advantage of using multiple cues. We present two main contributions: 1) the introduction of cue-driven equivalence class discrimination, and 2) the integration of this technique, the general MCOR algorithm, and a hierarchical activity recognition algorithm also presented in this paper, demonstrated on data taken from a static Sony QRIO robot observing a human interacting with objects. The hierarchical activity recognition provides an important cue for the object recognition.

[1]  Volker Graefe,et al.  Dependable multimodal communication and interaction with robotic assistants , 2002, Proceedings. 11th IEEE International Workshop on Robot and Human Interactive Communication.

[2]  Jochen Triesch,et al.  Shared Features for Scalable Appearance-Based Object Recognition , 2005, 2005 Seventh IEEE Workshops on Applications of Computer Vision (WACV/MOTION'05) - Volume 1.

[3]  Thomas Röfer,et al.  Realtime Object Recognition Using Decision Tree Learning , 2004, RoboCup.

[4]  Gertjan J. Burghouts,et al.  Performance evaluation of local colour invariants , 2009, Comput. Vis. Image Underst..

[5]  Paul A. Viola,et al.  Rapid object detection using a boosted cascade of simple features , 2001, Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. CVPR 2001.

[6]  Irfan A. Essa,et al.  Exploiting human actions and object context for recognition tasks , 1999, Proceedings of the Seventh IEEE International Conference on Computer Vision.

[7]  Teodiano Freire Bastos Filho,et al.  Object Recognition for an Agent-Based Controlled Mobile Robot , 1998 .

[8]  Manuela M. Veloso,et al.  FOCUS: a generalized method for object discovery for robots that observe and interact with humans , 2006, HRI '06.

[9]  Jannik Fritsch,et al.  A multi-modal object attention system for a mobile robot , 2005, 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[10]  Georg Hartmann,et al.  Agent-Based Object Recognition , 2000, ICEIS.

[11]  Manuela M. Veloso,et al.  Towards using multiple cues for robust object recognition , 2007, AAMAS '07.

[12]  G LoweDavid,et al.  Distinctive Image Features from Scale-Invariant Keypoints , 2004 .

[13]  Raphaël Marée,et al.  Decision Trees and Random Subwindows for Object Recognition , 2005 .

[14]  J. J. Gibson The theory of affordances , 1977 .

[15]  Antonio Torralba,et al.  Using the Forest to See the Trees: A Graphical Model Relating Features, Objects, and Scenes , 2003, NIPS.

[16]  Manuela M. Veloso,et al.  Simulation and weights of multiple cues for robust object recognition , 2007, 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[17]  Shigeki Aoki,et al.  Scene recognition based on relationship between human actions and objects , 2004, Proceedings of the 17th International Conference on Pattern Recognition, 2004. ICPR 2004..

[18]  Dan Roth,et al.  Learning to detect objects in images via a sparse, part-based representation , 2004, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[19]  Alex Pentland,et al.  Probabilistic visual learning for object detection , 1995, Proceedings of IEEE International Conference on Computer Vision.

[20]  Azriel Rosenfeld,et al.  Recognition by Functional Parts , 1995, Comput. Vis. Image Underst..

[21]  Raúl Rojas,et al.  Tracking regions and edges by shrinking and growing , 2003 .