Semantic Linking Maps for Active Visual Object Search

We aim for mobile robots to function in a variety of common human environments. Such robots need to be able to reason about the locations of previously unseen target objects. Landmark objects can help this reasoning by narrowing down the search space significantly. More specifically, we can exploit background knowledge about common spatial relations between landmark and target objects. For example, seeing a table and knowing that cups can often be found on tables aids the discovery of a cup. Such correlations can be expressed as distributions over possible pairing relationships of objects. In this paper, we propose an active visual object search strategy method through our introduction of the Semantic Linking Maps (SLiM) model. SLiM simultaneously maintains the belief over a target object’s location as well as landmark objects’ locations, while accounting for probabilistic inter-object spatial relations. Based on SLiM, we describe a hybrid search strategy that selects the next best view pose for searching for the target object based on the maintained belief. We demonstrate the efficiency of our SLiM-based search strategy through comparative experiments in simulated environments. We further demonstrate the realworld applicability of SLiM-based search in scenarios with a Fetch mobile manipulation robot.

[1]  Patric Jensfelt,et al.  Active Visual Object Search in Unknown Environments Using Uncertain Semantics , 2013, IEEE Transactions on Robotics.

[2]  David Hsu,et al.  Act to See and See to Act: POMDP planning for objects search in clutter , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[3]  Masayuki Inaba,et al.  Searching objects in large-scale indoor environments: A decision-theoretic approach , 2012, 2012 IEEE International Conference on Robotics and Automation.

[4]  Joris M. Mooij,et al.  libDAI: A Free and Open Source C++ Library for Discrete Approximate Inference in Graphical Models , 2010, J. Mach. Learn. Res..

[5]  Oliver Brock,et al.  Prior-assisted propagation of spatial information for object search , 2014, 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[6]  James J. Little,et al.  Curious George: An Integrated Visual Search Platform , 2010, 2010 Canadian Conference on Computer and Robot Vision.

[7]  Lucas Beyer,et al.  The STRANDS Project: Long-Term Autonomy in Everyday Environments , 2016, IEEE Robotics Autom. Mag..

[8]  Patric Jensfelt,et al.  Topological spatial relations for active visual search , 2012, Robotics Auton. Syst..

[9]  Dieter Fox,et al.  CRF-Filters: Discriminative Particle Filters for Sequential State Estimation , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[10]  Odest Chadwicke Jenkins,et al.  Semantic Mapping with Simultaneous Object Detection and Localization , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[11]  Anil K. Jain,et al.  Unsupervised Learning of Finite Mixture Models , 2002, IEEE Trans. Pattern Anal. Mach. Intell..

[12]  Stephanie Rosenthal,et al.  CoBots: Robust Symbiotic Autonomous Mobile Service Robots , 2015, IJCAI.

[13]  Lambert E. Wixson,et al.  Using intermediate objects to improve the efficiency of visual search , 1994, International Journal of Computer Vision.

[14]  James J. Little,et al.  Automated Spatial-Semantic Modeling with Applications to Place Labeling and Informed Search , 2009, 2009 Canadian Conference on Computer and Robot Vision.

[15]  John Folkesson,et al.  Search in the real world: Active visual object search based on spatial relations , 2011, 2011 IEEE International Conference on Robotics and Automation.

[16]  Sonia Chernova,et al.  Temporal persistence modeling for object search , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[17]  Christopher Amato,et al.  Online Planning for Target Object Search in Clutter under Partial Observability , 2019, 2019 International Conference on Robotics and Automation (ICRA).

[18]  Javier Ruiz-del-Solar,et al.  A Bayesian based Methodology for Indirect Object Search , 2018, J. Intell. Robotic Syst..

[19]  Danica Kragic,et al.  Object Search and Localization for an Indoor Mobile Robot , 2009, J. Comput. Inf. Technol..

[20]  Pietro Perona,et al.  Microsoft COCO: Common Objects in Context , 2014, ECCV.

[21]  Patric Jensfelt,et al.  Object search on a mobile robot using relational spatial information , 2010 .

[22]  Jiyu Cheng,et al.  Efficient Object Search With Belief Road Map Using Mobile Robot , 2018, IEEE Robotics and Automation Letters.

[23]  Brendan J. Frey,et al.  Factor graphs and the sum-product algorithm , 2001, IEEE Trans. Inf. Theory.

[24]  Nick Hawes,et al.  Using Qualitative Spatial Relations for indirect object search , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[25]  Benjamin Kuipers,et al.  Robot navigation with model predictive equilibrium point control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[26]  P. Cochat,et al.  Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.

[27]  T. Garvey Perceptual strategies for purposive vision , 1975 .

[28]  Jos Elfring,et al.  Active Object Search Exploiting Probabilistic Object-Object Relations , 2013, RoboCup.

[29]  Leslie Pack Kaelbling,et al.  Manipulation-based active search for occluded objects , 2013, 2013 IEEE International Conference on Robotics and Automation.

[30]  Kaiming He,et al.  Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[31]  Jake K. Aggarwal,et al.  BWIBots: A platform for bridging the gap between AI and human–robot interaction research , 2017, Int. J. Robotics Res..

[32]  Nicholas Roy,et al.  Utilizing object-object and object-scene context when planning to find things , 2009, 2009 IEEE International Conference on Robotics and Automation.

[33]  Wolfram Burgard,et al.  Searching for objects: Combining multiple cues to object locations using a maximum entropy model , 2010, 2010 IEEE International Conference on Robotics and Automation.

[34]  John K. Tsotsos,et al.  Visual search for an object in a 3D environment using a mobile robot , 2010, Comput. Vis. Image Underst..