What is That? Object Recognition from Natural Features on a Mobile Phone

Connecting the physical and the digital world is an upcoming trend that enables numberless use-cases. The mobile phone as the most pervasive digital device is often used to establish such a connection. The phone enables users to retrieve, use, and share digital information and services connected to physical objects. Recognizing physical objects is thus a fundamental precondition for such applications. Object recognition is usually enabled using visual marker (e.g. Qr Codes) or electronic marker (e.g. RFID). Marker based approaches are not feasible for a large range of objects such as sights, photos, and persons. Markerless approaches that use the image stream from the mobile phone’s camera are commonly server-based which dramatically limits the interactiveness. Recent work on image processing shows that interactive object recognition on mobile phones is at hand. In this paper we present a markerless object recognition that processes multiple camera images per second on recent mobile phones. The algorithm combines a stripped down SIFT with a scalable vocabulary tree and a simple feature matching. Based on this algorithm we implemented a simple application which recognizes poster segments and conducted an initial user study to get an understanding of the implications that accompany markerless interaction.

[1]  Tom Drummond,et al.  Machine Learning for High-Speed Corner Detection , 2006, ECCV.

[2]  Oliver Bimber,et al.  PhoneGuide: museum guidance supported by on-device object recognition on mobile phones , 2005, MUM '05.

[3]  Roy Want,et al.  Bridging physical and virtual worlds with electronic tags , 1999, CHI '99.

[4]  Qiong Liu,et al.  Mobile camera supported document redirection , 2006, MM '06.

[5]  Keith Cheverst,et al.  Understanding the role of image recognition in mobile tour guides , 2005, Mobile HCI.

[6]  Berna Erol,et al.  Linking presentation documents using image analysis , 2003, The Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, 2003.

[7]  Michael Rohs,et al.  USING CAMERA-EQUIPPED MOBILE PHONES FOR INTERACTING WITH REAL-WORLD OBJECTS , 2004 .

[8]  George W. Fitzmaurice,et al.  Situated information spaces and spatially aware palmtop computers , 1993, CACM.

[9]  Andrew W. Moore,et al.  An Investigation of Practical Approximate Nearest Neighbor Algorithms , 2004, NIPS.

[10]  Niels Henze,et al.  Evaluation of Camera Phone Based Interaction to Access Information Related to Posters , 2008, MIRW.

[11]  Cordelia Schmid,et al.  A performance evaluation of local descriptors , 2005, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[12]  G LoweDavid,et al.  Distinctive Image Features from Scale-Invariant Keypoints , 2004 .

[13]  Niels Henze,et al.  Snap and share your photobooks , 2008, ACM Multimedia.

[14]  Richard Szeliski,et al.  City-Scale Location Recognition , 2007, 2007 IEEE Conference on Computer Vision and Pattern Recognition.

[15]  Yan Xu,et al.  Designing a vision-based mobile interface for in-store shopping , 2008, NordiCHI.

[16]  David Nistér,et al.  Scalable Recognition with a Vocabulary Tree , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).

[17]  Dieter Schmalstieg,et al.  Pose tracking from natural features on mobile phones , 2008, 2008 7th IEEE/ACM International Symposium on Mixed and Augmented Reality.

[18]  Xing Xie,et al.  Inquiring of the Sights from the Web via Camera Mobiles , 2006, 2006 IEEE International Conference on Multimedia and Expo.