Browse-to-Search: Interactive Exploratory Search with Visual Entities

With the development of image search technology, users are no longer satisfied with searching for images using just metadata and textual descriptions. Instead, more search demands are focused on retrieving images based on similarities in their contents (textures, colors, shapes etc.). Nevertheless, one image may deliver rich or complex content and multiple interests. Sometimes users do not sufficiently define or describe their seeking demands for images even when general search interests appear, owing to a lack of specific knowledge to express their intents. A new form of information seeking activity, referred to as exploratory search, is emerging in the research community, which generally combines browsing and searching content together to help users gain additional knowledge and form accurate queries, thereby assisting the users with their seeking and investigation activities. However, there have been few attempts at addressing integrated exploratory search solutions when image browsing is incorporated into the exploring loop. In this work, we investigate the challenges of understanding users' search interests from the images being browsed and infer their actual search intentions. We develop a novel system to explore an effective and efficient way for allowing users to seamlessly switch between browse and search processes, and naturally complete visual-based exploratory search tasks. The system, called Browse-to-Search enables users to specify their visual search interests by circling any visual objects in the webpages being browsed, and then the system automatically forms the visual entities to represent users' underlying intent. One visual entity is not limited by the original image content, but also encapsulated by the textual-based browsing context and the associated heterogeneous attributes. We use large-scale image search technology to find the associated textual attributes from the repository. Users can then utilize the encapsulated visual entities to complete search tasks. The Browse-to-Search system is one of the first attempts to integrate browse and search activities for a visual-based exploratory search, which is characterized by four unique properties: (1) in session—searching is performed during browsing session and search results naturally accompany with browsing content; (2) in context—the pages being browsed provide text-based contextual cues for searching; (3) in focus—users can focus on the visual content of interest without worrying about the difficulties of query formulation, and visual entities will be automatically formed; and (4) intuitiveness—a touch and visual search-based user interface provides a natural user experience. We deploy the Browse-to-Search system on tablet devices and evaluate the system performance using millions of images. We demonstrate that it is effective and efficient in facilitating the user's exploratory search compared to the conventional image search methods and, more importantly, provides users with more robust results to satisfy their exploring experience.

[1]  Xian-Sheng Hua,et al.  Contextual image search , 2011, ACM Multimedia.

[2]  Richard Szeliski,et al.  City-Scale Location Recognition , 2007, 2007 IEEE Conference on Computer Vision and Pattern Recognition.

[3]  Tao Mei,et al.  Contextual Video Recommendation by Multimodal Relevance and User Feedback , 2011, TOIS.

[4]  Lei Zhang,et al.  A Unified Relevance Feedback Framework for Web Image Retrieval , 2009, IEEE Transactions on Image Processing.

[5]  Monica M. C. Schraefel,et al.  Challenges in Supporting Faceted Semantic Browsing of Multimedia Collections , 2007, SAMT.

[6]  Peter Pirolli,et al.  An elementary social information foraging model , 2009, CHI.

[7]  Nenghai Yu,et al.  Complementary hashing for approximate nearest neighbor search , 2011, 2011 International Conference on Computer Vision.

[8]  M. Sheelagh T. Carpendale,et al.  Navigating tomorrow's web: From searching and browsing to visual exploration , 2012, TWEB.

[9]  Ryen W. White,et al.  Supporting exploratory search , 2006 .

[10]  Qiang Yang,et al.  Building bridges for web query classification , 2006, SIGIR.

[11]  Qi Tian,et al.  SIFT match verification by geometric coding for large-scale partial-duplicate web image search , 2013, TOMCCAP.

[12]  Andrew M. Webb,et al.  combinFormation: Mixed-initiative composition of image and text surrogates promotes information discovery , 2008, TOIS.

[13]  Xiao Li,et al.  Understanding the Semantic Structure of Noun Phrase Queries , 2010, ACL.

[14]  Enhong Chen,et al.  Context-aware query classification , 2009, SIGIR.

[15]  Wen Gao,et al.  Towards low bit rate mobile visual search with multiple-channel coding , 2011, ACM Multimedia.

[16]  Stuart K. Card,et al.  The effects of information scent on visual search in the hyperbolic tree browser , 2003, TCHI.

[17]  Qi Tian,et al.  Multimedia search reranking: A literature survey , 2014, CSUR.

[18]  Monica M. C. Schraefel,et al.  Evaluating collaborative information-seeking interfaces with a search-oriented inspection method and re-framed information seeking theory , 2010, Inf. Process. Manag..

[19]  Tao Mei,et al.  CrowdReranking: exploring multiple search engines for visual search reranking , 2009, SIGIR.

[20]  Wai-Tat Fu,et al.  Facilitating exploratory search by model-based navigational cues , 2010, IUI '10.

[21]  Ian H. Witten,et al.  A competitive environment for exploratory query expansion , 2008, JCDL '08.

[22]  Simone Stumpf,et al.  This image smells good: effects of image information scent in search engine results pages , 2011, CIKM '11.

[23]  Meng Wang,et al.  Visual query suggestion , 2009, ACM Multimedia.

[24]  Abdigani Diriye,et al.  Designing a tool for exploratory information seeking , 2012, CHI EA '12.

[25]  David M. Nichols,et al.  That's 'é' not 'þ' '?' or '◓': a user-driven context-aware approach to erroneous metadata in digital libraries , 2011, JCDL '11.

[26]  Monica M. C. Schraefel,et al.  Backward highlighting: enhancing faceted search , 2008, UIST '08.

[27]  Huaiyu Zhu On Information and Sufficiency , 1997 .

[28]  Rongrong Ji,et al.  Active query sensing for mobile location search , 2011, ACM Multimedia.

[29]  W. Bruce Croft,et al.  Query expansion using local and global document analysis , 1996, SIGIR '96.

[30]  G LoweDavid,et al.  Distinctive Image Features from Scale-Invariant Keypoints , 2004 .

[31]  Liqing Zhang,et al.  MindFinder: interactive sketch-based image search on millions of images , 2010, ACM Multimedia.

[32]  Alberto Del Bimbo,et al.  Proceedings of the international conference on Multimedia , 2010 .

[33]  Hao Xu,et al.  Image search by concept map , 2010, SIGIR '10.

[34]  Ravin Balakrishnan,et al.  Keepin' it real: pushing the desktop metaphor with physics, piles and the pen , 2006, CHI.

[35]  Ryen W. White,et al.  Exploratory Search: Beyond the Query-Response Paradigm , 2009, Exploratory Search: Beyond the Query-Response Paradigm.

[36]  Gary Marchionini,et al.  Exploratory search , 2006, Commun. ACM.

[37]  Changsheng Xu,et al.  Interaction Design for Mobile Visual Search , 2013, IEEE Transactions on Multimedia.

[38]  Gary Marchionini,et al.  The Open Video Digital Library , 2002, D Lib Mag..

[39]  Shipeng Li,et al.  Query-driven iterated neighborhood graph search for large scale indexing , 2012, ACM Multimedia.

[40]  David M. Nichols,et al.  Interactive context-aware user-driven metadata correction in digital libraries , 2012, International Journal on Digital Libraries.

[41]  Luc Van Gool,et al.  Hello neighbor: Accurate object retrieval with k-reciprocal nearest neighbors , 2011, CVPR 2011.

[42]  Ben Shneiderman,et al.  Users can change their web search tactics: Design guidelines for categorized overviews , 2008, Inf. Process. Manag..

[43]  Shiyang Lu,et al.  Browse-to-search , 2012, ACM Multimedia.

[44]  Bernd Girod,et al.  CHoG: Compressed histogram of gradients A low bit-rate feature descriptor , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.