MediaQ: mobile multimedia management system

MediaQ is a novel online media management system to collect, organize, share, and search mobile multimedia contents using automatically tagged geospatial metadata. User-generated-videos can be uploaded to the MediaQ from users' smartphones, iPhone and Android, and displayed accurately on a map interface according to their automatically sensed geospatial and other metadata. The MediaQ system provides the following distinct features. First, individual frames of videos (or any meaningful video segments) are automatically annotated by objective metadata which capture four dimensions in the real world: the capture time (when), the camera location and viewing direction (where), several key-words (what) and people (who). We term this data W4-metadata and they are obtained by utilizing camera sensors, geospatial and computer vision techniques. Second, a new approach of collecting multimedia data from the public has been implemented using spatial crowdsourcing, which allows media content to be collected in a coordinated manner for a specific purpose. Lastly, flexible video search features are implemented using W4 metadata, such as directional queries for selecting multimedia with a specific viewing direction. This paper is to present the design of a comprehensive mobile multimedia management system, MediaQ, and to share our experience in its implementation. Our extensive real world experimental case studies demonstrate that MediaQ can be an effective and comprehensive solution for various mobile multimedia applications.

[1]  Roger Zimmermann,et al.  Automatic tag generation and ranking for sensor-rich outdoor videos , 2011, MM '11.

[2]  Cyrus Shahabi,et al.  GeoCrowd: enabling query answering with spatial crowdsourcing , 2012, SIGSPATIAL/GIS.

[3]  Roger Zimmermann,et al.  Relevance ranking in georeferenced video search , 2009, Multimedia Systems.

[4]  Brendan T. O'Connor,et al.  Cheap and Fast – But is it Good? Evaluating Non-Expert Annotations for Natural Language Tasks , 2008, EMNLP.

[5]  Alireza Sahami Shirazi,et al.  Location-based crowdsourcing: extending crowdsourcing to the real world , 2010, NordiCHI.

[6]  Jong-Hyun Park,et al.  The interactive geographic video , 2003, IGARSS 2003. 2003 IEEE International Geoscience and Remote Sensing Symposium. Proceedings (IEEE Cat. No.03CH37477).

[7]  Murat Demirbas,et al.  Crowdsourcing location-based queries , 2011, 2011 IEEE International Conference on Pervasive Computing and Communications Workshops (PERCOM Workshops).

[8]  Roger Zimmermann,et al.  Viewable scene modeling for geospatial video search , 2008, ACM Multimedia.

[9]  Prashant J. Shenoy,et al.  SEVA: Sensor-enhanced video annotation , 2009, TOMCCAP.

[10]  A. Kansal,et al.  Energy-Accuracy Aware Localization for Mobile Devices , 2009 .

[11]  Feng Zhao,et al.  Energy-accuracy trade-off for continuous mobile device location , 2010, MobiSys '10.

[12]  Cyrus Shahabi,et al.  A privacy-aware framework for participatory sensing , 2011, SKDD.

[13]  Roger Zimmermann,et al.  Automatic positioning data correction for sensor-annotated mobile videos , 2012, SIGSPATIAL/GIS.

[14]  Tim Kraska,et al.  CrowdDB: answering queries with crowdsourcing , 2011, SIGMOD '11.

[15]  Cyrus Shahabi,et al.  Key Frame Selection Algorithms for Automatic Generation of Panoramic Images from Crowdsourced Geo-tagged Videos , 2014, W2GIS.

[16]  Mahmood Fathy,et al.  Efficient key frames selection for panorama generation from video , 2011, J. Electronic Imaging.

[17]  Jong-Hun Lee,et al.  MPEG-7 metadata for video-based GIS applications , 2003, IGARSS 2003. 2003 IEEE International Geoscience and Remote Sensing Symposium. Proceedings (IEEE Cat. No.03CH37477).

[18]  Chin-Laung Lei,et al.  A crowdsourceable QoE evaluation framework for multimedia content , 2009, ACM Multimedia.

[19]  Frédo Durand,et al.  Learning to predict where humans look , 2009, 2009 IEEE 12th International Conference on Computer Vision.