CrowdLearner: rapidly creating mobile recognizers using crowdsourcing

Mobile applications can offer improved user experience through the use of novel modalities and user context. However, these new input dimensions often require recognition-based techniques, with which mobile app developers or designers may not be familiar. Furthermore, the recruiting, data collection and labeling, necessary for using these techniques, are usually time-consuming and expensive. We present CrowdLearner, a framework based on crowdsourcing to automatically generate recognizers using mobile sensor input such as accelerometer or touchscreen readings. CrowdLearner allows a developer to easily create a recognition task, distribute it to the crowd, and monitor its progress as more data becomes available. We deployed CrowdLearner to a crowd of 72 mobile users over a period of 2.5 weeks. We evaluated the system by experimenting with 6 recognition tasks concerning motion gestures, touchscreen gestures, and activity recognition. The experimental results indicated that CrowdLearner enables a developer to quickly acquire a usable recognizer for their specific application by spending a moderate amount of money, often less than $10, in a short period of time, often in the order of 2 hours. Our exploration also revealed challenges and provided insights into the design of future crowdsourcing systems for machine learning tasks.

[1]  Ramesh Govindan,et al.  Medusa: a programming framework for crowd-sensing applications , 2012, MobiSys '12.

[2]  Scott R. Klemmer,et al.  Authoring sensor-based interactions by demonstration with direct manipulation and pattern recognition , 2007, CHI.

[3]  Aniket Kittur,et al.  Crowdsourcing user studies with Mechanical Turk , 2008, CHI.

[4]  Bernhard E. Boser,et al.  A training algorithm for optimal margin classifiers , 1992, COLT '92.

[5]  Lorrie Faith Cranor,et al.  Are your participants gaming the system?: screening mechanical turk workers , 2010, CHI.

[6]  Yang Li,et al.  Tap, swipe, or move: attentional demands for distracted smartphone input , 2012, AVI.

[7]  Gaetano Borriello,et al.  A Practical Approach to Recognizing Physical Activities , 2006, Pervasive.

[8]  Nayan B. Ruparelia The history of version control , 2010, ACM SIGSOFT Softw. Eng. Notes.

[9]  John Langford,et al.  Telling humans and computers apart automatically , 2004, CACM.

[10]  Emiliano Miluzzo,et al.  People-centric urban sensing , 2006, WICON '06.

[11]  Bi Wu,et al.  CrowdLab: An architecture for volunteer mobile testbeds , 2011, 2011 Third International Conference on Communication Systems and Networks (COMSNETS 2011).

[12]  Michael S. Bernstein,et al.  Soylent: a word processor with a crowd inside , 2010, UIST.

[13]  Aniket Kittur,et al.  Harnessing the wisdom of crowds in wikipedia: quality through coordination , 2008, CSCW.

[14]  Yang Li,et al.  Bootstrapping personal gesture shortcuts with the wisdom of the crowd and handwriting recognition , 2012, CHI.

[15]  Walter S. Lasecki,et al.  Training Activity Recognition Systems Online Using Real-time Crowdsourcing , 2012 .

[16]  Yang Li,et al.  DoubleFlip: a motion gesture delimiter for mobile interaction , 2010, UIST '10.

[17]  Niels Henze,et al.  My App is an Experiment: Experience from User Studies in Mobile App Stores , 2011, Int. J. Mob. Hum. Comput. Interact..

[18]  Andrew T. Campbell,et al.  Cooperative Communities (CoCo): Exploiting Social Networks for Large-scale Modeling of Human Behavior , 2011 .

[19]  James A. Landay,et al.  Gestalt: integrated support for implementation and analysis in machine learning , 2010, UIST.

[20]  Yang Li,et al.  Gestures without libraries, toolkits or training: a $1 recognizer for user interface prototypes , 2007, UIST.

[21]  Clayton Shepard,et al.  LiveLab: measuring wireless networks and smartphone users in the field , 2011, SIGMETRICS Perform. Evaluation Rev..

[22]  Katharina Reinecke,et al.  Accurate measurements of pointing performance from in situ observations , 2012, CHI.

[23]  Thad Starner,et al.  MAGIC: a motion gesture design tool , 2010, CHI.

[24]  Ye Xu,et al.  Exploiting Social Networks for Large-Scale Human Behavior Modeling , 2011, IEEE Pervasive Computing.

[25]  Ian Witten,et al.  Data Mining , 2000 .

[26]  Ramachandran Ramjee,et al.  PRISM: platform for remote sensing using smartphones , 2010, MobiSys '10.

[27]  James A. Landay,et al.  MyExperience: a system for in situ tracing and capturing of user feedback on mobile phones , 2007, MobiSys '07.

[28]  อนิรุธ สืบสิงห์,et al.  Data Mining Practical Machine Learning Tools and Techniques , 2014 .

[29]  Jeffrey Heer,et al.  Crowdsourcing graphical perception: using mechanical turk to assess visualization design , 2010, CHI.

[30]  Burr Settles,et al.  Active Learning Literature Survey , 2009 .

[31]  Tracy L. Westeyn,et al.  Georgia tech gesture toolkit: supporting experiments in gesture recognition , 2003, ICMI '03.