Validity of Online Screening for Autism: Crowdsourcing Study Comparing Paid and Unpaid Diagnostic Tasks

Background Obtaining a diagnosis of neuropsychiatric disorders such as autism requires long waiting times that can exceed a year and can be prohibitively expensive. Crowdsourcing approaches may provide a scalable alternative that can accelerate general access to care and permit underserved populations to obtain an accurate diagnosis. Objective We aimed to perform a series of studies to explore whether paid crowd workers on Amazon Mechanical Turk (AMT) and citizen crowd workers on a public website shared on social media can provide accurate online detection of autism, conducted via crowdsourced ratings of short home video clips. Methods Three online studies were performed: (1) a paid crowdsourcing task on AMT (N=54) where crowd workers were asked to classify 10 short video clips of children as “Autism” or “Not autism,” (2) a more complex paid crowdsourcing task (N=27) with only those raters who correctly rated ≥8 of the 10 videos during the first study, and (3) a public unpaid study (N=115) identical to the first study. Results For Study 1, the mean score of the participants who completed all questions was 7.50/10 (SD 1.46). When only analyzing the workers who scored ≥8/10 (n=27/54), there was a weak negative correlation between the time spent rating the videos and the sensitivity (ρ=–0.44, P=.02). For Study 2, the mean score of the participants rating new videos was 6.76/10 (SD 0.59). The average deviation between the crowdsourced answers and gold standard ratings provided by two expert clinical research coordinators was 0.56, with an SD of 0.51 (maximum possible SD is 3). All paid crowd workers who scored 8/10 in Study 1 either expressed enjoyment in performing the task in Study 2 or provided no negative comments. For Study 3, the mean score of the participants who completed all questions was 6.67/10 (SD 1.61). There were weak correlations between age and score (r=0.22, P=.014), age and sensitivity (r=–0.19, P=.04), number of family members with autism and sensitivity (r=–0.195, P=.04), and number of family members with autism and precision (r=–0.203, P=.03). A two-tailed t test between the scores of the paid workers in Study 1 and the unpaid workers in Study 3 showed a significant difference (P<.001). Conclusions Many paid crowd workers on AMT enjoyed answering screening questions from videos, suggesting higher intrinsic motivation to make quality assessments. Paid crowdsourcing provides promising screening assessments of pediatric autism with an average deviation <20% from professional gold standard raters, which is potentially a clinically informative estimate for parents. Parents of children with autism likely overfit their intuition to their own affected child. This work provides preliminary demographic data on raters who may have higher ability to recognize and measure features of autism across its wide range of phenotypic manifestations.

[1]  M. Knapp,et al.  Costs of autism spectrum disorders in the United Kingdom and the United States. , 2014, JAMA pediatrics.

[2]  Q. Zeng-Treitler,et al.  Assessing Pictograph Recognition: A Comparison of Crowdsourcing and Traditional Survey Approaches , 2015, Journal of medical Internet research.

[3]  Cynthia D. Zubritsky,et al.  Factors Associated With Age of Diagnosis Among Children With Autism Spectrum Disorders , 2005, Pediatrics.

[4]  Peter Washington,et al.  Feasibility Testing of a Wearable Behavioral Aid for Social Learning in Children with Autism , 2018, Applied Clinical Informatics.

[5]  Peter Washington,et al.  Guess What? , 2018, Journal of Healthcare Informatics Research.

[6]  Arshya Vahabzadeh,et al.  Social Communication Coaching Smartglasses: Well Tolerated in a Diverse Sample of Children and Adults With Autism , 2017, JMIR mHealth and uHealth.

[7]  Philippe Ravaud,et al.  Mapping of Crowdsourcing in Health: Systematic Review , 2018, Journal of medical Internet research.

[8]  Peter Washington,et al.  SuperpowerGlass , 2017, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[9]  Arshya Vahabzadeh,et al.  Second Version of Google Glass as a Wearable Socio-Affective Aid: Positive School Desirability, High Usability, and Theoretical Framework in a Sample of Children with Autism , 2018, JMIR human factors.

[10]  Peter Washington,et al.  Exploratory study examining the at-home feasibility of a wearable tool for social-affective learning in children with autism , 2018, npj Digital Medicine.

[11]  H. S. S. Yee,et al.  Mobile technology for children with Autism Spectrum Disorder: Major trends and issues , 2012, 2012 IEEE Symposium on E-Learning, E-Management and E-Services.

[12]  Ashley N. D. Meyer,et al.  Crowdsourcing Diagnosis for Patients With Undiagnosed Illnesses: An Evaluation of CrowdMed , 2016, Journal of medical Internet research.

[13]  Aniket Kittur,et al.  Crowdsourcing user studies with Mechanical Turk , 2008, CHI.

[14]  Kevin M. Ayres,et al.  THE USE OF MOBILE TECHNOLOGIES TO ASSIST WITH LIFE SKILLS/INDEPENDENCE OF STUDENTS WITH MODERATE/SEVERE INTELLECTUAL DISABILITY AND/OR AUTISM SPECTRUM DISORDERS: CONSIDERATIONS FOR THE FUTURE OF SCHOOL PSYCHOLOGY , 2013 .

[15]  L. LeBlanc,et al.  Use of Technology in Interventions for Children with Autism , 2004 .

[16]  G. Abowd,et al.  A Novel System for Supporting Autism Diagnosis Using Home Videos: Iterative Development and Evaluation of System Design , 2015, JMIR mHealth and uHealth.

[17]  E. Yom-Tov,et al.  Risk Assessment for Parents Who Suspect Their Child Has Autism Spectrum Disorder: Machine Learning Approach , 2018, Journal of medical Internet research.

[18]  Peter Washington,et al.  Mobile detection of autism through machine learning on home video: A development and prospective validation study , 2018, PLoS medicine.

[19]  J. Paul Leigh,et al.  Brief Report: Forecasting the Economic Burden of Autism in 2015 and 2025 in the United States , 2015, Journal of autism and developmental disorders.

[20]  Dennis P. Wall,et al.  Can we accelerate autism discoveries through crowdsourcing , 2016 .

[21]  Stephen J. Blumberg,et al.  The Prevalence of Parent-Reported Autism Spectrum Disorder Among US Children , 2018, Pediatrics.

[22]  Arshya Vahabzadeh,et al.  Improvement of Attention-Deficit/Hyperactivity Disorder Symptoms in School-Aged Children, Adolescents, and Young Adults With Autism via a Digital Smartglasses-Based Socioemotional Coaching Aid: Short-Term, Uncontrolled Pilot Study , 2017, bioRxiv.

[23]  Fernando González-Ladrón-de-Guevara,et al.  Towards an integrated crowdsourcing definition , 2012, J. Inf. Sci..

[24]  Zachary F. Meisel,et al.  Crowdsourcing—Harnessing the Masses to Advance Health and Medicine, a Systematic Review , 2013, Journal of General Internal Medicine.

[25]  Illhoi Yoo,et al.  A Systematic Review of Healthcare Applications for Smartphones , 2012, BMC Medical Informatics and Decision Making.

[26]  Georgina Peacock,et al.  Whittling Down the Wait Time: Exploring Models to Minimize the Delay from Initial Concern to Diagnosis and Treatment of Autism Spectrum Disorder. , 2016, Pediatric clinics of North America.

[27]  Peter Washington,et al.  A Gamified Mobile System for Crowdsourcing Video for Autism Research , 2018, 2018 IEEE International Conference on Healthcare Informatics (ICHI).

[28]  Peter Washington,et al.  Superpower glass: delivering unobtrusive real-time social cues in wearable systems , 2016, UbiComp Adjunct.

[29]  Kevin M. Ayres,et al.  Video Self-Prompting and Mobile Technology to Increase Daily Living and Vocational Independence for Students with Autism Spectrum Disorders , 2012, Journal of Developmental and Physical Disabilities.

[30]  Lizbeth Escobedo,et al.  MOSOCO: a mobile assistive tool to support children with autism practicing social skills in real-life situations , 2012, CHI.