CrowdIA: Solving Mysteries with Crowdsourced Sensemaking

The increasing volume of text data is challenging the cognitive capabilities of expert analysts. Machine learning and crowdsourcing present new opportunities for large-scale sensemaking, but we must overcome the challenge of modeling the overall process so that many distributed agents can contribute to suitable components asynchronously and meaningfully. In this paper, we explore how to crowdsource the sensemaking process via a pipeline of modularized steps connected by clearly defined inputs and outputs. Our pipeline restructures and partitions information into "context slices" for individual workers. We implemented CrowdIA, a software platform to enable unsupervised crowd sensemaking using our pipeline. With CrowdIA, crowds successfully solved two mysteries, and were one step away from solving the third. The crowd's intermediate results revealed their reasoning process and provided evidence that justifies their conclusions. We suggest broader possibilities to optimize each component, as well as to evaluate and refine previous intermediate analyses to improve the final result.

[1]  Hao Wu,et al.  Where do I start?: algorithmic strategies to guide intelligence analysts , 2012, ISI-KDD '12.

[2]  Lydia B. Chilton,et al.  Cascade: crowdsourcing taxonomy creation , 2013, CHI.

[3]  Aniket Kittur,et al.  The Knowledge Accelerator: Big Picture Thinking in Small Pieces , 2016, CHI.

[4]  Krzysztof Z. Gajos,et al.  Human computation tasks with global constraints , 2012, CHI.

[5]  Stuart K. Card,et al.  Entity-based collaboration tools for intelligence analysis , 2008, 2008 IEEE Symposium on Visual Analytics Science and Technology.

[6]  Wei Wu,et al.  Structuring, Aggregating, and Evaluating Crowdsourced Design Critique , 2015, CSCW.

[7]  Michael S. Bernstein,et al.  Mechanical Novel: Crowdsourcing Complex Work through Reflection and Revision , 2016, CSCW.

[8]  Zoe Borovsky,et al.  Topic Modeling , 2017, Encyclopedia of Machine Learning and Data Mining.

[9]  Scott R. Klemmer,et al.  Gut Instinct: Creating Scientific Theories with Online Learners , 2017, CHI.

[10]  Steven Dow,et al.  Comparing Different Sensemaking Approaches for Large-Scale Ideation , 2016, CHI.

[11]  Aniket Kittur,et al.  Reviewing versus doing: learning and performance in crowd assessment , 2014, CSCW.

[12]  Brian P. Bailey,et al.  Voyant: generating structured feedback on visual designs using a crowd of non-experts , 2014, CSCW.

[13]  Paul Johns,et al.  Pathfinder: an online collaboration environment for citizen scientists , 2009, CHI.

[14]  Debbie Gary Patty Wagstaff's second act : an airshow superstar adds firefighting to her repertoire , 2011 .

[15]  Steven Dow,et al.  IdeaGens: Enabling Expert Facilitation of Crowd Brainstorming , 2016, CSCW '16 Companion.

[16]  Johnny Nhan,et al.  Digilantism: An Analysis of Crowdsourcing and the Boston Marathon Bombings , 2015 .

[17]  HeerJeffrey,et al.  Voyagers and voyeurs , 2009 .

[18]  Tim Kraska,et al.  CrowdER: Crowdsourcing Entity Resolution , 2012, Proc. VLDB Endow..

[19]  L. Elisa Celis,et al.  Assignment Techniques for Crowdsourcing Sensitive Tasks , 2016, CSCW.

[20]  Tianyi Li Crowdsourcing Intelligence Analysis with Context Slices , 2018 .

[21]  Weiai Xu,et al.  Network visualization and problem-solving support: A cognitive fit study , 2018, Soc. Networks.

[22]  Michael S. Bernstein,et al.  Crowd Guilds: Worker-led Reputation and Feedback on Crowdsourcing Platforms , 2016, CSCW.

[23]  Jeffrey V. Nickerson,et al.  Putting the Pieces Back Together Again: Contest Webs for Large-Scale Problem Solving , 2017, CSCW.

[24]  Alexander W. Skaburskis,et al.  The Sandbox for analysis: concepts and methods , 2006, CHI.

[25]  Richards J. Heuer,et al.  Psychology of Intelligence Analysis , 1999 .

[26]  George Chin,et al.  Exploring the analytical processes of intelligence analysts , 2009, CHI.

[27]  G. Stasser,et al.  Pooling of Unshared Information in Group Decision Making: Biased Information Sampling During Discussion , 1985 .

[28]  Michael S. Bernstein,et al.  Flash Organizations: Crowdsourcing Complex Work by Structuring Crowds As Organizations , 2017, CHI.

[29]  Cláudio T. Silva,et al.  VisTrails: enabling interactive multiple-view visualizations , 2005, VIS 05. IEEE Visualization, 2005..

[30]  Stephen Soderland,et al.  Learning Information Extraction Rules for Semi-Structured and Free Text , 1999, Machine Learning.

[31]  Aniket Kittur,et al.  Costs and benefits of structured information foraging , 2013, CHI.

[32]  Roman V. Yampolskiy,et al.  Turing Test as a Defining Feature of AI-Completeness , 2013, Artificial Intelligence, Evolutionary Computing and Metaheuristics.

[33]  Susan R. Fussell,et al.  Effects of Sensemaking Translucence on Distributed Collaborative Analysis , 2016, CSCW.

[34]  Chris North,et al.  Semantic interaction for visual text analytics , 2012, CHI.

[35]  Chris North,et al.  BiSet: Semantic Edge Bundling with Biclusters for Sensemaking , 2016, IEEE Transactions on Visualization and Computer Graphics.

[36]  R. Hastie,et al.  The common knowledge effect: Information sharing and group judgment. , 1993 .

[37]  Nikhil Sharma,et al.  Sensemaking handoff: When and how? , 2008, ASIST.

[38]  Aniket Kittur,et al.  CrowdForge: crowdsourcing complex work , 2011, UIST.

[39]  Martin Wattenberg,et al.  Voyagers and voyeurs: supporting asynchronous collaborative information visualization , 2007, CHI.

[40]  Christopher Andrews,et al.  VizCept: Supporting synchronous collaboration for constructing visualizations in intelligence analysis , 2010, 2010 IEEE Symposium on Visual Analytics Science and Technology.

[41]  Jeffrey Heer,et al.  CommentSpace: structured support for collaborative visual analysis , 2011, CHI.

[42]  Stephen Morse,et al.  Putting the pieces back together again: an illustration of the problem of interpreting development indicators using an African case study , 2004 .

[43]  Robert M. Clark,et al.  Intelligence Analysis: A Target-Centric Approach , 2003 .

[44]  Wen-Huang Cheng,et al.  Context-based page unit recommendation for web-based sensemaking tasks , 2009, IUI.

[45]  P. Pirolli,et al.  The Sensemaking Process and Leverage Points for Analyst Technology as Identified Through Cognitive Task Analysis , 2007 .

[46]  R. Jordan Crouser,et al.  Online Submission ID: 200 An Affordance-Based Framework for Human Computation and Human-Computer Collaboration , 2022 .

[47]  Kristin A. Cook,et al.  Illuminating the Path: The Research and Development Agenda for Visual Analytics , 2005 .

[48]  Susan R. Fussell,et al.  Effects of visualization and note-taking on sensemaking and analysis , 2013, CHI.

[49]  Christopher Andrews,et al.  Large high resolution displays for co-located collaborative sensemaking: Display usage and territoriality , 2013, Int. J. Hum. Comput. Stud..

[50]  Aditya G. Parameswaran,et al.  Crowdsourced Data Management: Industry and Academic Perspectives , 2015, Found. Trends Databases.

[51]  Hanna M. Wallach,et al.  Topic modeling: beyond bag-of-words , 2006, ICML.

[52]  Jian Zhao,et al.  Supporting Handoff in Asynchronous Collaborative Sensemaking Using Knowledge-Transfer Graphs , 2018, IEEE Transactions on Visualization and Computer Graphics.

[53]  Aniket Kittur,et al.  Crowd synthesis: extracting categories and clusters from complex data , 2014, CSCW.

[54]  Martin Wattenberg,et al.  Visualizing Dataflow Graphs of Deep Learning Models in TensorFlow , 2018, IEEE Transactions on Visualization and Computer Graphics.

[55]  Devi Parikh Human-Debugging of Machines , 2011 .

[56]  Michael S. Bernstein,et al.  Context Trees: Crowdsourcing Global Understanding from Local Views , 2014, HCOMP.

[57]  Mani B. Srivastava,et al.  Supporting Reasoning with Different Types of Evidence in Intelligence Analysis , 2015, AAMAS.

[58]  Aniket Kittur,et al.  Sensemaking : Improving Sensemaking by Leveraging the Efforts of Previous Users , 2012 .

[59]  Aniket Kittur,et al.  Standing on the schemas of giants: socially augmented information foraging , 2014, CSCW.

[60]  Bob Fields,et al.  Developing a Model of Distributed Sensemaking: A Case Study of Military Analysis , 2015, Informatics.

[61]  Lydia B. Chilton,et al.  Exploring iterative and parallel human computation processes , 2010, HCOMP '10.

[62]  Eric Gilbert,et al.  Piggyback Prototyping: Using Existing, Large-Scale Social Computing Systems to Prototype New Ones , 2015, CHI.

[63]  Michelle X. Zhou,et al.  Interactive Visual Synthesis of Analytic Knowledge , 2006, 2006 IEEE Symposium On Visual Analytics Science And Technology.

[64]  Michael S. Bernstein,et al.  Soylent: a word processor with a crowd inside , 2010, UIST.

[65]  John T. Stasko,et al.  Jigsaw: Supporting Investigative Analysis through Interactive Visualization , 2007, 2007 IEEE Symposium on Visual Analytics Science and Technology.

[66]  Philipp Drieger,et al.  Semantic Network Analysis as a Method for Visual Text Analytics , 2013 .

[67]  Nikhil Sharma,et al.  Artifact usefulness and usage in sensemaking handoffs , 2009, ASIST.

[68]  Yla R. Tausczik,et al.  Distributed Knowledge in Crowds: Crowd Performance on Hidden Profile Tasks , 2018, ICWSM.

[69]  Aniket Kittur,et al.  Alloy: Clustering with Crowds and Computation , 2016, CHI.

[70]  Scott R. Klemmer,et al.  Shepherding the crowd yields better work , 2012, CSCW.

[71]  Etienne Wenger,et al.  Situated Learning: Legitimate Peripheral Participation , 1991 .