Improving Crowd Labeling through Expert Evaluation
暂无分享,去创建一个
[1] Brendan T. O'Connor,et al. Cheap and Fast – But is it Good? Evaluating Non-Expert Annotations for Natural Language Tasks , 2008, EMNLP.
[2] Gerardo Hermosillo,et al. Learning From Crowds , 2010, J. Mach. Learn. Res..
[3] Devavrat Shah,et al. Budget-optimal crowdsourcing using low-rank matrix approximations , 2011, 2011 49th Annual Allerton Conference on Communication, Control, and Computing (Allerton).
[4] Nancy Ide,et al. Making Sense of Word Sense Variation , 2009, SEW@NAACL-HLT.
[5] K. Bretonnel Cohen,et al. Last Words: Amazon Mechanical Turk: Gold Mine or Coal Mine? , 2011, CL.
[6] Panagiotis G. Ipeirotis,et al. Get another label? improving data quality and data mining using multiple, noisy labelers , 2008, KDD.
[7] Jennifer G. Dy,et al. Active Learning from Crowds , 2011, ICML.
[8] Devavrat Shah,et al. Iterative Learning for Reliable Crowdsourcing Systems , 2011, NIPS.
[9] Gerardo Hermosillo,et al. Supervised learning from multiple experts: whom to trust when everyone lies a bit , 2009, ICML '09.
[10] Javier R. Movellan,et al. Whose Vote Should Count More: Optimal Integration of Labels from Labelers of Unknown Expertise , 2009, NIPS.
[11] Andrew McCallum,et al. Active Learning by Labeling Features , 2009, EMNLP.
[12] Panagiotis G. Ipeirotis,et al. Running Experiments on Amazon Mechanical Turk , 2010, Judgment and Decision Making.
[13] Devavrat Shah,et al. Budget-Optimal Task Allocation for Reliable Crowdsourcing Systems , 2011, Oper. Res..
[14] Mark W. Schmidt,et al. Modeling annotator expertise: Learning when everybody knows a bit of something , 2010, AISTATS.
[15] David A. Forsyth,et al. Utility data annotation with Amazon Mechanical Turk , 2008, 2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops.
[16] Nancy Ide,et al. Anveshan: A Framework for Analysis of Multiple Annotators’ Labeling Behavior , 2010, Linguistic Annotation Workshop.