Toward interactive training and evaluation

Machine learning often relies on costly labeled data, and this impedes its application to new classification and information extraction problems. This has motivated the development of methods for leveraging abundant prior knowledge about these problems, including methods for lightly supervised learning using model expectation constraints. Building on this work, we envision an interactive training paradigm in which practitioners perform evaluation, analyze errors, and provide and refine expectation constraints in a closed loop. In this paper, we focus on several key subproblems in this paradigm that can be cast as selecting a representative sample of the unlabeled data for the practitioner to inspect. To address these problems, we propose stratified sampling methods that use model expectations as a proxy for latent output variables. In classification and sequence labeling experiments, these sampling strategies reduce accuracy evaluation effort by as much as 53%, provide more reliable estimates of $F_1$ for rare labels, and aid in the specification and refinement of constraints.

[1]  Tom M. Mitchell,et al.  Text clustering with extended user feedback , 2006, SIGIR.

[2]  J. Neyman On the Two Different Aspects of the Representative Method: the Method of Stratified Sampling and the Method of Purposive Selection , 1934 .

[3]  Paul A. Viola,et al.  Corrective feedback and persistent learning for information extraction , 2006, Artif. Intell..

[4]  D. Holt,et al.  Multiple frame sampling for multivariate stratification , 1994 .

[5]  Burr Settles,et al.  Active Learning Literature Survey , 2009 .

[6]  Dan Roth,et al.  Interactive Feature Space Construction using Semantic Information , 2009, CoNLL.

[7]  Andrew McCallum,et al.  Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data , 2001, ICML.

[8]  T. Fearn The Jackknife , 2000 .

[9]  Tobias Scheffer,et al.  Active Estimation of F-Measures , 2010, NIPS.

[10]  Andrew McCallum,et al.  Active Learning by Labeling Features , 2009, EMNLP.

[11]  Steffen Bickel,et al.  Active Risk Estimation , 2010, ICML.

[12]  J. L. Hodges,et al.  Minimum Variance Stratification , 1959 .

[13]  Andrew McCallum,et al.  Generalized expectation criteria for lightly supervised learning , 2011 .

[14]  John Langford,et al.  The Epoch-Greedy Algorithm for Multi-armed Bandits with Side Information , 2007, NIPS.

[15]  Burr Settles,et al.  Closing the Loop: Fast, Interactive Semi-Supervised Annotation With Queries on Features and Instances , 2011, EMNLP.

[16]  Ben Taskar,et al.  Posterior Regularization for Structured Latent Variable Models , 2010, J. Mach. Learn. Res..

[17]  Steven A. Orszag,et al.  CBMS-NSF REGIONAL CONFERENCE SERIES IN APPLIED MATHEMATICS , 1978 .

[18]  J. Rao,et al.  Inference From Stratified Samples: Properties of the Linearization, Jackknife and Balanced Repeated Replication Methods , 1981 .

[19]  Dan Klein,et al.  Learning from measurements in exponential families , 2009, ICML '09.

[20]  Emine Yilmaz,et al.  A simple and efficient sampling method for estimating AP and NDCG , 2008, SIGIR '08.

[21]  Gideon S. Mann,et al.  Learning from labeled features using generalized expectation criteria , 2008, SIGIR '08.

[22]  M. Ruiz Espejo Sampling , 2013, Encyclopedic Dictionary of Archaeology.

[23]  Gideon S. Mann,et al.  Generalized Expectation Criteria for Semi-Supervised Learning of Conditional Random Fields , 2008, ACL.

[24]  B. Efron The jackknife, the bootstrap, and other resampling plans , 1987 .

[25]  Andrew McCallum,et al.  An Introduction to Conditional Random Fields for Relational Learning , 2007 .

[26]  Paul N. Bennett,et al.  Online stratified sampling: evaluating classifiers at web-scale , 2010, CIKM.