| { |
| "paper_id": "K15-1020", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T07:08:59.276556Z" |
| }, |
| "title": "Making the Most of Crowdsourced Document Annotations: Confused Supervised LDA", |
| "authors": [ |
| { |
| "first": "Paul", |
| "middle": [], |
| "last": "Felt", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Brigham Young University", |
| "location": {} |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Eric", |
| "middle": [ |
| "K" |
| ], |
| "last": "Ringger", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Brigham Young University", |
| "location": {} |
| }, |
| "email": "ringger@cs.byu.edu" |
| }, |
| { |
| "first": "Jordan", |
| "middle": [], |
| "last": "Boyd-Graber", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "University of Colorado", |
| "location": {} |
| }, |
| "email": "jordan.boyd.graber@colorado.edu" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Seppi", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Brigham Young University", |
| "location": {} |
| }, |
| "email": "kseppi@byu.edu" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "Corpus labeling projects frequently use low-cost workers from microtask marketplaces; however, these workers are often inexperienced or have misaligned incentives. Crowdsourcing models must be robust to the resulting systematic and nonsystematic inaccuracies. We introduce a novel crowdsourcing model that adapts the discrete supervised topic model sLDA to handle multiple corrupt, usually conflicting (hence \"confused\") supervision signals. Our model achieves significant gains over previous work in the accuracy of deduced ground truth.", |
| "pdf_parse": { |
| "paper_id": "K15-1020", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "Corpus labeling projects frequently use low-cost workers from microtask marketplaces; however, these workers are often inexperienced or have misaligned incentives. Crowdsourcing models must be robust to the resulting systematic and nonsystematic inaccuracies. We introduce a novel crowdsourcing model that adapts the discrete supervised topic model sLDA to handle multiple corrupt, usually conflicting (hence \"confused\") supervision signals. Our model achieves significant gains over previous work in the accuracy of deduced ground truth.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Supervised machine learning requires labeled training corpora, historically produced by laborious and costly annotation projects. Microtask markets such as Amazon's Mechanical Turk and Crowdflower have turned crowd labor into a commodity that can be purchased with relatively little overhead. However, crowdsourced judgments can suffer from high error rates. A common solution to this problem is to obtain multiple redundant human judgments, or annotations, 1 relying on the observation that, in aggregate, non-experts often rival or exceed experts by averaging over individual error patterns (Surowiecki, 2005; Snow et al., 2008; Jurgens, 2013) .", |
| "cite_spans": [ |
| { |
| "start": 593, |
| "end": 611, |
| "text": "(Surowiecki, 2005;", |
| "ref_id": "BIBREF31" |
| }, |
| { |
| "start": 612, |
| "end": 630, |
| "text": "Snow et al., 2008;", |
| "ref_id": "BIBREF30" |
| }, |
| { |
| "start": 631, |
| "end": 645, |
| "text": "Jurgens, 2013)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Modeling Annotators and Abilities", |
| "sec_num": "1" |
| }, |
| { |
| "text": "A crowdsourcing model harnesses the wisdom of the crowd and infers labels based on the evidence of the available annotations, imperfect though they be. A common baseline crowdsourcing method aggregates annotations by majority vote, but this approach ignores important information. For example, some annotators are more reliable than others and their judgments ought to be upweighted accordingly. State-ofthe-art crowdsourcing methods account for annotator expertise, often through a probabilistic formalism. Compounding the challenge, assessing unobserved annotator expertise is tangled with estimating ground truth from imperfect annotations; thus, joint inference of these interrelated quantities is necessary. State-of-the-art models also take the data into account, because data features can help ratify or veto human annotators.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Modeling Annotators and Abilities", |
| "sec_num": "1" |
| }, |
| { |
| "text": "We introduce a model that improves on state of the art crowdsourcing algorithms by modeling not only the annotations but also the features of the data (e.g., words in a document). Section 2 identifies modeling deficiencies affecting previous work and proposes a solution based on topic modeling; Section 2.4 presents inference for the new model. Experiments that contrast the proposed model with select previous work on several text classification datasets are presented in Section 3. In Section 4 we highlight additional related work.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Modeling Annotators and Abilities", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Most crowdsourcing models extend the itemresponse model of Dawid and Skene (1979) . The Bayesian version of this model, referred to here as ITEMRESP, is depicted in Figure 1 . In the generative story for this model, a confusion matrix \u03b3 j is drawn for each human annotator j. Each row \u03b3 jc of the confusion matrix \u03b3 j is drawn from \u03b8 \u03b3 jc jc ) and encodes a categorical probability distribution over label classes that annotator j is apt to choose when presented with a document whose true label is c. Then for each document d an unobserved document label y d is drawn. Annotations are generated as annotator j corrupts the true label y d according to the categorical distribution Cat(\u03b3 jy d ).", |
| "cite_spans": [ |
| { |
| "start": 59, |
| "end": 81, |
| "text": "Dawid and Skene (1979)", |
| "ref_id": "BIBREF7" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 165, |
| "end": 173, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Latent Representations that Reflect Labels and Confusion", |
| "sec_num": "2" |
| }, |
| { |
| "text": "b (\u03b8) b (\u03b3) y d a dj d = 1 . . . D c = 1 . . . C j = 1 . . . J j = 1 . . . J", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Latent Representations that Reflect Labels and Confusion", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Some extensions to ITEMRESP model the features of the data (e.g., words in a document). Many data-aware crowdsourcing models condition the labels on the data (Jin and Ghahramani, 2002; Raykar et al., 2010; Liu et al., 2012; Yan et al., 2014) , possibly because discriminative classifiers dominate supervised machine learning. Others model the data generatively (Bragg et al., 2013; Lam and Stork, 2005; Felt et al., 2014; Simpson and Roberts, 2015) . Felt et al. (2015) argue that generative models are better suited than conditional models to crowdsourcing scenarios because generative models often learn faster than their conditional counterparts (Ng and Jordan, 2001)especially early in the learning curve. This advantage is amplified by the annotation noise typical of crowdsourcing scenarios.", |
| "cite_spans": [ |
| { |
| "start": 158, |
| "end": 184, |
| "text": "(Jin and Ghahramani, 2002;", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 185, |
| "end": 205, |
| "text": "Raykar et al., 2010;", |
| "ref_id": "BIBREF27" |
| }, |
| { |
| "start": 206, |
| "end": 223, |
| "text": "Liu et al., 2012;", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 224, |
| "end": 241, |
| "text": "Yan et al., 2014)", |
| "ref_id": "BIBREF34" |
| }, |
| { |
| "start": 361, |
| "end": 381, |
| "text": "(Bragg et al., 2013;", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 382, |
| "end": 402, |
| "text": "Lam and Stork, 2005;", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 403, |
| "end": 421, |
| "text": "Felt et al., 2014;", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 422, |
| "end": 448, |
| "text": "Simpson and Roberts, 2015)", |
| "ref_id": "BIBREF28" |
| }, |
| { |
| "start": 451, |
| "end": 469, |
| "text": "Felt et al. (2015)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Leveraging Data", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Extensions to ITEMRESP that model document features generatively tend to share a common high-level architecture. After the document class label y d is drawn for each document d, features are drawn from class-conditional distributions. Felt et al. (2015) identify the MOMRESP model, reproduced in Figure 2 , as a strong representative of generative crowdsourcing models. In MOMRESP,", |
| "cite_spans": [ |
| { |
| "start": 235, |
| "end": 253, |
| "text": "Felt et al. (2015)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 296, |
| "end": 304, |
| "text": "Figure 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Leveraging Data", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "\u03b8 \u03b3 jc \u03c6 c b (\u03b8) b (\u03c6) b (\u03b3) y d x d a dj d = 1 . . . D c = 1 . . . C j = 1 . . . J j = 1 . . . J c = 1 . . . C Figure 2: MOMRESP as a plate diagram. |x d | = V", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Leveraging Data", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": ", the size of the vocabulary. Documents with similar feature vectors x tend to share a common label y. Reduces to mixture-of-multinomials clustering when no annotations a are observed. the feature vector x d for document d is drawn from the multinomial distribution with parameter vector \u03c6 y d . This class-conditional multinomial model of the data inherits many of the strengths and weaknesses of the na\u00efve Bayes model that it resembles. Strengths include easy inference and a strong inductive bias which helps the model be robust to annotation noise and scarcity. Weaknesses include overly strict conditional independence assumptions among features, leading to overconfidence in the document model and thereby causing the model to overweight feature evidence and underweight annotation evidence. This imbalance can result in degraded performance in the presence of high quality or many annotations.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Leveraging Data", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "We solve the problem of imbalanced feature and annotation evidence observed in MOMRESP by replacing the class-conditional structure of previous generative crowdsourcing models with a richer generative story where documents are drawn first and class labels y d are obtained afterwards via a log-linear mapping. This move towards conditioning classes on documents content is sensible because in many situations document content is authored first, whereas label structure is not imposed until afterwards. It is plausible to assume that there will exist some mapping from a latent document structure to the desired document label distinctions. Moreover, by jointly modeling topics and the mapping to labels, we can learn the latent document representations that best explain how best to predict and correct annotator errors.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Confused Supervised LDA (CSLDA)", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "N d Size of document d N dt n 1(z dn = t) N t d,n 1(z dn = t) N jcc d a djc 1(y d = c) N jc N jc1 \u2022 \u2022 \u2022 N jcC N vt dn 1(w dn = v \u2227 z dn = t) N t dn 1(z dn = t) N", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Term Definition", |
| "sec_num": null |
| }, |
| { |
| "text": "Count excludes variable being sampled z d", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Term Definition", |
| "sec_num": null |
| }, |
| { |
| "text": "Vector", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Term Definition", |
| "sec_num": null |
| }, |
| { |
| "text": "wherez dt = 1 N d n 1(z dn = t) z d", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Term Definition", |
| "sec_num": null |
| }, |
| { |
| "text": "Excludes the z dn being sampled We call our model confused supervised LDA (CSLDA, Figure 3 ), based on supervised topic modeling. Latent Dirichlet Allocation (Blei et al., 2003, LDA) models text documents as admixtures of word distributions, or topics. Although pre-calculated LDA topics as features can inform a crowdsourcing model (Levenberg et al., 2014) , supervised LDA (sLDA) provides a principled way of incorporating document class labels and topics into a single model, allowing topic variables and response variables to co-inform one another in joint inference. For example, when sLDA is given movie reviews labeled with sentiment, inferred topics cluster around sentimentheavy words (Mcauliffe and Blei, 2007) , which may be quite different from the topics inferred by unsupervised LDA. One way to view CSLDA is as a discrete sLDA in settings with noisy supervision from multiple, imprecise annotators.", |
| "cite_spans": [ |
| { |
| "start": 158, |
| "end": 182, |
| "text": "(Blei et al., 2003, LDA)", |
| "ref_id": null |
| }, |
| { |
| "start": 333, |
| "end": 357, |
| "text": "(Levenberg et al., 2014)", |
| "ref_id": "BIBREF16" |
| }, |
| { |
| "start": 694, |
| "end": 720, |
| "text": "(Mcauliffe and Blei, 2007)", |
| "ref_id": "BIBREF19" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 82, |
| "end": 90, |
| "text": "Figure 3", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Term Definition", |
| "sec_num": null |
| }, |
| { |
| "text": "The generative story for CSLDA is:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Term Definition", |
| "sec_num": null |
| }, |
| { |
| "text": "1. Draw per-topic word distributions \u03c6 t from Dir(b (\u03b8) ).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Term Definition", |
| "sec_num": null |
| }, |
| { |
| "text": "Gauss(\u00b5, \u03a3).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Draw per-class regression parameters \u03b7 c from", |
| "sec_num": "2." |
| }, |
| { |
| "text": "with row \u03b3 jc drawn from Dir(b ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Draw per-annotator confusion matrices \u03b3 j", |
| "sec_num": "3." |
| }, |
| { |
| "text": "(\u03b3) jc ). 4. For each document d, (a) Draw topic vector \u03b8 d from Dir(b (\u03b8) ). (b) For each token position n, draw topic z dn from Cat(\u03b8 d ) and word w dn from Cat(\u03c6 z dn ). (c) Draw class label y d with probability pro- portional to exp[\u03b7 y dz d ]. (d) For each annotator j draw annotation vector a dj from \u03b3 jy d . \u03b8 d \u03b3 jc \u03c6 t \u03b7 c b (\u03b8) b (\u03c6) b (\u03b3) \u00b5 \u03a3 y d z dn w dn a dj d = 1 . . . D n = 1 . . . Nd c = 1 . . . C j = 1 . . . J c = 1 . . . C j = 1 . . . J t = 1 . . . T", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Draw per-annotator confusion matrices \u03b3 j", |
| "sec_num": "3." |
| }, |
| { |
| "text": "We use stochastic expectation maximization (EM) for posterior inference in CSLDA, alternating between sampling values for topics z and document class labels y (the E-step) and optimizing values of regression parameters \u03b7 (the M-step). To sample z and y efficiently, we derive the full conditional distributions of z and y in a collapsed model where \u03b8, \u03c6, and \u03b3 have been analytically integrated out. Omitting multiplicative constants, the collapsed model joint probability is", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Stochastic EM", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "p(z, w, y, a|\u03b7) = p(z)p(w|z)p(y|z, \u03b7)p(a|y) (1) \u221d M (a)\u2022 d B(N d +b (\u03b8) ) \u2022 t B(N t +b (\u03c6) t ) \u2022 d exp(\u03b7 y dz d ) c exp(\u03b7 czd ) \u2022 \uf8eb \uf8ed j c B(N jc +b (\u03b3) jc ) \uf8f6 \uf8f8 where B(\u2022)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Stochastic EM", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "is the Beta function (multivariate as necessary), counts N and related symbols are defined in Table 1 , and", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 94, |
| "end": 101, |
| "text": "Table 1", |
| "ref_id": "TABREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Stochastic EM", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "M (a) = d,j M (a dj ) where M (a dj )", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Stochastic EM", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "is the multinomial coefficient. Simplifying Equation 1 yields full conditionals for each word z dn ,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Stochastic EM", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "p(z dn = t|\u1e91, w, y, a, \u03b7) \u221d N dt + b (\u03b8) t (2) \u2022N w dn t + b (\u03c6) w dn N t + |b (\u03c6) | 1 \u2022 exp( \u03b7y d t N d ) c exp( \u03b7ct N d + \u03b7 c\u1e91d )", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Stochastic EM", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": ", and similarly for document label y d :", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Stochastic EM", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "p(y d = c|z, w, y, a, \u03b7) \u221d exp(\u03b7 czd ) c exp(\u03b7 c z d ) (3) \u2022 j c N jcc + b (\u03b3) a djc c N jcc + b (\u03b3) jcc c a djc ,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Stochastic EM", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "where", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Stochastic EM", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "x k x(x + 1) \u2022 \u2022 \u2022 (x + k \u2212 1)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Stochastic EM", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "is the rising factorial. In Equation 2 the first and third terms are independent of word n and can be cached at the document level for efficiency. For the M-step, we train the regression parameters \u03b7 (containing one vector per class) by optimizing the same objective function as for training a logistic regression classifier, assuming that class y is given:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Stochastic EM", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "p(y = c|z, \u03b7) = d exp(\u03b7 czd ) c exp(\u03b7 c z d )", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Stochastic EM", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": ". 4We optimize the objective (Equation 4) using L-BFGS and a regularizing Gaussian prior with \u00b5 = 0, \u03c3 2 = 1.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Stochastic EM", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "While EM is sensitive to initialization, CSLDA is straightforward to initialize. Majority vote is used to set initial y values\u1ef9. Corresponding initial values for z and \u03b7 are obtained by clamping y to\u1ef9 and running stochastic EM on z and \u03b7.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Stochastic EM", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "Ideally, we would test CSLDA performance under all of the many algorithms available for inference in such a model. Although that is not feasible, Asuncion et al. (2009) demonstrate that hyperaparameter optimization in LDA topic models helps to bring the performance of alternative inference algorithms into approximate agreement. Accordingly, in Section 2.4 we implement hyperparameter optimization for CSLDA to make our results as general as possible.", |
| "cite_spans": [ |
| { |
| "start": 146, |
| "end": 168, |
| "text": "Asuncion et al. (2009)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Hyperparameter Optimization", |
| "sec_num": "2.4" |
| }, |
| { |
| "text": "Before moving on, however, we take a moment to validate that the observation of Asuncion et al. generalizes from LDA to the ITEMRESP model, which, together with LDA, comprises CSLDA. Figure 4 demonstrates that three ITEMRESP inference algorithms, Gibbs sampling (Gibbs), meanfield variational inference (Var), and the iterated conditional modes algorithm (ICM) (Besag, 1986) , are brought into better agreement after optimizing their hyperparameters via grid search. That is, the algorithms in Figure 4b are in better agreement, particularly near the extremes, than the algorithms in Figure 4a . This difference is subtle, but it holds to an equal and greater extent in other simulation conditions we tested (experiment details are similar to those reported in Section 3).", |
| "cite_spans": [ |
| { |
| "start": 361, |
| "end": 374, |
| "text": "(Besag, 1986)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 183, |
| "end": 191, |
| "text": "Figure 4", |
| "ref_id": null |
| }, |
| { |
| "start": 494, |
| "end": 503, |
| "text": "Figure 4b", |
| "ref_id": null |
| }, |
| { |
| "start": 584, |
| "end": 593, |
| "text": "Figure 4a", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Hyperparameter Optimization", |
| "sec_num": "2.4" |
| }, |
| { |
| "text": "Although a grid search is effective, it is not practical for a model with many hyperparameters such as CSLDA. For efficiency, therefore, we use the fixed-point updates of Minka (2000) . Our updates differ slightly from Minka's since we tie hyperparameters, allowing them to be learned more quickly from less data. In our implementation the matrices of hyperparameters b (\u03c6) and b (\u03b8) over the Dirichlet-multinomial distributions are completely tied such that b", |
| "cite_spans": [ |
| { |
| "start": 171, |
| "end": 183, |
| "text": "Minka (2000)", |
| "ref_id": "BIBREF21" |
| }, |
| { |
| "start": 370, |
| "end": 373, |
| "text": "(\u03c6)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Fixed-point Hyperparameter Updates", |
| "sec_num": null |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "(\u03c6) tv = b (\u03c6) \u2200t, v and b (\u03b8) t = b (\u03b8) \u2200t. This leads to b (\u03c6) \u2190b (\u03c6) \u2022 t,v [\u03a8(N tv +b (\u03c6) )]\u2212T V \u03a8(b (\u03c6) ) V [\u03a8(N t +V b (\u03c6) )\u2212\u03a8(V b (\u03c6) )]", |
| "eq_num": "(5)" |
| } |
| ], |
| "section": "Fixed-point Hyperparameter Updates", |
| "sec_num": null |
| }, |
| { |
| "text": "and", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Fixed-point Hyperparameter Updates", |
| "sec_num": null |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "b (\u03b8) \u2190b (\u03b8) \u2022 d,t [\u03a8(N dt +b (\u03b8) )]\u2212N T \u03a8(b (\u03b8) ) T [\u03a8(N d +T b (\u03b8) )\u2212\u03a8(T b (\u03b8) )] .", |
| "eq_num": "(6)" |
| } |
| ], |
| "section": "Fixed-point Hyperparameter Updates", |
| "sec_num": null |
| }, |
| { |
| "text": "The updates for b (\u03b3) are slightly more involved since we choose to tie the diagonal entries b o , updating each separately:", |
| "cite_spans": [ |
| { |
| "start": 18, |
| "end": 21, |
| "text": "(\u03b3)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Fixed-point Hyperparameter Updates", |
| "sec_num": null |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "b (\u03b3) d \u2190b (\u03b3) d \u2022 j,c [\u03a8(N jcc +b (\u03b3) d )]\u2212JC\u03a8(b (\u03b3) d ) Z(b (\u03b3) )", |
| "eq_num": "(7)" |
| } |
| ], |
| "section": "Fixed-point Hyperparameter Updates", |
| "sec_num": null |
| }, |
| { |
| "text": "and b", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Fixed-point Hyperparameter Updates", |
| "sec_num": null |
| }, |
| { |
| "text": "(\u03b3) o \u2190 b (\u03b3) o \u2022 j,c,c =c [\u03a8(N jcc +b (\u03b3) o )]\u2212JC(C\u22121)\u03a8(b (\u03b3) o )] (C\u22121)Z(b (\u03b3) ) (8) where Z(b (\u03b3) ) = j,c [\u03a8(N jc + b (\u03b3) d + (C \u2212 1)b (\u03b3) o )] \u2212 JC\u03a8(b (\u03b3) d + (C \u2212 1)b (\u03b3) o )", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Fixed-point Hyperparameter Updates", |
| "sec_num": null |
| }, |
| { |
| "text": ". As in the work of Asuncion et al. (2009) , we add an algorithmic gamma prior (b (\u2022) \u223c G(\u03b1, \u03b2)) for smoothing by adding \u03b1\u22121 b (\u2022) to the numerator and \u03b2 to the denominator of Equations 5-8. Note that these algorithmic gamma \"priors\" should not be understood as first-class members of the CSLDA model (Figure 3) . Rather, they are regularization terms that keep our hyperparameter search algorithm from straying towards problematic values such as 0 or \u221e.", |
| "cite_spans": [ |
| { |
| "start": 20, |
| "end": 42, |
| "text": "Asuncion et al. (2009)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 82, |
| "end": 85, |
| "text": "(\u2022)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 301, |
| "end": 311, |
| "text": "(Figure 3)", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Fixed-point Hyperparameter Updates", |
| "sec_num": null |
| }, |
| { |
| "text": "For all experiments we set CSLDA's number of topics T to 1.5 times the number of classes in each dataset. We found that model performance was reasonably robust to this parameter. Only when T drops below the number of label classes does performance suffer. As per Section 2.3, z and \u03b7 values are initialized with 500 rounds of stochastic EM, after which the full model is updated with 1000 additional rounds. Predictions are generated by aggregating samples from the last 100 rounds (the mode of the approximate marginal posterior).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments", |
| "sec_num": "3" |
| }, |
| { |
| "text": "We compare CSLDA with (1) a majority vote baseline, (2) the ITEMRESP model, and representatives of the two main classes of dataaware crowdsourcing models, namely (3) datagenerative and (4) data-conditional. MOMRESP represents a typical data-generative model (Bragg et al., 2013; Felt et al., 2014; Lam and Stork, 2005; Simpson and Roberts, 2015) . Data-conditional approaches typically model data features conditionally using a log-linear model (Jin and Ghahramani, 2002; Raykar et al., 2010; Liu et al., 2012; Yan et al., 2014) . For the purposes of this paper, we refer to this model as LOGRESP. For ITEMRESP, MOMRESP, and LOGRESP we use the variational inference methods presented by Felt et al. (2015) . Unlike that paper, in this work we have augmented inference with the in-line hyperparameter updates described in Section 2.4. ", |
| "cite_spans": [ |
| { |
| "start": 258, |
| "end": 278, |
| "text": "(Bragg et al., 2013;", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 279, |
| "end": 297, |
| "text": "Felt et al., 2014;", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 298, |
| "end": 318, |
| "text": "Lam and Stork, 2005;", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 319, |
| "end": 345, |
| "text": "Simpson and Roberts, 2015)", |
| "ref_id": "BIBREF28" |
| }, |
| { |
| "start": 445, |
| "end": 471, |
| "text": "(Jin and Ghahramani, 2002;", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 472, |
| "end": 492, |
| "text": "Raykar et al., 2010;", |
| "ref_id": "BIBREF27" |
| }, |
| { |
| "start": 493, |
| "end": 510, |
| "text": "Liu et al., 2012;", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 511, |
| "end": 528, |
| "text": "Yan et al., 2014)", |
| "ref_id": "BIBREF34" |
| }, |
| { |
| "start": 687, |
| "end": 705, |
| "text": "Felt et al. (2015)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments", |
| "sec_num": "3" |
| }, |
| { |
| "text": "To gauge the effectiveness of data-aware crowdsourcing models, we use the sentiment-annotated tweet dataset distributed by CrowdFlower as a part of its \"data for everyone\" initiative. 2 In the \"Weather Sentiment\" task, 20 annotators judged the sentiment of 1000 tweets as either positive, negative, neutral, or unrelated to the weather. In the secondary \"Weather Sentiment Evaluated\" task, 10 additional annotators judged the correctness of each consensus label. We construct a gold standard from the consensus labels that were judged to be correct by 9 of the 10 annotators in the secondary task. Figure 5 plots learning curves of the accuracy of model-inferred labels as annotations are added (ordered by timestamp). All methods, including majority vote, converge to roughly the same accuracy when all 20,000 annotations are added. When fewer annotations are available, statistical models beat majority vote, and CSLDA is considerably more accurate than other approaches. Learning curves are bumpy because annotation order is not random and because inferred label accuracy is calculated only over documents with at least one annotation. Learning curves collectively increase when average annotation depth (the number of annotations per item) increases and decrease when new documents are annotated and average annotation depth decreases. CSLDA stands out by being more robust to these changes than other algorithms, and also by maintaining a higher level of accuracy across the board. This is important because high accuracy using fewer annotations translates to decreased annotations costs. News 16,995 20 22,851 111 WebKB 3,543 4 5,781 131 Reuters8 6,523 8 6,776 53 Reuters52 7,735 52 5,579 58 CADE12 34,801 12 41,628 110 Enron 3,854 32 14,069 431 ", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 598, |
| "end": 606, |
| "text": "Figure 5", |
| "ref_id": "FIGREF4" |
| }, |
| { |
| "start": 1594, |
| "end": 1738, |
| "text": "News 16,995 20 22,851 111 WebKB 3,543 4 5,781 131 Reuters8 6,523 8 6,776 53 Reuters52 7,735 52 5,579 58 CADE12 34,801 12 41,628", |
| "ref_id": "TABREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Human-generated Annotations", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "D C V 1 N d N d 20", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Human-generated Annotations", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "Datasets including both annotations and gold standard labels are in short supply. Although plenty of text categorization datasets have been annotated, common practice reflects that initial noisy annotations be discarded and only consensus labels be published. Consequently, we follow previous work in achieving broad validation by constructing synthetic annotators that corrupt known gold standard labels. We base our experimental setup on the annotations gathered by Felt et al. (2015) , 3 who paid CrowdFlower annotators to relabel 1000 documents from the well-known 20 Newsgroups classification dataset. In that experiment, 136 annotators contributed, each instance was labeled an average of 6.9 times, and annotator accuracies were distributed approximately according to a Beta(3.6, 5.1) distribution. Accordingly we construct 100 synthetic annotators, each parametrized by an accuracy drawn from Beta(3.6, 5.1) and with errors drawn from a symmetric Dirichlet Dir(1). Datasets are annotated by selecting an instance (at random without replacement) and then selecting K annotators (at random without replacement) to annotate it before moving on. We choose K = 7 to mirror the empirical average in the CrowdFlower annotation set.", |
| "cite_spans": [ |
| { |
| "start": 468, |
| "end": 486, |
| "text": "Felt et al. (2015)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Synthetic Annotations", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "We evaluate on six text classification datasets, summarized in Table 2 . The 20 Newsgroups, We-bKB, Cade12, Reuters8, and Reuters52 datasets are described in more detail by Cardoso-Cachopo (2007) . The LDC-labeled Enron emails dataset is described by Berry et al. (2001) . Each dataset is Figure 6 : Inferred label accuracy of models on synthetic annotations. The first instance is annotated 7 times, then the second, and so on.", |
| "cite_spans": [ |
| { |
| "start": 173, |
| "end": 195, |
| "text": "Cardoso-Cachopo (2007)", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 251, |
| "end": 270, |
| "text": "Berry et al. (2001)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 63, |
| "end": 70, |
| "text": "Table 2", |
| "ref_id": "TABREF1" |
| }, |
| { |
| "start": 289, |
| "end": 297, |
| "text": "Figure 6", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Synthetic Annotations", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "preprocessed via Porter stemming and by removal of the stopwords from MALLET's stopword list (McCallum, 2002) . Features occurring fewer than 5 times in the corpus are discarded. In the case of MOMRESP, features are fractionally scaled so that each document is the same length, in keeping with previous work in multinomial document models (Nigam et al., 2006) . Figure 6 plots learning curves on three representative datasets (Enron resembles Cade12, and the Reuters datasets resemble WebKB). CSLDA consistently outperforms LOGRESP, ITEMRESP, and majority vote.", |
| "cite_spans": [ |
| { |
| "start": 93, |
| "end": 109, |
| "text": "(McCallum, 2002)", |
| "ref_id": "BIBREF20" |
| }, |
| { |
| "start": 339, |
| "end": 359, |
| "text": "(Nigam et al., 2006)", |
| "ref_id": "BIBREF24" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 362, |
| "end": 370, |
| "text": "Figure 6", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Synthetic Annotations", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "The generative models (CSLDA and MOMRESP) tend to excel in lowannotation portions of the learning curve, partially because generative models tend to converge quickly and partially because generative models naturally learn from unlabeled documents (i.e., semi-supervision). However, MOMRESP tends to quickly reach a performance plateau after which additional annotations do little good. The performance of MOMRESP is also highly dataset de- Table 3 : The number of annotations \u00d71000 at which the algorithm reaches 95% inferred label accuracy on the indicated dataset (average annotations per instance are in parenthesis). All instances are annotated once, then twice, and so on. Empty entries ('-') do not reach 95% even with 20 annotations per instance.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 440, |
| "end": 447, |
| "text": "Table 3", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Synthetic Annotations", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "pendent: it is good on 20 Newsgroups, mediocre on WebKB, and poor on CADE12. By contrast, CSLDA is relatively stable across datasets.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Synthetic Annotations", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "To understand the different behavior of the two generative models, recall that MOMRESP is identical to ITEMRESP save for its multinomial data model. Indeed, the equations governing inference of label y in MOMRESP simply sum together terms from an ITEMRESP model and terms from a mixture of multinomials clustering model (and for reasons explained in Section 2.1, the multinomial data model terms tend to dominate). Therefore when MOMRESP diverges from ITEMRESP it is because MOMRESP is attracted toward a y assignment that satisfies the multinomial data model, grouping similar documents together. This can both help and hurt. When data clusters and label classes are misaligned, MOMRESP falters (as in the case of the Cade12 dataset). In contrast, CSLDA's flexible mapping from topics to labels is less sensitive: topics can diverge from label classes so long as there exists some linear transformation from the topics to the labels.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Synthetic Annotations", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Many corpus annotation projects are not complete until the corpus achieves some target level of quality. We repeat the experiment reported in Figure 6, but rather than simulating seven annotations for each instance before moving on, we simulate one annotation for each instance, then two, and so on until each instance in the dataset is annotated 20 times. Table 3 reports the minimal number of annotations before an algorithm's inferred labels reach an accuracy of 95%, a lofty goal that can require significant amounts of annotation when using poor quality annotations. CSLDA achieves 95% accuracy with fewer annotations, corresponding to reduced annotation cost. Figure 7 : Joint inference for CSLDA vs pipeline inference (CSLDA-P).", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 142, |
| "end": 148, |
| "text": "Figure", |
| "ref_id": null |
| }, |
| { |
| "start": 357, |
| "end": 364, |
| "text": "Table 3", |
| "ref_id": null |
| }, |
| { |
| "start": 666, |
| "end": 674, |
| "text": "Figure 7", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Synthetic Annotations", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "To isolate the effectiveness of joint inference in CSLDA, we compare against the pipeline alternative where topics are inferred first and then held constant during inference (Levenberg et al., 2014) . Joint inference yields modest but consistent benefits over a pipeline approach. Figure 7 highlights a portion of the learning curve on the Newsgroups dataset (based on the experiments summarized in Table 3 ). This trend holds across all of the datasets that we examined.", |
| "cite_spans": [ |
| { |
| "start": 174, |
| "end": 198, |
| "text": "(Levenberg et al., 2014)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 281, |
| "end": 289, |
| "text": "Figure 7", |
| "ref_id": null |
| }, |
| { |
| "start": 399, |
| "end": 406, |
| "text": "Table 3", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Joint vs Pipeline Inference", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "Class-conditional models like MOMRESP include a feature that data-conditional models like CSLDA lack: an explicit prior over class prevalence. Figure 8a shows that CSLDA performs poorly on the CrowdFlower-annotated Newsgroups documents described at the beginning of Section 3 (not the synthetic annotations). Error analysis uncovers that CSLDA lumps related classes together in this dataset. This is because annotators could specify up to 3 simultaneous labels for each annotation, so that similar labels (e.g., \"talk.politics.misc\" and \"talk.politics.mideast\") are usually chosen in blocks. Suppose each member of a set of documents with similar topical content is annotated with both label A and B. In this scenario it is apparent that CSLDA will achieve its best fit by inferring all documents to have the same label either A or B. By contrast, MOMRESP's uniform prior distribution over \u03b8 leads it to prefer solutions with a balance of A and B. The hypothesis that class combination explains CSLDA's performance is supported by Figure 8b , which shows that CSLDA recovers after combining the classes that were most frequently coannotated. We greedily combine label class pairs to maximize Krippendorf's \u03b1 until only 10 labels were left: \"alt.atheism,\" religion, and politics classes were combined; also, \"sci.electronics\" and the computing classes. The remaining eight classes were unaltered. However, one could also argue that the original behavior of CSLDA is in some ways desirable. That is, if two classes of documents are mostly the same both topically and in terms of annotator decisions, perhaps those classes ought to be collapsed. We are not overly concerned that MOMRESP beats CSLDA in Figure 8 , since this result is consistent with early relative performance in simulation.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 143, |
| "end": 152, |
| "text": "Figure 8a", |
| "ref_id": null |
| }, |
| { |
| "start": 1031, |
| "end": 1040, |
| "text": "Figure 8b", |
| "ref_id": null |
| }, |
| { |
| "start": 1699, |
| "end": 1707, |
| "text": "Figure 8", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Error Analysis", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "This section reviews related work not already discussed. A growing body of work extends the itemresponse model to account for variables such as item difficulty (Whitehill et al., 2009; Passonneau and Carpenter, 2013; Zhou et al., 2012) , annotator trustworthiness (Hovy et al., 2013) , correlation among various combinations of these variables , and change in annotator behavior over time (Simpson and Roberts, 2015) . Welinder et al. (2010) carefully model the process of annotating objects in images, including variables for item difficulty, item class, and classconditional perception noise. In follow-up work, Liu et al. (2012) demonstrate that similar levels of performance can be achieved with the simple item-response model by using variational inference rather than EM. Alternative inference algorithms have been proposed for crowdsourcing models (Dalvi et al., 2013; Ghosh et al., 2011; Karger et al., 2013; Zhang et al., 2014) . Some crowdsourcing work regards labeled data not as an end in itself, but rather as a means to train classifiers (Lin et al., 2014) . The fact-finding literature assigns trust scores to assertions made by untrusted sources (Pasternack and Roth, 2010) .", |
| "cite_spans": [ |
| { |
| "start": 160, |
| "end": 184, |
| "text": "(Whitehill et al., 2009;", |
| "ref_id": "BIBREF33" |
| }, |
| { |
| "start": 185, |
| "end": 216, |
| "text": "Passonneau and Carpenter, 2013;", |
| "ref_id": "BIBREF25" |
| }, |
| { |
| "start": 217, |
| "end": 235, |
| "text": "Zhou et al., 2012)", |
| "ref_id": "BIBREF36" |
| }, |
| { |
| "start": 264, |
| "end": 283, |
| "text": "(Hovy et al., 2013)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 389, |
| "end": 416, |
| "text": "(Simpson and Roberts, 2015)", |
| "ref_id": "BIBREF28" |
| }, |
| { |
| "start": 419, |
| "end": 441, |
| "text": "Welinder et al. (2010)", |
| "ref_id": "BIBREF32" |
| }, |
| { |
| "start": 614, |
| "end": 631, |
| "text": "Liu et al. (2012)", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 855, |
| "end": 875, |
| "text": "(Dalvi et al., 2013;", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 876, |
| "end": 895, |
| "text": "Ghosh et al., 2011;", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 896, |
| "end": 916, |
| "text": "Karger et al., 2013;", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 917, |
| "end": 936, |
| "text": "Zhang et al., 2014)", |
| "ref_id": "BIBREF35" |
| }, |
| { |
| "start": 1052, |
| "end": 1070, |
| "text": "(Lin et al., 2014)", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 1162, |
| "end": 1189, |
| "text": "(Pasternack and Roth, 2010)", |
| "ref_id": "BIBREF26" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Additional Related Work", |
| "sec_num": "4" |
| }, |
| { |
| "text": "We describe CSLDA, a generative, data-aware crowdsourcing model that addresses important modeling deficiencies identified in previous work. In particular, CSLDA handles data in which the natural document clusters are at odds with the intended document labels. It also transitions smoothly from situations in which few annotations are available to those in which many annotations are available. Because of the flexible mapping in CSLDA to class labels, many structural variants are possible in future work. For example, this mapping could depend not just on inferred topical content but also directly on data features (c.f. Nguyen et al. (2013)) or learned embedded feature representations.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion and Future Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "The large number of parameters in the learned confusion matrices of crowdsourcing models present difficulty at scale. This could be addressed by modeling structure both inside of the annotators and classes. Redundant annotations give unique insights into both inter-annotator and inter-class relationships and could be used to induce annotator or label class hierarchies with parsimonious representations. Simpson et al. (2013) identify annotator clusters using community detection algorithms but do not address annotator hierarchy or scalable confusion representations.", |
| "cite_spans": [ |
| { |
| "start": 406, |
| "end": 427, |
| "text": "Simpson et al. (2013)", |
| "ref_id": "BIBREF29" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion and Future Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "In this paper, we call human judgments annotations to distinguish them from gold standard class labels.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "http://www.crowdflower.com/data-for-everyone", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "The dataset is available via git at git://nlp.cs. byu.edu/plf1/crowdflower-newsgroups.git", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "Acknowledgments This work was supported by the collaborative NSF Grant IIS-1409739 (BYU) and IIS-1409287 (UMD). Boyd-Graber is also supported by NSF grants IIS-1320538 and NCSE-1422492. Any opinions, findings, conclusions, or recommendations expressed here are those of the authors and do not necessarily reflect the view of the sponsor.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "acknowledgement", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "On smoothing and inference for topic models", |
| "authors": [ |
| { |
| "first": "Arthur", |
| "middle": [], |
| "last": "Asuncion", |
| "suffix": "" |
| }, |
| { |
| "first": "Max", |
| "middle": [], |
| "last": "Welling", |
| "suffix": "" |
| }, |
| { |
| "first": "Padhraic", |
| "middle": [], |
| "last": "Smyth", |
| "suffix": "" |
| }, |
| { |
| "first": "Yee Whye", |
| "middle": [], |
| "last": "Teh", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Proceedings of Uncertainty in Artificial Intelligence", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Arthur Asuncion, Max Welling, Padhraic Smyth, and Yee Whye Teh. 2009. On smoothing and inference for topic models. In Proceedings of Uncertainty in Artificial Intelligence.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Topic annotated Enron email data set. Linguistic Data Consortium", |
| "authors": [ |
| { |
| "first": "Michael", |
| "middle": [ |
| "W" |
| ], |
| "last": "Berry", |
| "suffix": "" |
| }, |
| { |
| "first": "Murray", |
| "middle": [], |
| "last": "Browne", |
| "suffix": "" |
| }, |
| { |
| "first": "Ben", |
| "middle": [], |
| "last": "Signer", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Michael W. Berry, Murray Browne, and Ben Signer. 2001. Topic annotated Enron email data set. Lin- guistic Data Consortium.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "On the statistical analysis of dirty pictures", |
| "authors": [ |
| { |
| "first": "Julian", |
| "middle": [], |
| "last": "Besag", |
| "suffix": "" |
| } |
| ], |
| "year": 1986, |
| "venue": "Journal of the Royal Statistical Society", |
| "volume": "48", |
| "issue": "3", |
| "pages": "259--302", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Julian Besag. 1986. On the statistical analysis of dirty pictures. Journal of the Royal Statistical Society, 48(3):259-302.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Latent Dirichlet allocation", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [ |
| "M" |
| ], |
| "last": "Blei", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew", |
| "middle": [ |
| "Y" |
| ], |
| "last": "Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [ |
| "I" |
| ], |
| "last": "Jordan", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Journal of Machine Learning Research", |
| "volume": "3", |
| "issue": "", |
| "pages": "993--1022", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "David M. Blei, Andrew Y. Ng, and Michael I. Jordan. 2003. Latent Dirichlet allocation. Journal of Ma- chine Learning Research, 3:993-1022.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Crowdsourcing multi-label classification for taxonomy creation", |
| "authors": [ |
| { |
| "first": "Jonathan", |
| "middle": [], |
| "last": "Bragg", |
| "suffix": "" |
| }, |
| { |
| "first": "Mausam", |
| "middle": [], |
| "last": "", |
| "suffix": "" |
| }, |
| { |
| "first": "Daniel", |
| "middle": [ |
| "S" |
| ], |
| "last": "Weld", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "First AAAI Conference on Human Computation and Crowdsourcing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jonathan Bragg, Mausam, and Daniel S. Weld. 2013. Crowdsourcing multi-label classification for taxon- omy creation. In First AAAI Conference on Human Computation and Crowdsourcing.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Improving Methods for Single-label Text Categorization", |
| "authors": [ |
| { |
| "first": "Ana", |
| "middle": [], |
| "last": "Margarida De", |
| "suffix": "" |
| }, |
| { |
| "first": "Jesus", |
| "middle": [], |
| "last": "Cardoso-Cachopo", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ana Margarida de Jesus Cardoso-Cachopo. 2007. Im- proving Methods for Single-label Text Categoriza- tion. Ph.D. thesis, Universidade Tecnica de Lisboa.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Aggregating crowdsourced binary ratings", |
| "authors": [ |
| { |
| "first": "Nilesh", |
| "middle": [], |
| "last": "Dalvi", |
| "suffix": "" |
| }, |
| { |
| "first": "Anirban", |
| "middle": [], |
| "last": "Dasgupta", |
| "suffix": "" |
| }, |
| { |
| "first": "Ravi", |
| "middle": [], |
| "last": "Kumar", |
| "suffix": "" |
| }, |
| { |
| "first": "Vibhor", |
| "middle": [], |
| "last": "Rastogi", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of World Wide Web Conference", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nilesh Dalvi, Anirban Dasgupta, Ravi Kumar, and Vib- hor Rastogi. 2013. Aggregating crowdsourced bi- nary ratings. In Proceedings of World Wide Web Conference.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Maximum likelihood estimation of observer error-rates using the EM algorithm", |
| "authors": [ |
| { |
| "first": "Alexander", |
| "middle": [ |
| "P" |
| ], |
| "last": "Dawid", |
| "suffix": "" |
| }, |
| { |
| "first": "Allan", |
| "middle": [ |
| "M" |
| ], |
| "last": "Skene", |
| "suffix": "" |
| } |
| ], |
| "year": 1979, |
| "venue": "Applied Statistics", |
| "volume": "", |
| "issue": "", |
| "pages": "20--28", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alexander P. Dawid and Allan M. Skene. 1979. Max- imum likelihood estimation of observer error-rates using the EM algorithm. Applied Statistics, pages 20-28.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "MomResp: A Bayesian model for multi-annotator document labeling", |
| "authors": [ |
| { |
| "first": "Paul", |
| "middle": [], |
| "last": "Felt", |
| "suffix": "" |
| }, |
| { |
| "first": "Robbie", |
| "middle": [], |
| "last": "Haertel", |
| "suffix": "" |
| }, |
| { |
| "first": "Eric", |
| "middle": [], |
| "last": "Ringger", |
| "suffix": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Seppi", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "International Language Resources and Evaluation", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Paul Felt, Robbie Haertel, Eric Ringger, and Kevin Seppi. 2014. MomResp: A Bayesian model for multi-annotator document labeling. In International Language Resources and Evaluation.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Early gains matter: A case for preferring generative over discriminative crowdsourcing models", |
| "authors": [ |
| { |
| "first": "Paul", |
| "middle": [], |
| "last": "Felt", |
| "suffix": "" |
| }, |
| { |
| "first": "Eric", |
| "middle": [], |
| "last": "Ringger", |
| "suffix": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Seppi", |
| "suffix": "" |
| }, |
| { |
| "first": "Robbie", |
| "middle": [], |
| "last": "Haertel", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Conference of the North American Chapter", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Paul Felt, Eric Ringger, Kevin Seppi, and Robbie Haer- tel. 2015. Early gains matter: A case for preferring generative over discriminative crowdsourcing mod- els. In Conference of the North American Chapter of the Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Who moderates the moderators?: crowdsourcing abuse detection in user-generated content", |
| "authors": [ |
| { |
| "first": "Arpita", |
| "middle": [], |
| "last": "Ghosh", |
| "suffix": "" |
| }, |
| { |
| "first": "Satyen", |
| "middle": [], |
| "last": "Kale", |
| "suffix": "" |
| }, |
| { |
| "first": "Preston", |
| "middle": [], |
| "last": "Mcafee", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the 12th ACM conference on Electronic commerce", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Arpita Ghosh, Satyen Kale, and Preston McAfee. 2011. Who moderates the moderators?: crowd- sourcing abuse detection in user-generated content. In Proceedings of the 12th ACM conference on Elec- tronic commerce.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Learning whom to trust with MACE", |
| "authors": [ |
| { |
| "first": "Dirk", |
| "middle": [], |
| "last": "Hovy", |
| "suffix": "" |
| }, |
| { |
| "first": "Taylor", |
| "middle": [], |
| "last": "Berg-Kirkpatrick", |
| "suffix": "" |
| }, |
| { |
| "first": "Ashish", |
| "middle": [], |
| "last": "Vaswani", |
| "suffix": "" |
| }, |
| { |
| "first": "Eduard", |
| "middle": [ |
| "H" |
| ], |
| "last": "Hovy", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Conference of the North American Chapter of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dirk Hovy, Taylor Berg-Kirkpatrick, Ashish Vaswani, and Eduard H. Hovy. 2013. Learning whom to trust with MACE. In Conference of the North American Chapter of the Association for Computational Lin- guistics.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Learning with multiple labels", |
| "authors": [ |
| { |
| "first": "Rong", |
| "middle": [], |
| "last": "Jin", |
| "suffix": "" |
| }, |
| { |
| "first": "Zoubin", |
| "middle": [], |
| "last": "Ghahramani", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "Proceedings of Advances in Neural Information Processing Systems", |
| "volume": "", |
| "issue": "", |
| "pages": "897--904", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Rong Jin and Zoubin Ghahramani. 2002. Learning with multiple labels. In Proceedings of Advances in Neural Information Processing Systems, pages 897- 904.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Embracing ambiguity: A comparison of annotation methodologies for crowdsourcing word sense labels", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Jurgens", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of NAACL-HLT", |
| "volume": "", |
| "issue": "", |
| "pages": "556--562", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "David Jurgens. 2013. Embracing ambiguity: A comparison of annotation methodologies for crowd- sourcing word sense labels. In Proceedings of NAACL-HLT, pages 556-562.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Efficient crowdsourcing for multi-class labeling", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [ |
| "R" |
| ], |
| "last": "Karger", |
| "suffix": "" |
| }, |
| { |
| "first": "Sewoong", |
| "middle": [], |
| "last": "Oh", |
| "suffix": "" |
| }, |
| { |
| "first": "Devavrat", |
| "middle": [], |
| "last": "Shah", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "In ACM SIGMETRICS Performance Evaluation Review", |
| "volume": "41", |
| "issue": "", |
| "pages": "81--92", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "David R. Karger, Sewoong Oh, and Devavrat Shah. 2013. Efficient crowdsourcing for multi-class label- ing. In ACM SIGMETRICS Performance Evaluation Review, volume 41, pages 81-92. ACM.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Toward optimal labeling strategy under multiple unreliable labelers", |
| "authors": [ |
| { |
| "first": "Chuck", |
| "middle": [ |
| "P" |
| ], |
| "last": "Lam", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [ |
| "G" |
| ], |
| "last": "Stork", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "AAAI Spring Symposium: Knowledge Collection from Volunteer Contributors", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Chuck P. Lam and David G. Stork. 2005. Toward optimal labeling strategy under multiple unreliable labelers. In AAAI Spring Symposium: Knowledge Collection from Volunteer Contributors.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Predicting economic indicators from web text using sentiment composition", |
| "authors": [ |
| { |
| "first": "Abby", |
| "middle": [], |
| "last": "Levenberg", |
| "suffix": "" |
| }, |
| { |
| "first": "Stephen", |
| "middle": [], |
| "last": "Pulman", |
| "suffix": "" |
| }, |
| { |
| "first": "Karo", |
| "middle": [], |
| "last": "Moilanen", |
| "suffix": "" |
| }, |
| { |
| "first": "Edwin", |
| "middle": [], |
| "last": "Simpson", |
| "suffix": "" |
| }, |
| { |
| "first": "Stephen", |
| "middle": [], |
| "last": "Roberts", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "International Journal of Computer and Communication Engineering", |
| "volume": "3", |
| "issue": "2", |
| "pages": "109--115", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Abby Levenberg, Stephen Pulman, Karo Moilanen, Edwin Simpson, and Stephen Roberts. 2014. Predicting economic indicators from web text us- ing sentiment composition. International Jour- nal of Computer and Communication Engineering, 3(2):109-115.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "To re (label), or not to re (label)", |
| "authors": [ |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Christopher", |
| "suffix": "" |
| }, |
| { |
| "first": "Mausam", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| }, |
| { |
| "first": "Daniel", |
| "middle": [ |
| "S" |
| ], |
| "last": "Weld", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Second AAAI Conference on Human Computation and Crowdsourcing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Christopher H. Lin, Mausam, and Daniel S. Weld. 2014. To re (label), or not to re (label). In Sec- ond AAAI Conference on Human Computation and Crowdsourcing.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Variational inference for crowdsourcing", |
| "authors": [ |
| { |
| "first": "Qiang", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Jian", |
| "middle": [], |
| "last": "Peng", |
| "suffix": "" |
| }, |
| { |
| "first": "Alex", |
| "middle": [ |
| "T" |
| ], |
| "last": "Ihler", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Proceedings of Advances in Neural Information Processing Systems", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Qiang Liu, Jian Peng, and Alex T. Ihler. 2012. Varia- tional inference for crowdsourcing. In Proceedings of Advances in Neural Information Processing Sys- tems.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Supervised topic models", |
| "authors": [ |
| { |
| "first": "Jon", |
| "middle": [ |
| "D" |
| ], |
| "last": "Mcauliffe", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Blei", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proceedings of Advances in Neural Information Processing Systems", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jon D. Mcauliffe and David Blei. 2007. Supervised topic models. In Proceedings of Advances in Neural Information Processing Systems.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "Mallet: A machine learning for language toolkit", |
| "authors": [ |
| { |
| "first": "Andrew", |
| "middle": [], |
| "last": "Mccallum", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Andrew McCallum. 2002. Mallet: A machine learning for language toolkit. http://mallet.cs.umass.edu.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "Estimating a Dirichlet distribution", |
| "authors": [ |
| { |
| "first": "Thomas", |
| "middle": [], |
| "last": "Minka", |
| "suffix": "" |
| } |
| ], |
| "year": 2000, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Thomas Minka. 2000. Estimating a Dirichlet distribu- tion.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "On discriminative vs. generative classifiers: A comparison of logistic regression and Naive Bayes", |
| "authors": [ |
| { |
| "first": "Y", |
| "middle": [], |
| "last": "Andrew", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [ |
| "I" |
| ], |
| "last": "Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Jordan", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of Advances in Neural Information Processing Systems", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Andrew Y. Ng and Michael I. Jordan. 2001. On dis- criminative vs. generative classifiers: A comparison of logistic regression and Naive Bayes. Proceedings of Advances in Neural Information Processing Sys- tems.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "Lexical and hierarchical topic regression", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Viet-An", |
| "suffix": "" |
| }, |
| { |
| "first": "Jordan", |
| "middle": [ |
| "L" |
| ], |
| "last": "Nguyen", |
| "suffix": "" |
| }, |
| { |
| "first": "Philip", |
| "middle": [], |
| "last": "Boyd-Graber", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Resnik", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of Advances in Neural Information Processing Systems", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Viet-An Nguyen, Jordan L. Boyd-Graber, and Philip Resnik. 2013. Lexical and hierarchical topic regres- sion. In Proceedings of Advances in Neural Infor- mation Processing Systems.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "Semi-supervised text classification using EM. Semi-Supervised Learning", |
| "authors": [ |
| { |
| "first": "Kamal", |
| "middle": [], |
| "last": "Nigam", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew", |
| "middle": [], |
| "last": "Mccallum", |
| "suffix": "" |
| }, |
| { |
| "first": "Tom", |
| "middle": [], |
| "last": "Mitchell", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "33--56", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kamal Nigam, Andrew McCallum, and Tom Mitchell. 2006. Semi-supervised text classification using EM. Semi-Supervised Learning, pages 33-56.", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "The benefits of a model of annotation", |
| "authors": [ |
| { |
| "first": "Rebecca", |
| "middle": [ |
| "J" |
| ], |
| "last": "Passonneau", |
| "suffix": "" |
| }, |
| { |
| "first": "Bob", |
| "middle": [], |
| "last": "Carpenter", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the 7th Linguistic Annotation Workshop and Interoperability with Discourse", |
| "volume": "", |
| "issue": "", |
| "pages": "187--195", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Rebecca J. Passonneau and Bob Carpenter. 2013. The benefits of a model of annotation. In Proceedings of the 7th Linguistic Annotation Workshop and Inter- operability with Discourse, pages 187-195.", |
| "links": null |
| }, |
| "BIBREF26": { |
| "ref_id": "b26", |
| "title": "Knowing what to believe (when you already know something)", |
| "authors": [ |
| { |
| "first": "Jeff", |
| "middle": [], |
| "last": "Pasternack", |
| "suffix": "" |
| }, |
| { |
| "first": "Dan", |
| "middle": [], |
| "last": "Roth", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of International Conference on Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jeff Pasternack and Dan Roth. 2010. Knowing what to believe (when you already know something). In Proceedings of International Conference on Compu- tational Linguistics.", |
| "links": null |
| }, |
| "BIBREF27": { |
| "ref_id": "b27", |
| "title": "Learning from crowds", |
| "authors": [ |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Vikas", |
| "suffix": "" |
| }, |
| { |
| "first": "Shipeng", |
| "middle": [], |
| "last": "Raykar", |
| "suffix": "" |
| }, |
| { |
| "first": "Linda", |
| "middle": [ |
| "H" |
| ], |
| "last": "Yu", |
| "suffix": "" |
| }, |
| { |
| "first": "Gerardo", |
| "middle": [ |
| "Hermosillo" |
| ], |
| "last": "Zhao", |
| "suffix": "" |
| }, |
| { |
| "first": "Charles", |
| "middle": [], |
| "last": "Valadez", |
| "suffix": "" |
| }, |
| { |
| "first": "Luca", |
| "middle": [], |
| "last": "Florin", |
| "suffix": "" |
| }, |
| { |
| "first": "Linda", |
| "middle": [], |
| "last": "Bogoni", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Moy", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Journal of Machine Learning Research", |
| "volume": "11", |
| "issue": "", |
| "pages": "1297--1322", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Vikas C. Raykar, Shipeng Yu, Linda H. Zhao, Ger- ardo Hermosillo Valadez, Charles Florin, Luca Bo- goni, and Linda Moy. 2010. Learning from crowds. Journal of Machine Learning Research, 11:1297- 1322.", |
| "links": null |
| }, |
| "BIBREF28": { |
| "ref_id": "b28", |
| "title": "Bayesian methods for intelligent task assignment in crowdsourcing systems", |
| "authors": [ |
| { |
| "first": "Edwin", |
| "middle": [], |
| "last": "Simpson", |
| "suffix": "" |
| }, |
| { |
| "first": "Stephen", |
| "middle": [], |
| "last": "Roberts", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Decision Making: Uncertainty, Imperfection, Deliberation and Scalability", |
| "volume": "", |
| "issue": "", |
| "pages": "1--32", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Edwin Simpson and Stephen Roberts. 2015. Bayesian methods for intelligent task assignment in crowd- sourcing systems. In Decision Making: Uncer- tainty, Imperfection, Deliberation and Scalability, pages 1-32. Springer.", |
| "links": null |
| }, |
| "BIBREF29": { |
| "ref_id": "b29", |
| "title": "Dynamic bayesian combination of multiple imperfect classifiers", |
| "authors": [ |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Simpson", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Roberts", |
| "suffix": "" |
| }, |
| { |
| "first": "I", |
| "middle": [], |
| "last": "Psorakis", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Smith", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Decision Making and Imperfection", |
| "volume": "", |
| "issue": "", |
| "pages": "1--35", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "E. Simpson, S. Roberts, I. Psorakis, and A. Smith. 2013. Dynamic bayesian combination of multiple imperfect classifiers. In Decision Making and Im- perfection, pages 1-35. Springer.", |
| "links": null |
| }, |
| "BIBREF30": { |
| "ref_id": "b30", |
| "title": "Cheap and fast-but is it good?: Evaluating non-expert annotations for natural language tasks", |
| "authors": [ |
| { |
| "first": "Rion", |
| "middle": [], |
| "last": "Snow", |
| "suffix": "" |
| }, |
| { |
| "first": "O'", |
| "middle": [], |
| "last": "Brendan", |
| "suffix": "" |
| }, |
| { |
| "first": "Daniel", |
| "middle": [], |
| "last": "Connor", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew", |
| "middle": [ |
| "Y" |
| ], |
| "last": "Jurafsky", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Ng", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proceedings of EMNLP. ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Rion Snow, Brendan O'Connor, Daniel Jurafsky, and Andrew Y. Ng. 2008. Cheap and fast-but is it good?: Evaluating non-expert annotations for natu- ral language tasks. In Proceedings of EMNLP. ACL.", |
| "links": null |
| }, |
| "BIBREF31": { |
| "ref_id": "b31", |
| "title": "The Wisdom of Crowds", |
| "authors": [ |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Surowiecki", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "James Surowiecki. 2005. The Wisdom of Crowds. Random House LLC.", |
| "links": null |
| }, |
| "BIBREF32": { |
| "ref_id": "b32", |
| "title": "The multidimensional wisdom of crowds", |
| "authors": [ |
| { |
| "first": "Peter", |
| "middle": [], |
| "last": "Welinder", |
| "suffix": "" |
| }, |
| { |
| "first": "Steve", |
| "middle": [], |
| "last": "Branson", |
| "suffix": "" |
| }, |
| { |
| "first": "Pietro", |
| "middle": [], |
| "last": "Perona", |
| "suffix": "" |
| }, |
| { |
| "first": "Serge", |
| "middle": [ |
| "J" |
| ], |
| "last": "Belongie", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "NIPS", |
| "volume": "", |
| "issue": "", |
| "pages": "2424--2432", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Peter Welinder, Steve Branson, Pietro Perona, and Serge J. Belongie. 2010. The multidimensional wis- dom of crowds. In NIPS, pages 2424-2432.", |
| "links": null |
| }, |
| "BIBREF33": { |
| "ref_id": "b33", |
| "title": "Whose vote should count more: Optimal integration of labels from labelers of unknown expertise", |
| "authors": [ |
| { |
| "first": "Jacob", |
| "middle": [], |
| "last": "Whitehill", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Ting-Fan", |
| "suffix": "" |
| }, |
| { |
| "first": "Jacob", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Bergsma", |
| "suffix": "" |
| }, |
| { |
| "first": "Paul", |
| "middle": [ |
| "L" |
| ], |
| "last": "Javier R Movellan", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Ruvolo", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "NIPS", |
| "volume": "22", |
| "issue": "", |
| "pages": "2035--2043", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jacob Whitehill, Ting-fan Wu, Jacob Bergsma, Javier R Movellan, and Paul L. Ruvolo. 2009. Whose vote should count more: Optimal integration of la- bels from labelers of unknown expertise. NIPS, 22:2035-2043.", |
| "links": null |
| }, |
| "BIBREF34": { |
| "ref_id": "b34", |
| "title": "Learning from multiple annotators with varying expertise", |
| "authors": [ |
| { |
| "first": "Yan", |
| "middle": [], |
| "last": "Yan", |
| "suffix": "" |
| }, |
| { |
| "first": "R\u00f3mer", |
| "middle": [], |
| "last": "Rosales", |
| "suffix": "" |
| }, |
| { |
| "first": "Glenn", |
| "middle": [], |
| "last": "Fung", |
| "suffix": "" |
| }, |
| { |
| "first": "Ramanathan", |
| "middle": [], |
| "last": "Subramanian", |
| "suffix": "" |
| }, |
| { |
| "first": "Jennifer", |
| "middle": [], |
| "last": "Dy", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Machine Learning", |
| "volume": "95", |
| "issue": "", |
| "pages": "291--327", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yan Yan, R\u00f3mer Rosales, Glenn Fung, Ramanathan Subramanian, and Jennifer Dy. 2014. Learn- ing from multiple annotators with varying expertise. Machine Learning, 95(3):291-327.", |
| "links": null |
| }, |
| "BIBREF35": { |
| "ref_id": "b35", |
| "title": "Spectral methods meet em: A provably optimal algorithm for crowdsourcing", |
| "authors": [ |
| { |
| "first": "Yuchen", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "Xi", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Dengyong", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [ |
| "I" |
| ], |
| "last": "Jordan", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Advances in Neural Information Processing Systems", |
| "volume": "27", |
| "issue": "", |
| "pages": "1260--1268", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yuchen Zhang, Xi Chen, Dengyong Zhou, and Michael I. Jordan. 2014. Spectral methods meet em: A provably optimal algorithm for crowd- sourcing. In Advances in Neural Information Pro- cessing Systems 27, pages 1260-1268. Curran As- sociates, Inc.", |
| "links": null |
| }, |
| "BIBREF36": { |
| "ref_id": "b36", |
| "title": "Learning from the wisdom of crowds by minimax entropy", |
| "authors": [ |
| { |
| "first": "Dengyong", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| }, |
| { |
| "first": "Sumit", |
| "middle": [], |
| "last": "Basu", |
| "suffix": "" |
| }, |
| { |
| "first": "Yi", |
| "middle": [], |
| "last": "Mao", |
| "suffix": "" |
| }, |
| { |
| "first": "John", |
| "middle": [ |
| "C" |
| ], |
| "last": "Platt", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "NIPS", |
| "volume": "25", |
| "issue": "", |
| "pages": "2204--2212", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dengyong Zhou, Sumit Basu, Yi Mao, and John C. Platt. 2012. Learning from the wisdom of crowds by minimax entropy. In NIPS, volume 25, pages 2204-2212.", |
| "links": null |
| }, |
| "BIBREF37": { |
| "ref_id": "b37", |
| "title": "Aggregating ordinal labels from crowds by minimax conditional entropy", |
| "authors": [ |
| { |
| "first": "Dengyong", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| }, |
| { |
| "first": "Qiang", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "John", |
| "middle": [], |
| "last": "Platt", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [], |
| "last": "Meek", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the International Conference of Machine Learning", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dengyong Zhou, Qiang Liu, John Platt, and Christo- pher Meek. 2014. Aggregating ordinal labels from crowds by minimax conditional entropy. In Pro- ceedings of the International Conference of Machine Learning.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "text": "ITEMRESP as a plate diagram. Round nodes are random variables. Rectangular nodes are free parameters. Shaded nodes are observed. D, J, C are the number of documents, annotators, and classes, respectively. a symmetric Dirichlet distribution Dir(b (\u03b3)", |
| "num": null, |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "FIGREF1": { |
| "text": "CSLDA as a plate diagram. D, J, C, T are the number of documents, annotators, classes, and topics, respectively. N d is the size of document d. |\u03c6 t | = V , the size of the vocabulary. \u03b7 c is a vector of regression parameters. Reduces to LDA when no annotations a are observed.", |
| "num": null, |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "FIGREF2": { |
| "text": "Hyperparameters optimized via grid search on validation dataFigure 4: Differences among the inferred label accuracy learning curves of three ITEMRESP inference algorithms are reduced when hyperparameters are optimized.", |
| "num": null, |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "FIGREF4": { |
| "text": "Inferred label accuracy of models on sentiment-annotated weather tweets.", |
| "num": null, |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "FIGREF7": { |
| "text": "After combining frequently co-annotated label classesFigure 8: An illustrative failure case. CSLDA, lacking a class label prior, prefers to combine label classes that are highly co-annotated.", |
| "num": null, |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "TABREF0": { |
| "type_str": "table", |
| "text": "Definition of counts and select notation.", |
| "content": "<table><tr><td>1(\u2022) is the indicator function.</td></tr></table>", |
| "html": null, |
| "num": null |
| }, |
| "TABREF1": { |
| "type_str": "table", |
| "text": "Dataset statistics. D is number of documents, C is number of classes, V is number of features, and 1", |
| "content": "<table><tr><td>d N d is average document size. Values are calculated after setting aside 15% as N</td></tr><tr><td>validation data and doing feature selection.</td></tr></table>", |
| "html": null, |
| "num": null |
| } |
| } |
| } |
| } |