| { |
| "paper_id": "P99-1004", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T09:32:10.786722Z" |
| }, |
| "title": "Measures of Distributional Similarity", |
| "authors": [ |
| { |
| "first": "Lillian", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Cornell University Ithaca", |
| "location": { |
| "postCode": "14853-7501", |
| "region": "NY" |
| } |
| }, |
| "email": "llee@cs@edu" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "We study distributional similarity measures for the purpose of improving probability estimation for unseen cooccurrences. Our contributions are threefold: an empirical comparison of a broad range of measures; a classification of similarity functions based on the information that they incorporate; and the introduction of a novel function that is superior at evaluating potential proxy distributions.", |
| "pdf_parse": { |
| "paper_id": "P99-1004", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "We study distributional similarity measures for the purpose of improving probability estimation for unseen cooccurrences. Our contributions are threefold: an empirical comparison of a broad range of measures; a classification of similarity functions based on the information that they incorporate; and the introduction of a novel function that is superior at evaluating potential proxy distributions.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "An inherent problem for statistical methods in natural language processing is that of sparse data --the inaccurate representation in any training corpus of the probability of low frequency events. In particular, reasonable events that happen to not occur in the training set may mistakenly be assigned a probability of zero.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "These unseen events generally make up a substantial portion of novel data; for example, Essen and Steinbiss (1992) report that 12% of the test-set bigrams in a 75%-25% split of one million words did not occur in the training partition.", |
| "cite_spans": [ |
| { |
| "start": 88, |
| "end": 114, |
| "text": "Essen and Steinbiss (1992)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "We consider here the question of how to estimate the conditional cooccurrence probability P(v[n) of an unseen word pair (n, v) drawn from some finite set N x V. Two state-of-the-art technologies are Katz's (1987) backoff method and Jelinek and Mercer's (1980) interpolation method. Both use P(v) to estimate P(v[n) when (n, v) is unseen, essentially ignoring the identity of n.", |
| "cite_spans": [ |
| { |
| "start": 90, |
| "end": 96, |
| "text": "P(v[n)", |
| "ref_id": null |
| }, |
| { |
| "start": 199, |
| "end": 212, |
| "text": "Katz's (1987)", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 232, |
| "end": 259, |
| "text": "Jelinek and Mercer's (1980)", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 308, |
| "end": 314, |
| "text": "P(v[n)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "An alternative approach is distance-weighted averaging, which arrives at an estimate for unseen cooccurrences by combining estimates for 25 cooccurrences involving similar words: 1 /P(v[n) ----~-~mES(n) sim(n, m)P (v[m) ~-]mES(n) sim(n, m) ,", |
| "cite_spans": [ |
| { |
| "start": 214, |
| "end": 219, |
| "text": "(v[m)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "where S(n) is a set of candidate similar words and sim(n, m) is a function of the similarity between n and m. We focus on distributional rather than semantic similarity (e.g., Resnik (1995) ) because the goal of distance-weighted averaging is to smooth probability distributions --although the words \"chance\" and \"probability\" are synonyms, the former may not be a good model for predicting what cooccurrences the latter is likely to participate in. There are many plausible measures of distributional similarity. In previous work (Dagan et al., 1999) , we compared the performance of three different functions: the Jensen-Shannon divergence (total divergence to the average), the L1 norm, and the confusion probability. Our experiments on a frequency-controlled pseudoword disambiguation task showed that using any of the three in a distance-weighted averaging scheme yielded large improvements over Katz's backoff smoothing method in predicting unseen coocurrences. Furthermore, by using a restricted version of model (1) that stripped incomparable parameters, we were able to empirically demonstrate that the confusion probability is fundamentally worse at selecting useful similar words. D. Lin also found that the choice of similarity function can affect the quality of automatically-constructed thesauri to a statistically significant degree (1998a) and the ability to determine common morphological roots by as much as 49% in precision (1998b).", |
| "cite_spans": [ |
| { |
| "start": 176, |
| "end": 189, |
| "text": "Resnik (1995)", |
| "ref_id": "BIBREF26" |
| }, |
| { |
| "start": 531, |
| "end": 551, |
| "text": "(Dagan et al., 1999)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "These empirical results indicate that investigating different similarity measures can lead to improved natural language processing. On the other hand, while there have been many similarity measures proposed and analyzed in the information retrieval literature (Jones and Furnas, 1987) , there has been some doubt expressed in that community that the choice of similarity metric has any practical impact:", |
| "cite_spans": [ |
| { |
| "start": 260, |
| "end": 284, |
| "text": "(Jones and Furnas, 1987)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Several authors have pointed out that the difference in retrieval performance achieved by different measures of association is insignificant, providing that these are appropriately normalised. (van Rijsbergen, 1979, pg. 38) But no contradiction arises because, as van Rijsbergen continues, \"one would expect this since most measures incorporate the same information\". In the language-modeling domain, there is currently no agreed-upon best similarity metric because there is no agreement on what the \"same information\"-the key data that a similarity function should incorporate --is.", |
| "cite_spans": [ |
| { |
| "start": 193, |
| "end": 223, |
| "text": "(van Rijsbergen, 1979, pg. 38)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The overall goal of the work described here was to discover these key characteristics. To this end, we first compared a number of common similarity measures, evaluating them in a parameter-free way on a decision task. When grouped by average performance, they fell into several coherent classes, which corresponded to the extent to which the functions focused on the intersection of the supports (regions of positive probability) of the distributions. Using this insight, we developed an information-theoretic metric, the skew divergence, which incorporates the support-intersection data in an asymmetric fashion. This function yielded the best performance overall: an average error rate reduction of 4% (significant at the .01 level) with respect to the Jensen-Shannon divergence, the best predictor of unseen events in our earlier experiments (Dagan et al., 1999) .", |
| "cite_spans": [ |
| { |
| "start": 845, |
| "end": 865, |
| "text": "(Dagan et al., 1999)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Our contributions are thus three-fold: an empirical comparison of a broad range of similarity metrics using an evaluation methodology that factors out inessential degrees of freedom; a proposal, building on this comparison, of a characteristic for classifying similarity functions; and the introduction of a new similarity metric incorporating this characteristic that is superior at evaluating potential proxy distributions.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this section, we describe the seven distributional similarity functions we initally evaluated. 2 For concreteness, we choose N and V to be the set of nouns and the set of transitive verbs, respectively; a cooccurrence pair (n, v) results when n appears as the head noun of the direct object of v. We use P to denote probabilities assigned by a base language model (in our experiments, we simply used unsmoothed relative frequencies derived from training corpus counts).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2{} 2 Distributional Similarity Functions", |
| "sec_num": null |
| }, |
| { |
| "text": "Let n and m be two nouns whose distributional similarity is to be determined; for notational simplicity, we write q(v) for P(vln ) and r(v) for P(vlm), their respective conditional verb cooccurrence probabilities. Figure 1 lists several familiar functions. The cosine metric and Jaccard's coefficient are commonly used in information retrieval as measures of association (Salton and McGill, 1983) . Note that Jaccard's coefficient differs from all the other measures we consider in that it is essentially combinatorial, being based only on the sizes of the supports of q, r, and q \u2022 r rather than the actual values of the distributions.", |
| "cite_spans": [ |
| { |
| "start": 371, |
| "end": 396, |
| "text": "(Salton and McGill, 1983)", |
| "ref_id": "BIBREF28" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 214, |
| "end": 222, |
| "text": "Figure 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "2{} 2 Distributional Similarity Functions", |
| "sec_num": null |
| }, |
| { |
| "text": "Previously, we found the Jensen-Shannon divergence (Rao, 1982; J. Lin, 1991) to be a useful measure of the distance between distributions:", |
| "cite_spans": [ |
| { |
| "start": 51, |
| "end": 62, |
| "text": "(Rao, 1982;", |
| "ref_id": "BIBREF25" |
| }, |
| { |
| "start": 63, |
| "end": 76, |
| "text": "J. Lin, 1991)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2{} 2 Distributional Similarity Functions", |
| "sec_num": null |
| }, |
| { |
| "text": "JS(q,r)=-~l [D(q aVgq,r)+D(r aVgq,r) ]", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2{} 2 Distributional Similarity Functions", |
| "sec_num": null |
| }, |
| { |
| "text": "The function D is the KL divergence, which measures the (always nonnegative) average inefficiency in using one distribution to code for another (Cover and Thomas, 1991) :", |
| "cite_spans": [ |
| { |
| "start": 155, |
| "end": 168, |
| "text": "Thomas, 1991)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2{} 2 Distributional Similarity Functions", |
| "sec_num": null |
| }, |
| { |
| "text": "(v) D(pl(V) IIp2(V)) = EPl(V)log Pl p2(v) \" V", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2{} 2 Distributional Similarity Functions", |
| "sec_num": null |
| }, |
| { |
| "text": "The function avga, r denotes the average distribution avgq,r(V ) --= (q(v)+r(v))/2; observe that its use ensures that the Jensen-Shannon divergence is always defined. In contrast, D(qllr ) is undefined if q is not absolutely continuous with respect to r (i.e., the support of q is not a subset of the support of r).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2{} 2 Distributional Similarity Functions", |
| "sec_num": null |
| }, |
| { |
| "text": "2Strictly speaking, some of these functions are dissimilarity measures, but each such function f can be recast as a similarity function via the simple transformation C -f, where C is an appropriate constant. Whether we mean f or C -f should be clear from context. ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2{} 2 Distributional Similarity Functions", |
| "sec_num": null |
| }, |
| { |
| "text": "L2(q,r) = Ll(q,r) = cos(q, r) = Jac(q, r) = ~v (q(v) -r(v)) 2 Iq(v) -r(v)l V ~-~v q(v)r(v) X/~-~v q(v) 2 V/Y~-v r(v) 2 I{v : q(v) > 0 and r(v) > 0}l I{v I q(v) > 0 or r(v) > O}l Figure 1: Well-known functions", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2{} 2 Distributional Similarity Functions", |
| "sec_num": null |
| }, |
| { |
| "text": "The confusion probability has been used by several authors to smooth word cooccurrence probabilities (Sugawara et al., 1985; Essen and Steinbiss, 1992; Grishman and Sterling, 1993) ; it measures the degree to which word m can be substituted into the contexts in which n appears. If the base language model probabilities obey certain Bayesian consistency conditions (Dagan et al., 1999) , as is the case for relative frequencies, then we may write the confusion probability as follows:", |
| "cite_spans": [ |
| { |
| "start": 101, |
| "end": 124, |
| "text": "(Sugawara et al., 1985;", |
| "ref_id": "BIBREF31" |
| }, |
| { |
| "start": 125, |
| "end": 151, |
| "text": "Essen and Steinbiss, 1992;", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 152, |
| "end": 180, |
| "text": "Grishman and Sterling, 1993)", |
| "ref_id": null |
| }, |
| { |
| "start": 365, |
| "end": 385, |
| "text": "(Dagan et al., 1999)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2{} 2 Distributional Similarity Functions", |
| "sec_num": null |
| }, |
| { |
| "text": "P(m) conf(q, r, P(m) ) = E q(v)r(v) -p-~(v) \" V", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2{} 2 Distributional Similarity Functions", |
| "sec_num": null |
| }, |
| { |
| "text": "Note that it incorporates unigram probabilities as well as the two distributions q and r.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2{} 2 Distributional Similarity Functions", |
| "sec_num": null |
| }, |
| { |
| "text": "Finally, Kendall's % which appears in work on clustering similar adjectives (Hatzivassiloglou and McKeown, 1993; Hatzivassiloglou, 1996) , is a nonparametric measure of the association between random variables (Gibbons, 1993) . In our context, it looks for correlation between the behavior of q and r on pairs of verbs. Three versions exist; we use the simplest, Ta, here:", |
| "cite_spans": [ |
| { |
| "start": 76, |
| "end": 112, |
| "text": "(Hatzivassiloglou and McKeown, 1993;", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 113, |
| "end": 136, |
| "text": "Hatzivassiloglou, 1996)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 210, |
| "end": 225, |
| "text": "(Gibbons, 1993)", |
| "ref_id": "BIBREF7" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2{} 2 Distributional Similarity Functions", |
| "sec_num": null |
| }, |
| { |
| "text": "r(q,r) = E sign [(q(vl) -q(v2))(r(vl) -r(v2))] v,,v 2(l t)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2{} 2 Distributional Similarity Functions", |
| "sec_num": null |
| }, |
| { |
| "text": "where sign(x) is 1 for positive arguments, -1 for negative arguments, and 0 at 0. The intuition behind Kendall's T is as follows. Assume all verbs have distinct conditional probabilities. If sorting the verbs by the likelihoods assigned by q yields exactly the same ordering as that which results from ranking them according to r, then T(q, r) = 1; if it yields exactly the opposite ordering, then T(q, r) ---1. We treat a value of -1 as indicating extreme dissimilarity. 3 It is worth noting at this point that there are several well-known measures from the NLP literature that we have omitted from our experiments. Arguably the most widely used is the mutual information (Hindle, 1990; Church and Hanks, 1990; Dagan et al., 1995; Luk, 1995; D. Lin, 1998a) . It does not apply in the present setting because it does not measure the similarity between two arbitrary probability distributions (in our case, P(VIn ) and P(VIm)) , but rather the similarity between a joint distribution P(X1,X2) and the corresponding product distribution P(X1)P(X2).", |
| "cite_spans": [ |
| { |
| "start": 673, |
| "end": 687, |
| "text": "(Hindle, 1990;", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 688, |
| "end": 711, |
| "text": "Church and Hanks, 1990;", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 712, |
| "end": 731, |
| "text": "Dagan et al., 1995;", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 732, |
| "end": 742, |
| "text": "Luk, 1995;", |
| "ref_id": "BIBREF22" |
| }, |
| { |
| "start": 743, |
| "end": 757, |
| "text": "D. Lin, 1998a)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2{} 2 Distributional Similarity Functions", |
| "sec_num": null |
| }, |
| { |
| "text": "Hamming-type metrics (Cardie, 1993; Zavrel and Daelemans, 1997) are intended for data with symbolic features, since they count feature label mismatches, whereas we are dealing feature Values that are probabilities. Variations of the value difference metric (Stanfill and Waltz, 1986 ) have been employed for supervised disambiguation (Ng and H.B. Lee, 1996; Ng, 1997) ; but it is not reasonable in language modeling to expect training data tagged with correct probabilities. The Dice coej~cient (Smadja et al., 1996; D. Lin, 1998a D. Lin, , 1998b is monotonic in Jaccard's coefficient (van Rijsbergen, 1979) , so its inclusion in our experiments would be redundant. Finally, we did not use the KL divergence because it requires a smoothed base language model. SZero would also be a reasonable choice, since it indicates zero correlation between q and r. However, it would then not be clear how to average in the estimates of negatively correlated words in equation 1.", |
| "cite_spans": [ |
| { |
| "start": 21, |
| "end": 35, |
| "text": "(Cardie, 1993;", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 36, |
| "end": 63, |
| "text": "Zavrel and Daelemans, 1997)", |
| "ref_id": "BIBREF33" |
| }, |
| { |
| "start": 257, |
| "end": 282, |
| "text": "(Stanfill and Waltz, 1986", |
| "ref_id": "BIBREF30" |
| }, |
| { |
| "start": 347, |
| "end": 357, |
| "text": "Lee, 1996;", |
| "ref_id": "BIBREF23" |
| }, |
| { |
| "start": 358, |
| "end": 367, |
| "text": "Ng, 1997)", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 495, |
| "end": 516, |
| "text": "(Smadja et al., 1996;", |
| "ref_id": "BIBREF29" |
| }, |
| { |
| "start": 517, |
| "end": 530, |
| "text": "D. Lin, 1998a", |
| "ref_id": null |
| }, |
| { |
| "start": 531, |
| "end": 546, |
| "text": "D. Lin, , 1998b", |
| "ref_id": null |
| }, |
| { |
| "start": 585, |
| "end": 607, |
| "text": "(van Rijsbergen, 1979)", |
| "ref_id": "BIBREF32" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2{} 2 Distributional Similarity Functions", |
| "sec_num": null |
| }, |
| { |
| "text": "We evaluated the similarity functions introduced in the previous section on a binary decision task, using the same experimental framework as in our previous preliminary compari- son (Dagan et al., 1999) . That is, the data consisted of the verb-object cooccurrence pairs in the 1988 Associated Press newswire involving the 1000 most frequent nouns, extracted via Church's (1988) and Yarowsky's processing tools. 587,833 (80%) of the pairs served as a training set from which to calculate base probabilities. From the other 20%, we prepared test sets as follows: after discarding pairs occurring in the training data (after all, the point of similarity-based estimation is to deal with unseen pairs), we split the remaining pairs into five partitions, and replaced each nounverb pair (n, vl) with a noun-verb-verb triple (n, vl, v2) such that P(v2) ~ P(vl). The task for the language model under evaluation was to reconstruct which of (n, vl) and (n, v2) was the original cooccurrence. Note that by construction, (n, Vl) was always the correct answer, and furthermore, methods relying solely on unigram frequencies would perform no better than chance. Test-set performance was measured by the error rate, defined as T(# of incorrect choices + (# of ties)/2), where T is the number of test triple tokens in the set, and a tie results when both alternatives are deemed equally likely by the language model in question.", |
| "cite_spans": [ |
| { |
| "start": 178, |
| "end": 202, |
| "text": "son (Dagan et al., 1999)", |
| "ref_id": null |
| }, |
| { |
| "start": 363, |
| "end": 378, |
| "text": "Church's (1988)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Empirical Comparison", |
| "sec_num": "3" |
| }, |
| { |
| "text": "To perform the evaluation, we incorporated each similarity function into a decision rule as follows. For a given similarity measure f and neighborhood size k, let 3f, k(n) denote the k most similar words to n according to f. We define the evidence according to f for the cooccurrence ( n, v~) as", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Empirical Comparison", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Then, the decision rule was to choose the alternative with the greatest evidence.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ef, k(n, vi) = [(m E SLk(n) : P(vilm) > l }l \u2022", |
| "sec_num": null |
| }, |
| { |
| "text": "The reason we used a restricted version of the distance-weighted averaging model was that we sought to discover fundamental differences in behavior. Because we have a binary decision task, Ef,k(n, vl) simply counts the number of k nearest neighbors to n that make the right decision. If we have two functions f and g such that Ef,k(n, Vl) > Eg,k(n, vi), then the k most similar words according to f are on the whole better predictors than the k most similar words according to g; hence, f induces an inherently better similarity ranking for distance-weighted averaging. The difficulty with using the full model (Equation (1)) for comparison purposes is that fundamental differences can be obscured by issues of weighting. For example, suppose the probability estimate ~v(2 -Ll(q, r)). r (v) (suitably normalized) performed poorly. We would not be able to tell whether the cause was an inherent deficiency in the L1 norm or just a poor choice of weight function --perhaps (2-Ll(q,r)) 2 would have yielded better estimates. Figure 2 shows how the average error rate varies with k for the seven similarity metrics introduced above. As previously mentioned, a steeper slope indicates a better similarity ranking.", |
| "cite_spans": [ |
| { |
| "start": 189, |
| "end": 200, |
| "text": "Ef,k(n, vl)", |
| "ref_id": null |
| }, |
| { |
| "start": 787, |
| "end": 790, |
| "text": "(v)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 1022, |
| "end": 1030, |
| "text": "Figure 2", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Ef, k(n, vi) = [(m E SLk(n) : P(vilm) > l }l \u2022", |
| "sec_num": null |
| }, |
| { |
| "text": "All the curves have a generally upward trend but always lie far below backoff (51% error rate). They meet at k = 1000 because Sf, looo(n) is always the set of all nouns. We see that the functions fall into four groups: (1) the L2 norm;", |
| "cite_spans": [ |
| { |
| "start": 126, |
| "end": 137, |
| "text": "Sf, looo(n)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ef, k(n, vi) = [(m E SLk(n) : P(vilm) > l }l \u2022", |
| "sec_num": null |
| }, |
| { |
| "text": "(2) Kendall's T; (3) the confusion probability and the cosine metric; and (4) the L1 norm, Jensen-Shannon divergence, and Jaccard's coefficient.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ef, k(n, vi) = [(m E SLk(n) : P(vilm) > l }l \u2022", |
| "sec_num": null |
| }, |
| { |
| "text": "We can account for the similar performance of various metrics by analyzing how they incorporate information from the intersection of the supports of q and r. (Recall that we are using q and r for the conditional verb cooccurrrence probabilities of two nouns n and m.) Consider the following supports (illustrated in Figure 3) :", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 316, |
| "end": 325, |
| "text": "Figure 3)", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Ef, k(n, vi) = [(m E SLk(n) : P(vilm) > l }l \u2022", |
| "sec_num": null |
| }, |
| { |
| "text": "Vq = {veV : q(v)>O} = {v\u2022V:r(v)>0} Yqr = {v \u2022 V : q(v)r(v) > 0} = Yq n", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ef, k(n, vi) = [(m E SLk(n) : P(vilm) > l }l \u2022", |
| "sec_num": null |
| }, |
| { |
| "text": "We can rewrite the similarity functions from Section 2 in terms of these sets, making use of cos(q, r) We see that for the non-combinatorial functions, the groups correspond to the degree to which the measures rely on the verbs in Vat. The Jensen-Shannon divergence and the L1 norm can be computed simply by knowing the val- and Rousseeuw (1990) note that it is \"extremely sensitive to the effect of one or more outliers\" (pg. 117).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ef, k(n, vi) = [(m E SLk(n) : P(vilm) > l }l \u2022", |
| "sec_num": null |
| }, |
| { |
| "text": "the identities ~-~veyq\\yq~ q(v) + ~veyq~ q(v) = ~'~-v~U~\\Vq~ r(v) + ~v~Vq~ r(v) = 1.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ef, k(n, vi) = [(m E SLk(n) : P(vilm) > l }l \u2022", |
| "sec_num": null |
| }, |
| { |
| "text": "= P(ra) Y] q(v)r(v)/P(v) v e Vq~ = E q(v)r(v)( E q(v) 2 E r(v)2) -1/2 v~ Vqr ve Vq v~ Vr Ll(q,r) JS(q, r) Jac(q, r) = 2--E (Iq(v)-r(v)l-q(v)-r(v)) vE Vqr = log2 + 1 E (h(q(v) + r(v)) -h(q(v)) -h(r(v))) , v ~ Vq~ = IV~l/IV~ u v~l h( x ) = -x log x", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ef, k(n, vi) = [(m E SLk(n) : P(vilm) > l }l \u2022", |
| "sec_num": null |
| }, |
| { |
| "text": "The superior performance of Jac(q, r) seems to underscore the importance of the set Vqr.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ef, k(n, vi) = [(m E SLk(n) : P(vilm) > l }l \u2022", |
| "sec_num": null |
| }, |
| { |
| "text": "Jaccard's coefficient ignores the values of q and r on Vqr; but we see that simply knowing the size of Vqr relative to the supports of q and r leads to good rankings.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ef, k(n, vi) = [(m E SLk(n) : P(vilm) > l }l \u2022", |
| "sec_num": null |
| }, |
| { |
| "text": "Based on the results just described, it appears that it is desirable to have a similarity function that focuses on the verbs that cooccur with both of the nouns being compared. However, we can make a further observation: with the exception of the confusion probability, all the functions we compared are symmetric, that is, f(q, r) -= f (r, q) . But the substitutability of one word for another need not symmetric. For instance, \"fruit\" may be the best possible approximation to \"apple\", but the distribution of \"apple\" may not be a suitable proxy for the distribution of \"fruit\".a In accordance with this insight, we developed a novel asymmetric generalization of the KL divergence, the a-skew divergence: sa (q,r) = D(r [[a'q + (1 -a)-r) for 0 <_ a < 1. It can easily be shown that sa depends only on the verbs in Vat. Note that at a --1, the skew divergence is exactly the KL divergence, and su2 is twice one of the summands of JS (note that it is still asymmetric). 40n a related note, an anonymous reviewer cited the following example from the psychology literature: we can say Smith's lecture is like a sleeping pill, but \"not the other way round\".", |
| "cite_spans": [ |
| { |
| "start": 337, |
| "end": 343, |
| "text": "(r, q)", |
| "ref_id": null |
| }, |
| { |
| "start": 710, |
| "end": 729, |
| "text": "(q,r) = D(r [[a'q +", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Skew Divergence", |
| "sec_num": "4" |
| }, |
| { |
| "text": "We can think of a as a degree of confidence in the empirical distribution q; or, equivalently, (1 -a) can be thought of as controlling the amount by which one smooths q by r. Thus, we can view the skew divergence as an approximation to the KL divergence to be used when sparse data problems would cause the latter measure to be undefined. Figure 4 shows the performance of sa for a = .99. It performs better than all the other functions; the difference with respect to Jaccard's coefficient is statistically significant, according to the paired t-test, at all k (except k = 1000), with significance level .01 at all k except 100, 400, and 1000.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 339, |
| "end": 347, |
| "text": "Figure 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "The Skew Divergence", |
| "sec_num": "4" |
| }, |
| { |
| "text": "In this paper, we empirically evaluated a number of distributional similarity measures, including the skew divergence, and analyzed their information sources. We observed that the ability of a similarity function f(q, r) to select useful nearest neighbors appears to be correlated with its focus on the intersection Vqr of the supports of q and r. This is of interest from a computational point of view because Vqr tends to be a relatively small subset of V, the set of all verbs. Furthermore, it suggests downplaying the role of negative information, which is encoded by verbs appearing with exactly one noun, although the Jaccard coefficient does take this type of information into account.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussion", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Our explicit division of V-space into various support regions has been implicitly considered in other work. Smadja et al. (1996) observe that for two potential mutual translations X and Y, the fact that X occurs with translation Y indicates association; X's occurring with a translation other than Y decreases one's belief in their association; but the absence of both X and Y yields no information. In essence, Smadja et al. argue that information from the union of supports, rather than the just the intersection, is important. D. Lin (1997; 1998a) takes an axiomatic approach to determining the characteristics of a good similarity measure. Starting with a formalization (based on certain assumptions) of the intuition that the similarity between two events depends on both their commonality and their differences, he derives a unique similarity function schema. The Figure 4 : Performance of the skew divergence with respect to the best functions from Figure 2 . definition of commonality is left to the user (several different definitions are proposed for different tasks).", |
| "cite_spans": [ |
| { |
| "start": 108, |
| "end": 128, |
| "text": "Smadja et al. (1996)", |
| "ref_id": "BIBREF29" |
| }, |
| { |
| "start": 533, |
| "end": 543, |
| "text": "Lin (1997;", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 544, |
| "end": 550, |
| "text": "1998a)", |
| "ref_id": "BIBREF19" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 870, |
| "end": 878, |
| "text": "Figure 4", |
| "ref_id": null |
| }, |
| { |
| "start": 956, |
| "end": 964, |
| "text": "Figure 2", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Discussion", |
| "sec_num": "5" |
| }, |
| { |
| "text": "We view the empirical approach taken in this paper as complementary to Lin's. That is, we are working in the context of a particular application, and, while we have no mathematical certainty of the importance of the \"common support\" information, we did not assume it a priori;", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussion", |
| "sec_num": "5" |
| }, |
| { |
| "text": "rather, we let the performance data guide our thinking.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussion", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Finally, we observe that the skew metric seems quite promising. We conjecture that appropriate values for a may inversely correspond to the degree of sparseness in the data, and intend in the future to test this conjecture on larger-scale prediction tasks. We also plan to evaluate skewed versions of the Jensen-Shannon divergence proposed by Rao (1982) and J. Lin (1991) .", |
| "cite_spans": [ |
| { |
| "start": 343, |
| "end": 353, |
| "text": "Rao (1982)", |
| "ref_id": "BIBREF25" |
| }, |
| { |
| "start": 361, |
| "end": 371, |
| "text": "Lin (1991)", |
| "ref_id": "BIBREF21" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussion", |
| "sec_num": "5" |
| }, |
| { |
| "text": "1The term \"similarity-based\", which we have used previously, has been applied to describe other models as well (L.Lee, 1997;Karov and Edelman, 1998).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "Thanks to Claire Cardie, Jon Kleinberg, Fernando Pereira, and Stuart Shieber for helpful discussions, the anonymous reviewers for their insightful comments, Fernando Pereira for access to computational resources at AT&T, and Stuart Shieber for the opportunity to pursue this work at Harvard University under NSF Grant No. IRI9712068.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgements", |
| "sec_num": "6" |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "A case-based approach to knowledge acquisition for domain-specific sentence analysis", |
| "authors": [ |
| { |
| "first": "Claire", |
| "middle": [], |
| "last": "Cardie", |
| "suffix": "" |
| } |
| ], |
| "year": 1993, |
| "venue": "11th National Conference on Artifical Intelligence", |
| "volume": "", |
| "issue": "", |
| "pages": "798--803", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Claire Cardie. 1993. A case-based approach to knowledge acquisition for domain-specific sentence analysis. In 11th National Confer- ence on Artifical Intelligence, pages 798-803.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Word association norms, mutual information, and lexicography", |
| "authors": [ |
| { |
| "first": "Kenneth", |
| "middle": [ |
| "Ward" |
| ], |
| "last": "Church", |
| "suffix": "" |
| }, |
| { |
| "first": "Patrick", |
| "middle": [], |
| "last": "Hanks", |
| "suffix": "" |
| } |
| ], |
| "year": 1990, |
| "venue": "Computational Linguistics", |
| "volume": "16", |
| "issue": "1", |
| "pages": "22--29", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kenneth Ward Church and Patrick Hanks. 1990. Word association norms, mutual in- formation, and lexicography. Computational Linguistics, 16(1):22-29.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "A stochastic parts program and noun phrase parser for unrestricted text", |
| "authors": [ |
| { |
| "first": "Kenneth", |
| "middle": [ |
| "W" |
| ], |
| "last": "Church", |
| "suffix": "" |
| } |
| ], |
| "year": 1988, |
| "venue": "Second Conference on Applied Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "136--143", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kenneth W. Church. 1988. A stochastic parts program and noun phrase parser for un- restricted text. In Second Conference on Applied Natural Language Processing, pages 136-143.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Elements of Information Theory", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Thomas", |
| "suffix": "" |
| }, |
| { |
| "first": "Joy", |
| "middle": [ |
| "A" |
| ], |
| "last": "Cover", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Thomas", |
| "suffix": "" |
| } |
| ], |
| "year": 1991, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Thomas M. Cover and Joy A. Thomas. 1991. Elements of Information Theory. John Wiley.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Contextual word similarity and estimation from sparse data", |
| "authors": [ |
| { |
| "first": "Shanl", |
| "middle": [], |
| "last": "Ido Dagan", |
| "suffix": "" |
| }, |
| { |
| "first": "Shanl", |
| "middle": [], |
| "last": "Marcus", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Markovitch", |
| "suffix": "" |
| } |
| ], |
| "year": 1995, |
| "venue": "Computer Speech and Language", |
| "volume": "9", |
| "issue": "", |
| "pages": "123--152", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ido Dagan, Shanl Marcus, and Shanl Marko- vitch. 1995. Contextual word similarity and estimation from sparse data. Computer Speech and Language, 9:123-152.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Similarity-based models of cooccurrence probabilities", |
| "authors": [ |
| { |
| "first": "Lillian", |
| "middle": [], |
| "last": "Ido Dagan", |
| "suffix": "" |
| }, |
| { |
| "first": "Fernando", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Pereira", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "Machine Learning", |
| "volume": "34", |
| "issue": "", |
| "pages": "43--69", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ido Dagan, Lillian Lee, and Fernando Pereira. 1999. Similarity-based models of cooccur- rence probabilities. Machine Learning, 34(1- 3) :43-69.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Cooccurrence smoothing for stochastic language modeling", |
| "authors": [ |
| { |
| "first": "Ute", |
| "middle": [], |
| "last": "Essen", |
| "suffix": "" |
| }, |
| { |
| "first": "Volker", |
| "middle": [], |
| "last": "Steinbiss", |
| "suffix": "" |
| } |
| ], |
| "year": 1992, |
| "venue": "ICASSP 92", |
| "volume": "1", |
| "issue": "", |
| "pages": "161--164", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ute Essen and Volker Steinbiss. 1992. Co- occurrence smoothing for stochastic language modeling. In ICASSP 92, volume 1, pages 161-164.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Nonparametric Measures of Association. Sage University Paper series on Quantitative Applications in the", |
| "authors": [ |
| { |
| "first": "Jean Dickinson", |
| "middle": [], |
| "last": "Gibbons", |
| "suffix": "" |
| } |
| ], |
| "year": 1993, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jean Dickinson Gibbons. 1993. Nonparametric Measures of Association. Sage University Pa- per series on Quantitative Applications in the", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Smoothing of automatically generated selectional constraints", |
| "authors": [], |
| "year": 1993, |
| "venue": "Human Language Technology: Proceedings of the ARPA Workshop", |
| "volume": "", |
| "issue": "", |
| "pages": "254--259", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Social Sciences, 07-091. Sage Publications. Ralph Grishman and John Sterling. 1993. Smoothing of automatically generated selec- tional constraints. In Human Language Tech- nology: Proceedings of the ARPA Workshop, pages 254-259.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Towards the automatic identification of adjectival scales: Clustering of adjectives according to meaning", |
| "authors": [ |
| { |
| "first": "Vasileios", |
| "middle": [], |
| "last": "Hatzivassiloglou", |
| "suffix": "" |
| }, |
| { |
| "first": "Kathleen", |
| "middle": [], |
| "last": "Mckeown", |
| "suffix": "" |
| } |
| ], |
| "year": 1993, |
| "venue": "31st Annual Meeting of the ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "172--182", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Vasileios Hatzivassiloglou and Kathleen McKe- own. 1993. Towards the automatic identifica- tion of adjectival scales: Clustering of adjec- tives according to meaning. In 31st Annual Meeting of the ACL, pages 172-182.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Do we need linguistics when we have statistics? A comparative analysis of the contributions of linguistic cues to a statistical word grouping system", |
| "authors": [ |
| { |
| "first": "Vasileios", |
| "middle": [], |
| "last": "Hatzivassiloglou", |
| "suffix": "" |
| } |
| ], |
| "year": 1996, |
| "venue": "The Balancing Act", |
| "volume": "", |
| "issue": "", |
| "pages": "67--94", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Vasileios Hatzivassiloglou. 1996. Do we need linguistics when we have statistics? A com- parative analysis of the contributions of lin- guistic cues to a statistical word grouping system. In Judith L. Klavans and Philip Resnik, editors, The Balancing Act, pages 67- 94. MIT Press.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Noun classification from predicate-argument structures", |
| "authors": [ |
| { |
| "first": "Don", |
| "middle": [], |
| "last": "Hindle", |
| "suffix": "" |
| } |
| ], |
| "year": 1990, |
| "venue": "28th Annual Meeting of the A CL", |
| "volume": "", |
| "issue": "", |
| "pages": "268--275", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Don Hindle. 1990. Noun classification from predicate-argument structures. In 28th An- nual Meeting of the A CL, pages 268-275.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Interpolated estimation of Markov source parameters from sparse data", |
| "authors": [ |
| { |
| "first": "Frederick", |
| "middle": [], |
| "last": "Jelinek", |
| "suffix": "" |
| }, |
| { |
| "first": "Robert", |
| "middle": [ |
| "L" |
| ], |
| "last": "Mercer", |
| "suffix": "" |
| } |
| ], |
| "year": 1980, |
| "venue": "Proceedings of the Workshop on Pattern Recognition in Practice", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Frederick Jelinek and Robert L. Mercer. 1980. Interpolated estimation of Markov source pa- rameters from sparse data. In Proceedings of the Workshop on Pattern Recognition in Practice.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Pictures of relevance", |
| "authors": [ |
| { |
| "first": "P", |
| "middle": [], |
| "last": "William", |
| "suffix": "" |
| }, |
| { |
| "first": "George", |
| "middle": [ |
| "W" |
| ], |
| "last": "Jones", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Furnas", |
| "suffix": "" |
| } |
| ], |
| "year": 1987, |
| "venue": "Journal of the American Society for Information Science", |
| "volume": "38", |
| "issue": "6", |
| "pages": "420--442", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "William P. Jones and George W. Furnas. 1987. Pictures of relevance. Journal of the American Society for Information Science, 38(6):420-442.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Similarity-based word sense disambiguation", |
| "authors": [ |
| { |
| "first": "Yael", |
| "middle": [], |
| "last": "Karov", |
| "suffix": "" |
| }, |
| { |
| "first": "Shimon", |
| "middle": [], |
| "last": "Edelman", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "Computational Linguistics", |
| "volume": "24", |
| "issue": "1", |
| "pages": "41--59", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yael Karov and Shimon Edelman. 1998. Similarity-based word sense disambiguation. Computational Linguistics, 24(1):41-59.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Estimation of probabilities from sparse data for the language model component of a speech recognizer", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Slava", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Katz", |
| "suffix": "" |
| } |
| ], |
| "year": 1987, |
| "venue": "IEEE Transactions on Acoustics, Speech and Signal Processing", |
| "volume": "35", |
| "issue": "3", |
| "pages": "400--401", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Slava M. Katz. 1987. Estimation of probabili- ties from sparse data for the language model component of a speech recognizer. IEEE Transactions on Acoustics, Speech and Signal Processing, ASSP-35(3):400--401, March.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Finding Groups in Data: An Introduction to Cluster Analysis", |
| "authors": [ |
| { |
| "first": "Leonard", |
| "middle": [], |
| "last": "Kanfman", |
| "suffix": "" |
| }, |
| { |
| "first": "Peter", |
| "middle": [ |
| "J" |
| ], |
| "last": "Rousseeuw", |
| "suffix": "" |
| } |
| ], |
| "year": 1990, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Leonard Kanfman and Peter J. Rousseeuw. 1990. Finding Groups in Data: An Intro- duction to Cluster Analysis. John Wiley and Sons.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Similarity-Based Approaches to Natural Language Processing", |
| "authors": [ |
| { |
| "first": "Lillian", |
| "middle": [ |
| "Lee" |
| ], |
| "last": "", |
| "suffix": "" |
| } |
| ], |
| "year": 1997, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lillian Lee. 1997. Similarity-Based Approaches to Natural Language Processing. Ph.D. the- sis, Harvard University.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Using syntactic dependency as local context to resolve word sense ambiguity", |
| "authors": [ |
| { |
| "first": "Dekang", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| } |
| ], |
| "year": 1997, |
| "venue": "35th Annual Meeting of the ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "64--71", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dekang Lin. 1997. Using syntactic dependency as local context to resolve word sense ambi- guity. In 35th Annual Meeting of the ACL, pages 64-71.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Automatic retrieval and clustering of similar words", |
| "authors": [ |
| { |
| "first": "Dekang", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "COLING-A CL '98", |
| "volume": "", |
| "issue": "", |
| "pages": "768--773", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dekang Lin. 1998a. Automatic retrieval and clustering of similar words. In COLING-A CL '98, pages 768-773.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "An information theoretic definition of similarity", |
| "authors": [ |
| { |
| "first": "Dekang", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "Machine Learning: Proceedings of the Fiftheenth International Conference (ICML '98)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dekang Lin. 1998b. An information theoretic definition of similarity. In Machine Learn- ing: Proceedings of the Fiftheenth Interna- tional Conference (ICML '98).", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "Divergence measures based on the Shannon entropy", |
| "authors": [ |
| { |
| "first": "Jianhua", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| } |
| ], |
| "year": 1991, |
| "venue": "IEEE Transactions on Information Theory", |
| "volume": "37", |
| "issue": "1", |
| "pages": "145--151", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jianhua Lin. 1991. Divergence measures based on the Shannon entropy. IEEE Transactions on Information Theory, 37(1):145-151.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "Statistical sense disambiguation with relatively small corpora using dictionary definitions", |
| "authors": [ |
| { |
| "first": "Alpha", |
| "middle": [ |
| "K" |
| ], |
| "last": "Luk", |
| "suffix": "" |
| } |
| ], |
| "year": 1995, |
| "venue": "33rd Annual Meeting of the ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "181--188", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alpha K. Luk. 1995. Statistical sense disam- biguation with relatively small corpora using dictionary definitions. In 33rd Annual Meet- ing of the ACL, pages 181-188.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "Integrating multiple knowledge sources to disambiguate word sense: An exemplar-based approach", |
| "authors": [ |
| { |
| "first": "Tou", |
| "middle": [], |
| "last": "Hwee", |
| "suffix": "" |
| }, |
| { |
| "first": "Hian Beng", |
| "middle": [], |
| "last": "Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| } |
| ], |
| "year": 1996, |
| "venue": "3~th Annual Meeting of the ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "40--47", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hwee Tou Ng and Hian Beng Lee. 1996. Inte- grating multiple knowledge sources to disam- biguate word sense: An exemplar-based ap- proach. In 3~th Annual Meeting of the ACL, pages 40--47.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "Exemplar-based word sense disambiguation: Some recent improvements", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Hwee Tou Ng", |
| "suffix": "" |
| } |
| ], |
| "year": 1997, |
| "venue": "Second Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "208--213", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hwee Tou Ng. 1997. Exemplar-based word sense disambiguation: Some recent improve- ments. In Second Conference on Empiri- cal Methods in Natural Language Processing (EMNLP-2), pages 208-213.", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "Diversity: Its measurement, decomposition, apportionment and analysis", |
| "authors": [ |
| { |
| "first": "C", |
| "middle": [ |
| "Radhakrishna" |
| ], |
| "last": "Rao", |
| "suffix": "" |
| } |
| ], |
| "year": 1982, |
| "venue": "SankyhZt: The Indian Journal of Statistics", |
| "volume": "44", |
| "issue": "", |
| "pages": "1--22", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "C. Radhakrishna Rao. 1982. Diversity: Its measurement, decomposition, apportionment and analysis. SankyhZt: The Indian Journal of Statistics, 44(A):1-22.", |
| "links": null |
| }, |
| "BIBREF26": { |
| "ref_id": "b26", |
| "title": "Using information content to evaluate semantic similarity in a taxonomy", |
| "authors": [ |
| { |
| "first": "Philip", |
| "middle": [], |
| "last": "Resnik", |
| "suffix": "" |
| } |
| ], |
| "year": 1995, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Philip Resnik. 1995. Using information content to evaluate semantic similarity in a taxonomy.", |
| "links": null |
| }, |
| "BIBREF27": { |
| "ref_id": "b27", |
| "title": "Proceedings of IJCAI-95", |
| "authors": [], |
| "year": null, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "448--453", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "In Proceedings of IJCAI-95, pages 448-453.", |
| "links": null |
| }, |
| "BIBREF28": { |
| "ref_id": "b28", |
| "title": "Introduction to Modern Information Retrieval", |
| "authors": [ |
| { |
| "first": "Gerard", |
| "middle": [], |
| "last": "Salton", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [ |
| "J" |
| ], |
| "last": "Mcgill", |
| "suffix": "" |
| } |
| ], |
| "year": 1983, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Gerard Salton and Michael J. McGill. 1983. In- troduction to Modern Information Retrieval. McGraw-Hill.", |
| "links": null |
| }, |
| "BIBREF29": { |
| "ref_id": "b29", |
| "title": "Translating collocations for bilingual lexicons: A statistical approach", |
| "authors": [ |
| { |
| "first": "Frank", |
| "middle": [], |
| "last": "Smadja", |
| "suffix": "" |
| }, |
| { |
| "first": "Kathleen", |
| "middle": [ |
| "R" |
| ], |
| "last": "Mckeown", |
| "suffix": "" |
| }, |
| { |
| "first": "Vasileios", |
| "middle": [], |
| "last": "Hatzivassiloglou", |
| "suffix": "" |
| } |
| ], |
| "year": 1996, |
| "venue": "Computational Linguistics", |
| "volume": "22", |
| "issue": "1", |
| "pages": "1--38", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Frank Smadja, Kathleen R. McKeown, and Vasileios Hatzivassiloglou. 1996. Translat- ing collocations for bilingual lexicons: A sta- tistical approach. Computational Linguistics, 22(1):1-38.", |
| "links": null |
| }, |
| "BIBREF30": { |
| "ref_id": "b30", |
| "title": "Toward memory-based reasoning", |
| "authors": [ |
| { |
| "first": "Craig", |
| "middle": [], |
| "last": "Stanfill", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Waltz", |
| "suffix": "" |
| } |
| ], |
| "year": 1986, |
| "venue": "Communications of the ACM", |
| "volume": "29", |
| "issue": "12", |
| "pages": "1213--1228", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Craig Stanfill and David Waltz. 1986. To- ward memory-based reasoning. Communica- tions of the ACM, 29(12):1213-1228.", |
| "links": null |
| }, |
| "BIBREF31": { |
| "ref_id": "b31", |
| "title": "Isolated word recognition using hidden Markov models", |
| "authors": [ |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Sugawara", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Nishimura", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Toshioka", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Okochi", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Kaneko", |
| "suffix": "" |
| } |
| ], |
| "year": 1985, |
| "venue": "ICASSP 85", |
| "volume": "", |
| "issue": "", |
| "pages": "1--4", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "K. Sugawara, M. Nishimura, K. Toshioka, M. Okochi, and T. Kaneko. 1985. Isolated word recognition using hidden Markov mod- els. In ICASSP 85, pages 1-4.", |
| "links": null |
| }, |
| "BIBREF32": { |
| "ref_id": "b32", |
| "title": "Information Retrieval", |
| "authors": [ |
| { |
| "first": "C", |
| "middle": [ |
| "J" |
| ], |
| "last": "Van Rijsbergen", |
| "suffix": "" |
| } |
| ], |
| "year": 1979, |
| "venue": "Butterworths", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "C. J. van Rijsbergen. 1979. Information Re- trieval. Butterworths, second edition.", |
| "links": null |
| }, |
| "BIBREF33": { |
| "ref_id": "b33", |
| "title": "Memory-based learning: Using similarity for smoothing", |
| "authors": [ |
| { |
| "first": "Jakub", |
| "middle": [], |
| "last": "Zavrel", |
| "suffix": "" |
| }, |
| { |
| "first": "Walter", |
| "middle": [], |
| "last": "Daelemans", |
| "suffix": "" |
| } |
| ], |
| "year": 1997, |
| "venue": "35th Annual Meeting of the A CL", |
| "volume": "", |
| "issue": "", |
| "pages": "436--443", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jakub Zavrel and Walter Daelemans. 1997. Memory-based learning: Using similarity for smoothing. In 35th Annual Meeting of the A CL, pages 436-443.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "text": "Similarity metric performance. Errorbars denote the range of error rates over the five test sets. Backoff's average error rate was 51%.", |
| "type_str": "figure", |
| "uris": null, |
| "num": null |
| }, |
| "FIGREF1": { |
| "text": "ues of q and r on Vqr. For the cosine and the confusion probability, the distribution values on Vqr are key, but other information is also incorporated. The statistic Ta takes into account all verbs, including those that occur neither with v Supports on V n nor m. Finally, the Euclidean distance is quadratic in verbs outside Vat; indeed, Kaufman", |
| "type_str": "figure", |
| "uris": null, |
| "num": null |
| }, |
| "TABREF1": { |
| "content": "<table><tr><td>lists</td></tr></table>", |
| "type_str": "table", |
| "html": null, |
| "text": "", |
| "num": null |
| }, |
| "TABREF2": { |
| "content": "<table/>", |
| "type_str": "table", |
| "html": null, |
| "text": "Similarity functions, written in terms of sums over supports and grouped by average performance. \\ denotes set difference; A denotes symmetric set difference.", |
| "num": null |
| } |
| } |
| } |
| } |