| { |
| "paper_id": "P01-1025", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T09:29:47.498974Z" |
| }, |
| "title": "Methods for the Qualitative Evaluation of Lexical Association Measures", |
| "authors": [ |
| { |
| "first": "Stefan", |
| "middle": [], |
| "last": "Evert", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "University of Stuttgart", |
| "location": { |
| "addrLine": "Azenbergstr. 12", |
| "postCode": "D-70174", |
| "settlement": "Stuttgart", |
| "country": "Germany" |
| } |
| }, |
| "email": "evert@ims.uni-stuttgart.de" |
| }, |
| { |
| "first": "Brigitte", |
| "middle": [], |
| "last": "Krenn", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "Austrian Research Institute for Artificial Intelligence (\u00d6FAI)", |
| "institution": "", |
| "location": { |
| "addrLine": "Schottengasse 3 A", |
| "postCode": "1010", |
| "settlement": "Vienna", |
| "country": "Austria" |
| } |
| }, |
| "email": "brigitte@ai.univie.ac.at" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "This paper presents methods for a qualitative, unbiased comparison of lexical association measures and the results we have obtained for adjective-noun pairs and preposition-noun-verb triples extracted from German corpora. In our approach, we compare the entire list of candidates, sorted according to the particular measures, to a reference set of manually identified \"true positives\". We also show how estimates for the very large number of hapaxlegomena and double occurrences can be inferred from random samples.", |
| "pdf_parse": { |
| "paper_id": "P01-1025", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "This paper presents methods for a qualitative, unbiased comparison of lexical association measures and the results we have obtained for adjective-noun pairs and preposition-noun-verb triples extracted from German corpora. In our approach, we compare the entire list of candidates, sorted according to the particular measures, to a reference set of manually identified \"true positives\". We also show how estimates for the very large number of hapaxlegomena and double occurrences can be inferred from random samples.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "In computational linguistics, a variety of (statistical) measures have been proposed for identifying lexical associations between words in lexical tuples extracted from text corpora. Methods used range from pure frequency counts to information theoretic measures and statistical significance tests. While the mathematical properties of those measures have been extensively discussed, 1 the strategies employed for evaluating the identification results are far from adequate. Another crucial but still unsolved issue in statistical collocation identification is the treatment of lowfrequency data.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this paper, we first specify requirements for a qualitative evaluation of lexical association mea-sures (AMs). Based on these requirements, we introduce an experimentation procedure, and discuss the evaluation results for a number of widely used AMs. Finally, methods and strategies for handling low-frequency data are suggested.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The measures 2 -Mutual Information ( \u00a2 \u00a1 ) (Church and Hanks, 1989) , the log-likelihood ratio test (Dunning, 1993) , two statistical tests: t-test and", |
| "cite_spans": [ |
| { |
| "start": 43, |
| "end": 67, |
| "text": "(Church and Hanks, 1989)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 100, |
| "end": 115, |
| "text": "(Dunning, 1993)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "-test, and co-occurrence frequencyare applied to two sets of data: adjective-noun (AdjN) pairs and preposition-noun-verb (PNV) triples, where the AMs are applied to (PN,V) pairs. See section 3 for a description of the base data. For evaluation of the association measures, \u00a6 -best strategies (section 4.1) are supplemented with precision and recall graphs (section 4.2) over the complete data sets. Samples comprising particular frequency strata (high versus low frequencies) are examined (section 4.3). In section 5, methods for the treatment of low-frequency data, single (hapaxlegomena) and double occurrences are discussed. The significance of differences between the AMs is addressed in section 6.", |
| "cite_spans": [ |
| { |
| "start": 165, |
| "end": 171, |
| "text": "(PN,V)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u00a3 \u00a5 \u00a4", |
| "sec_num": null |
| }, |
| { |
| "text": "A standard procedure for the evaluation of AMs is manual judgment of the \u00a6 -best candidates identified in a particular corpus by the measure in question. Typically, the number of true positives (TPs) among the 50 or 100 (or slightly more) highest ranked word combinations is manually identified by a human evaluator, in most cases the author of the paper in which the evaluation is presented. This method leads to a very superficial judgment of AMs for the following reasons:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "State-of-the-art", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "(1) The identification results are based on small subsets of the candidates extracted from the corpus. Consequently, results achieved by individual measures may very well be due to chance (cf. sections 4.1 and 4.2), and evaluation with respect to frequency strata is not possible (cf. section 4.3). (2) For the same reason, it is impossible to determine recall values, which are important for many practical applications. 3The introduction of new measures or changes to the calculation methods require additional manual evaluation, as new \u00a6 -best lists are generated.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "State-of-the-art", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "To improve the reliability of the evaluation results, a number of properties need to be controlled. We distinguish between two classes:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Requirements", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "(1) Characteristics of the set of candidate data employed for collocation identification: (i) the syntactic homogeneity of the base data, i.e., whether the set of candidate data consists only of adjective-noun, noun-verb, etc. pairs or whether different types of word combinations are mixed; (ii) the grammatical status of the individual word combinations in the base set, i.e., whether they are part of or constitute a phrase or simply cooccur within a given text window; (iii) the percentage of TPs in the base set, which is typically higher among high-frequency data than among low-frequency data.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Requirements", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "(2) The evaluation strategies applied: Instead of examining only a small sample of \u00a6 -best candidates for each measure as it is common practice, we make use of recall and precision values for \u00a6 best samples of arbitrary size, which allows us to plot recall and precision curves for the whole set of candidate data. In addition, we compare precision curves for different frequency strata.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Requirements", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "The base data for our experiments are extracted from two corpora which differ with respect to size and text type. The base sets also differ with re-spect to syntactic homogeneity and grammatical correctness. Both candidate sets have been manually inspected for TPs.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Base Data", |
| "sec_num": "3" |
| }, |
| { |
| "text": "The first set comprises bigrams of adjacent, lemmatized AdjN pairs extracted from a small (", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Base Data", |
| "sec_num": "3" |
| }, |
| { |
| "text": "\u00a2 \u00a1 \u00a4 \u00a3 \u00a6 \u00a5 \u00a7 \u00a9", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Base Data", |
| "sec_num": "3" |
| }, |
| { |
| "text": "word) corpus of freely available German law texts. 3 Due to the extraction strategy, the data are homogeneous and grammatically correct, i.e., there is (almost) always a grammatical dependency between adjacent adjectives and nouns in running text. Two human annotators independently marked candidate pairs perceived as \"typical\" combinations, including idioms ((die) hohe See, 'the high seas'), legal terms (\u00fcble Nachrede, 'slander'), and proper names (Rotes Kreuz, 'Red Cross'). Candidates accepted by either one of the annotators were considered TPs.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Base Data", |
| "sec_num": "3" |
| }, |
| { |
| "text": "The second set consists of PNV triples extracted from an 8 million word portion of the Frankfurter Rundschau Corpus 4 , in which partof-speech tags and minimal PPs were identified. 5 The PNV triples were selected automatically such that the preposition and the noun are constituents of the same PP, and the PP and the verb co-occur within a sentence. Only main verbs were considered and full forms were reduced to bases. 6 The PNV data are partially inhomogeneous and not fully grammatically correct, because they include combinations with no grammatical relation between PN and V. PNV collocations were manually annotated. The criteria used for the distinction between collocations and arbitrary word combinations are: There is a grammatical relation between the verb and the PP, and the triple can be interpreted as support verb construction and/or a metaphoric or idiomatic reading is available, e.g.: zur Verf\u00fcgung stellen (at_the availability put, 'make available'), am Herzen liegen (at the heart lie, 'have at heart'). 7 ", |
| "cite_spans": [ |
| { |
| "start": 181, |
| "end": 182, |
| "text": "5", |
| "ref_id": null |
| }, |
| { |
| "start": 421, |
| "end": 422, |
| "text": "6", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Base Data", |
| "sec_num": "3" |
| }, |
| { |
| "text": "After extraction of the base data and manual identification of TPs, the AMs are applied, resulting in an ordered candidate list for each measure (henceforth significance list, SL). The order indicates the degree of collocativity. Multiple candidates with identical scores are listed in random order. This is necessary, in particular, when co-occurrence frequency is used as an association measure.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental Setup", |
| "sec_num": "4" |
| }, |
| { |
| "text": "In this approach, the set of the \u00a6 highest ranked word combinations is evaluated for each measure, and the proportion of TPs among this \u00a6 -best list (the precision) is computed. Another measure of goodness is the proportion of TPs in the base data that are also contained in the \u00a6 -best list (the recall). While precision measures the quality of the \u00a6 -best lists produced, recall measures their coverage, i.e., how many of all true collocations in the corpus were identified. The most problematic aspect here is that conclusions drawn from \u00a6 -best lists for a single (and often small) value of \u00a6 are only snapshots and likely to be misleading.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "4.1\u00a8-Best Lists", |
| "sec_num": null |
| }, |
| { |
| "text": "For instance, considering the set of AdjN base data with \u00a9 \u00a1 \u00a5 we might arrive at the following results ( Table 2 gives the precision values of the \u00a6 highest ranked word combinations with", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 106, |
| "end": 113, |
| "text": "Table 2", |
| "ref_id": "TABREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "4.1\u00a8-Best Lists", |
| "sec_num": null |
| }, |
| { |
| "text": "\u00a6 \u00a1 \u00a7 \u00a7 \u00a7 \u00a7 ):", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "4.1\u00a8-Best Lists", |
| "sec_num": null |
| }, |
| { |
| "text": "As expected from the results of other studies (e.g. Lezius (1999) ", |
| "cite_spans": [ |
| { |
| "start": 52, |
| "end": 65, |
| "text": "Lezius (1999)", |
| "ref_id": "BIBREF7" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "4.1\u00a8-Best Lists", |
| "sec_num": null |
| }, |
| { |
| "text": "For a clearer picture, however, larger portions of the SLs need to be examined. A well suited means for comparing the goodness of different AMs are the precision and recall graphs obtained by stepwise processing of the complete SLs (Figures 1 to 10 below). 9", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Precision and Recall Graphs", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "The -axis represents the percentage of data processed in the respective SL, while the ) contain a particularly small number of TPs. Therefore, they are more susceptible to random variation, which illustrates that evaluation based on a small number of \u00a6 -best candidate pairs cannot be reliable.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Precision and Recall Graphs", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "With respect to the recall curves (Figures 3 and 4 ), we find: (i) Examination of 50% of the data in the SLs leads to identification of between 75% (AdjN) and 80% (PNV) of the TPs. (ii) For the first 40% of the SLs, Examining the precision and recall graphs in more detail, we find that for the AdjN data (Figure 1) , log-likelihood and t-test lead to the best results, with log-likelihood giving an overall better result than the t-test. The picture differs slightly for the PNV data ( Figure 2 ). Here t-test outperforms log-likelihood, and even precision gained by frequency is better than or at least comparable to log-likelihood. These pairings -log-likelihood and t-test for AdjN, and t-test and frequency for PNV -are also visible in the recall curves (Figures 3 and 4) . Moreover, for the PNV data the t-test leads to a recall of over 60% when approx. 20% of the SL has been considered.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 34, |
| "end": 51, |
| "text": "(Figures 3 and 4", |
| "ref_id": "FIGREF3" |
| }, |
| { |
| "start": 306, |
| "end": 316, |
| "text": "(Figure 1)", |
| "ref_id": null |
| }, |
| { |
| "start": 488, |
| "end": 496, |
| "text": "Figure 2", |
| "ref_id": "FIGREF2" |
| }, |
| { |
| "start": 760, |
| "end": 777, |
| "text": "(Figures 3 and 4)", |
| "ref_id": "FIGREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Precision and Recall Graphs", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "In the Figures above, there are a number of positions on the -axis where the precision and recall values of different measures are almost identical. This shows that a simple \u00a6 -best approach will often produce misleading results. For instance, if we just look at the first \u00a9 \u00a7 & of the SLs for the PNV data, we might conclude that the t-test and frequency measures are equally well suited for the extraction of PNV collocations. However, the full curves in Figures 2 and 4 show that t-test is consistently better than frequency.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 457, |
| "end": 472, |
| "text": "Figures 2 and 4", |
| "ref_id": "FIGREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Precision and Recall Graphs", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "While we have previously considered data from a broad frequency range (i.e., frequencies \u00a1 \u00a5 for AdjN and \u00a1 \u00a9 for PNV), we will now split up the candidate sets into high-frequency and low-frequency occurrences. This procedure allows us to assess the performance of AMs within different frequency strata. For instance, there is a widely held belief that \u00a2 \u00a1 and \u00a3 \u00a4 are inferior to other measures because they overestimate the collocativity of low-frequency candidates (cf. the remarks on the \u00a3 \u00a4 measure in (Dunning, 1993) ", |
| "cite_spans": [ |
| { |
| "start": 507, |
| "end": 522, |
| "text": "(Dunning, 1993)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Frequency Strata", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Considering our high-frequency AdjN data (Figure 5) , we find that all precision curves decline as more of the data in the SLs is examined. Especially for \u00a1 , this is markedly different from the results obtained before. As the full curves show, log-likelihood is obviously the best measure . Surprisingly, the precision curves of \u00a3 \u00a4 and in particular \u00a1 increase over the first 60% of the SLs for high-frequency PNV data, whereas the curves for t-test, log-likelihood, and frequency have the usual downward slope (see Figure 6 ). Log-likelihood achieves precision values above 50% for the first 10% of the list, but is outperformed by the t-test afterwards. Looking at the first 40% of the data, there is a big gap between the good measures (t-test, log-likelihood, and frequency) and the weak measures (\u00a3 \u00a4 and", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 41, |
| "end": 51, |
| "text": "(Figure 5)", |
| "ref_id": null |
| }, |
| { |
| "start": 518, |
| "end": 526, |
| "text": "Figure 6", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "High Frequencies", |
| "sec_num": null |
| }, |
| { |
| "text": "). In the second half of the data in the SLs, however, there is virtually no difference between \u00a2 \u00a1 , \u00a3 \u00a5 \u00a4 , and the other measures, with the exception of mere co-occurrence frequency.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u00a2 \u00a1", |
| "sec_num": null |
| }, |
| { |
| "text": "Summing up, t-test -with a few exceptions around the first 5% of the data in the SLsleads to the overall best precision results for high-frequency PNV data. Log-likelihood is second best but achieves the best results for highfrequency AdjN data.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u00a2 \u00a1", |
| "sec_num": null |
| }, |
| { |
| "text": "0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% show that there is little difference between the AMs for low-frequency data, except for co-occurrence frequency, which leads to worse results than all other measures.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Low Frequencies", |
| "sec_num": null |
| }, |
| { |
| "text": "For AdjN data, the AMs at best lead to an improvement of factor 3 compared to random selection (when up to \u00a5 & of the SL is examined, log-likelihood achieves precision values above 30%). Log-likelihood is the overall best measure for identifying AdjN collocations, except forcoordinates between 15% and 20% where t-test outperforms log-likelihood.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Low Frequencies", |
| "sec_num": null |
| }, |
| { |
| "text": "For PNV data, the curves of all measures (except for frequency) are nearly identical. Their precision values are not significantly 10 different from the baseline obtained by random selection.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Low Frequencies", |
| "sec_num": null |
| }, |
| { |
| "text": "In contrast to our expectation stated at the beginning of this section, the performance of \u00a2 \u00a1 and \u00a3 \u00a5 \u00a4 relative to the other AMs is not better for high-frequency data than for low-frequency data. Instead, the poor performance observed in section 4.2 is explained by the considerably higher baseline precision of the high-frequency data (cf. ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Low Frequencies", |
| "sec_num": null |
| }, |
| { |
| "text": "As the frequency distribution of word combinations in texts is characterised by a large number of rare events, low-frequency data are a serious challenge for AMs. One way to deal with lowfrequency candidates is the introduction of cutoff thresholds. This is a widely used strategy, and it is motivated by the fact that it is in general highly problematic to draw conclusions from low-frequency data with statistical methods (cf. Weeber et al. (2000) and Figure 8) . A practical reason for cutting off low-frequency data is the need to reduce the amount of manual work when the complete data set has to be evaluated, which is a precondition for the exact calculation of recall and for plotting precision curves. The major drawback of an approach where all low-frequency candidates are excluded is that a large part of the data is lost for collocation extraction. In our data, for instance, 80% of the full set of PNV data and 58% of the AdjN data are hapaxes. Thus it is important to know how many (and which) true collocations there are among the excluded low-frequency candidates.", |
| "cite_spans": [ |
| { |
| "start": 429, |
| "end": 449, |
| "text": "Weeber et al. (2000)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 454, |
| "end": 463, |
| "text": "Figure 8)", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Hapaxlegomena and Double Occurrences", |
| "sec_num": "5" |
| }, |
| { |
| "text": "In this section, we estimate the number of collocations in the data excluded from our experiments (i.e., AdjN pairs with \u00a1 and PNV triples with", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Statistical Estimation of TPs among Low-Frequency Data", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "). Because of the large number of candidates in those sets (6 435 for AdjN, 10 According to the \u00a7 -test as described in section 6. 279 880 for PNV), manual inspection of the entire data is impractical. Therefore, we use random samples from the candidate sets to obtain estimates for the proportion of true collocations among the low-frequency data. We randomly selected 965 items (15%) from the AdjN hapaxes, and 983 items ( 0.35%) from the low-frequency PNV triples. Manual examination of the samples yielded 31 TPs for AdjN (a proportion of 3.2%) and 6 TPs for PNV (0.6%).", |
| "cite_spans": [ |
| { |
| "start": 76, |
| "end": 78, |
| "text": "10", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u00a1 ) \u00a5", |
| "sec_num": null |
| }, |
| { |
| "text": "Considering the low proportion of collocations in the samples, we must expect highly skewed frequency distributions (where is very small), which are problematic for standard statistical tests. In order to obtain reliable estimates, we have used an exact test based on the following model: Assuming a proportion of TPs in the full low-frequency data (AdjN or PNV), the number of TPs in a random sample of size\u00a8is described by a binomially distributed random variable \u00a1 with parameter . 11 Consequently, the probability of finding \u00a2 or less TPs in the sample is", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u00a1 ) \u00a5", |
| "sec_num": null |
| }, |
| { |
| "text": "\u00a3 \u00a5 \u00a4 \u00a3 \u00a1 \u00a1 \u00a2 \u00a4 \u00a7 \u00a6 \u00a9 \u00a3\u00a1 ! \u00a4 # \"", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u00a1 ) \u00a5", |
| "sec_num": null |
| }, |
| { |
| "text": ". We apply a one-tailed statistical test based on the probabilities at the same confidence level of 99%. However, due to the very large number of low-frequency candidates, there may be as many as 4200 collocations in the PNV data with \u00a1 ) \u00a5 , more than 4 times the number identified in our experiment.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u00a1 ) \u00a5", |
| "sec_num": null |
| }, |
| { |
| "text": "It is imaginable, then, that one of the AMs 11 To be precise, the binomial distribution is itself an approximation of the exact hypergeometric probabilities (cf. Pedersen (1996) ). This approximation is sufficiently accurate as long as the sample size ) is small compared to the size of the base set (i.e., the number of low-frequency candidates). ", |
| "cite_spans": [ |
| { |
| "start": 162, |
| "end": 177, |
| "text": "Pedersen (1996)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u00a1 ) \u00a5", |
| "sec_num": null |
| }, |
| { |
| "text": "\u00a6 \u00a1 \u00a7 \u00a7 \u00a7 \u00a5 \u00a7 \u00a7 \u00a7 \u00a7 \u00a7 \u00a7", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u00a1 ) \u00a5", |
| "sec_num": null |
| }, |
| { |
| "text": "). In order to reduce the amount of manual work, the precision values for each AM are based on a 10% random sample from the 10 000 highest ranked candidates. We have applied the statistical test described above to obtain confidence intervals for the true precision values of the bestperforming AM (frequency), given our 10% sample. The upper and lower bounds of the 95% confidence intervals are shown as thin lines. Even the highest precision estimates fall well below the 6.41% precision baseline of the PNV data with \u00a1 \u00a9 . Again, we conclude that the exclusion of low-frequency candidates was well justified.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u00a1 ) \u00a5", |
| "sec_num": null |
| }, |
| { |
| "text": "We have assessed the significance of differences between AMs using the well-known \u00a3 \u00a4 t est as described in (Krenn, 2000) . 12 The thin lines in Figure 10 delimit 95% confidence intervals around the best-performing measure for the AdjN data with \u00a2 \u00a1 \u00a5 (log-likelihood). There is no significant difference between loglikelihood and t-test. And only for \u00a6 -best lists with \u00a6 \u00a1 \u00a7 \u00a7 \u00a7", |
| "cite_spans": [ |
| { |
| "start": 108, |
| "end": 121, |
| "text": "(Krenn, 2000)", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 124, |
| "end": 126, |
| "text": "12", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 145, |
| "end": 151, |
| "text": "Figure", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Significance Testing", |
| "sec_num": "6" |
| }, |
| { |
| "text": ", frequency performs marginally significantly worse than log-likelihood. For the PNV data (not shown), the t-test is significantly better than log-likelihood, but the difference between frequency and the t-test is at best marginally significant. ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Significance Testing", |
| "sec_num": "6" |
| }, |
| { |
| "text": "We have shown that simple \u00a6 -best approaches are not suitable for a qualitative evaluation of lexical association measures, mainly for the following reasons: the instability of precision values obtained from the first few percent of the data in the SLs; the lack of significant differences between the AMs after approx. 50% of the data in the SLs have been examined; and the lack of significant differences between the measures except for certain specific values of \u00a6 . We have also shown that the evaluation results and the ranking of AMs differ depending on the kind of collocations to be identified, and the proportion of hapaxes in the candidate sets. Finally, our results question the widely accepted argument that the strength of loglikelihood lies in handling low-frequency data. In our experiments, none of the AMs was able to extract a substantial number of collocations from the set of hapaxlegomena.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "7" |
| }, |
| { |
| "text": "See for instance (Manning and Sch\u00fctze, 1999, chapter 5),(Kilgarriff, 1996), and(Pedersen, 1996).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "For a more detailed description of these measures and relevant literature, see(Manning and Sch\u00fctze, 1999, chapter 5) or http://www.collocations.de/EK/, where several other AMs are discussed as well.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "See(Schmid, 1995) for a description of the part-ofspeech tagger used to identify adjectives and nouns in the corpus.4 The Frankfurter Rundschau Corpus is part of the European Corpus Initiative Multilingual Corpus I.5 See(Skut and Brants, 1998) for a description of the tagger and chunker.6 Mmorph -the MULTEXT morphology tool provided by ISSCO/SUISSETRA, Geneva, Switzerland -has been employed for determining verb infinitives.7 For definitions of and literature on idioms, metaphors and support verb constructions (Funktionsverbgef\u00fcge) see for instance(Bu\u00dfmann, 1990).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "Colour versions of all plots in this paper will be available from http://www.collocations.de/EK/", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "See(Krenn and Evert, 2001) for a short discussion of the applicability of this test.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "The work of B. Krenn has been sponsored by the Fonds zur F\u00f6rderung der wissenschaftlichen Forschung (FWF), Grant No. P12920. Financial support for \u00d6FAI is provided by the Austrian Federal Ministry of Education, Science and Culture. The AdjN data is the result of joint research with Ulrich Heid and Wolfgang Lezius.The authors would like to thank the anonymous reviewers for many helpful comments and interesting references.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgement", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Lexikon der Sprachwissenschaft", |
| "authors": [ |
| { |
| "first": "Hadumod", |
| "middle": [], |
| "last": "Bu\u00dfmann", |
| "suffix": "" |
| } |
| ], |
| "year": 1990, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hadumod Bu\u00dfmann. 1990. Lexikon der Sprachwis- senschaft. Kr\u00f6ner, 2nd edition.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Word association norms, mutual information, and lexicography", |
| "authors": [ |
| { |
| "first": "K", |
| "middle": [ |
| "W" |
| ], |
| "last": "Church", |
| "suffix": "" |
| }, |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Hanks", |
| "suffix": "" |
| } |
| ], |
| "year": 1989, |
| "venue": "Proceedings of the 27th Annual Meeting of the Association for Computational Linguistics, Vancouver", |
| "volume": "", |
| "issue": "", |
| "pages": "76--83", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "K.W. Church and P. Hanks. 1989. Word association norms, mutual information, and lexicography. In Proceedings of the 27th Annual Meeting of the As- sociation for Computational Linguistics, Vancou- ver, Canada, 76-83.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Accurate methods for the statistics of surprise and coincidence", |
| "authors": [ |
| { |
| "first": "Ted", |
| "middle": [], |
| "last": "Dunning", |
| "suffix": "" |
| } |
| ], |
| "year": 1993, |
| "venue": "Computational Linguistics", |
| "volume": "19", |
| "issue": "1", |
| "pages": "61--74", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ted Dunning. 1993. Accurate methods for the statis- tics of surprise and coincidence. Computational Linguistics, 19(1):61-74.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Methoden zum Vergleich von Signifikanzma\u00dfen zur Kollokationsidentifikation", |
| "authors": [ |
| { |
| "first": "Stefan", |
| "middle": [], |
| "last": "Evert", |
| "suffix": "" |
| }, |
| { |
| "first": "Ulrich", |
| "middle": [], |
| "last": "Heid", |
| "suffix": "" |
| }, |
| { |
| "first": "Wolfgang", |
| "middle": [], |
| "last": "Lezius", |
| "suffix": "" |
| } |
| ], |
| "year": 2000, |
| "venue": "Proceedings of KONVENS 2000", |
| "volume": "", |
| "issue": "", |
| "pages": "215--220", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Stefan Evert, Ulrich Heid, and Wolfgang Lezius. 2000. Methoden zum Vergleich von Signifikanz- ma\u00dfen zur Kollokationsidentifikation. In Proceed- ings of KONVENS 2000, VDE-Verlag, Germany, pages 215 -220.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Which words are particularly characteristic of a text? A survey of statistical approaches", |
| "authors": [ |
| { |
| "first": "Adam", |
| "middle": [], |
| "last": "Kilgarriff", |
| "suffix": "" |
| } |
| ], |
| "year": 1996, |
| "venue": "Proceedings of the AISB Workshop on Language Engineering for Document Analysis and Recognition", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Adam Kilgarriff. 1996. Which words are particularly characteristic of a text? A survey of statistical ap- proaches. In Proceedings of the AISB Workshop on Language Engineering for Document Analysis and Recognition, Sussex University, GB.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "The Usual Suspects: Data-Oriented Models for the Identification and Representation of Lexical Collocations", |
| "authors": [ |
| { |
| "first": "Brigitte", |
| "middle": [], |
| "last": "Krenn", |
| "suffix": "" |
| } |
| ], |
| "year": 2000, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Brigitte Krenn. 2000. The Usual Suspects: Data- Oriented Models for the Identification and Repre- sentation of Lexical Collocations. DFKI & Univer- sit\u00e4t des Saarlandes, Saarbr\u00fccken.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Can we do better than frequency? A case study on extracting PP-verb collocations", |
| "authors": [ |
| { |
| "first": "Brigitte", |
| "middle": [], |
| "last": "Krenn", |
| "suffix": "" |
| }, |
| { |
| "first": "Stefan", |
| "middle": [], |
| "last": "Evert", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the ACL Workshop on Collocations", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Brigitte Krenn and Stefan Evert. 2001. Can we do better than frequency? A case study on extracting PP-verb collocations. In Proceedings of the ACL Workshop on Collocations, Toulouse, France.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Automatische Extrahierung idiomatischer Bigramme aus Textkorpora", |
| "authors": [ |
| { |
| "first": "Wolfgang", |
| "middle": [], |
| "last": "Lezius", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "Tagungsband des 34. Linguistischen Kolloquiums", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wolfgang Lezius. 1999. Automatische Extrahierung idiomatischer Bigramme aus Textkorpora. In Tagungsband des 34. Linguistischen Kolloquiums, Germersheim.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Foundations of Statistical Natural Language Processing", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Christopher", |
| "suffix": "" |
| }, |
| { |
| "first": "Hinrich", |
| "middle": [], |
| "last": "Manning", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Sch\u00fctze", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Christopher D. Manning and Hinrich Sch\u00fctze. 1999. Foundations of Statistical Natural Language Pro- cessing. MIT Press, Cambridge, MA.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Fishing for Exactness", |
| "authors": [ |
| { |
| "first": "Ted", |
| "middle": [], |
| "last": "Pedersen", |
| "suffix": "" |
| } |
| ], |
| "year": 1996, |
| "venue": "Proceedings of the South-Central SAS Users Group Conference", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ted Pedersen. 1996. Fishing for Exactness. In Pro- ceedings of the South-Central SAS Users Group Conference, Austin, TX.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Improvements in part-ofspeech tagging with an application to german", |
| "authors": [ |
| { |
| "first": "Helmut", |
| "middle": [], |
| "last": "Schmid", |
| "suffix": "" |
| } |
| ], |
| "year": 1995, |
| "venue": "Proceedings of the ACL SIGDAT-Workshop", |
| "volume": "", |
| "issue": "", |
| "pages": "47--50", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Helmut Schmid. 1995. Improvements in part-of- speech tagging with an application to german. In Proceedings of the ACL SIGDAT-Workshop, 47-50.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Chunk Tagger. Stochastic Recognition of Noun Phrases", |
| "authors": [ |
| { |
| "first": "Wojciech", |
| "middle": [], |
| "last": "Skut", |
| "suffix": "" |
| }, |
| { |
| "first": "Thorsten", |
| "middle": [], |
| "last": "Brants", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "ESSLI Workshop on Automated Acquisition of Syntax and Parsing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wojciech Skut and Thorsten Brants. 1998. Chunk Tagger. Stochastic Recognition of Noun Phrases. In ESSLI Workshop on Automated Acquisition of Syn- tax and Parsing, Saarbr\u00fccken, Germany.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Extracting the lowest-frequency words: Pitfalls and possibilities", |
| "authors": [ |
| { |
| "first": "Mark", |
| "middle": [], |
| "last": "Weeber", |
| "suffix": "" |
| }, |
| { |
| "first": "Rein", |
| "middle": [], |
| "last": "Vos", |
| "suffix": "" |
| }, |
| { |
| "first": "Harald", |
| "middle": [ |
| "R" |
| ], |
| "last": "Baayen", |
| "suffix": "" |
| } |
| ], |
| "year": 2000, |
| "venue": "Computational Linguistics", |
| "volume": "", |
| "issue": "3", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Mark Weeber, Rein Vos, and Harald R. Baayen 2000. Extracting the lowest-frequency words: Pitfalls and possibilities. Computational Linguistics, 26(3).", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "text": "precision (or recall) values achieved. For instance, the precision values for vertical lines). The dotted horizontal line represents the percentage of true collocations in the base set. This value corresponds to the expected precision value for random selection, and provides a baseline for the interpretation of the precision curves.General findings from the precision graphs are: (i) It is only useful to consider the first halves of the SLs, as the measures approximate afterwards. (ii) Precision of log-likelihood,\u00a3 \u00a4, t-test and frequency strongly decreases in the first part of the SLs, whereas precision of \u00a2 \u00a1 remains almost constant (cf.", |
| "num": null, |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "FIGREF1": { |
| "text": "Figure 1) or even increases slightly (cf.", |
| "num": null, |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "FIGREF2": { |
| "text": "Figure 2). (iii) The identification results are instable for the first few percent of the data, with log-likelihood, t-test and frequency sta-Precision graphs for PNV data. stabilizing earlier than the AdjN data. This instability is caused by \"random fluctuations\", i.e., whether a particular TP ends up on rank \u00a6", |
| "num": null, |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "FIGREF3": { |
| "text": "Recall graphs for PNV data.", |
| "num": null, |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "FIGREF4": { |
| "text": "data in the SLs are examined. In the remaining part of the lists, \u00a2 \u00a1 yields better results than frequency and is practically identical to the best-performing measures.", |
| "num": null, |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "FIGREF5": { |
| "text": "Figure 7: AdjN data with \u00a5 \u00a1 \u00a4 \u00a3 .", |
| "num": null, |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "FIGREF7": { |
| "text": "in order to obtain an upper estimate for the actual proportion of collocations among the low-frequency data: the estimate , there should be at most 320 TPs among the AdjN candidates with \u00a1 . Compared to the 737 TPs identified in the AdjN data with \" \u00a1 \u00a5 , our decision to exclude the hapaxlegomena was well justified. The proportion of TPs in the PNV sample (\u00a2", |
| "num": null, |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "FIGREF8": { |
| "text": "extracting a substantial number of collocations from the low-frequency PNV data.Figure 9shows precision curves for the 10 000 highest ranked word combinations from each SL for PNV combinations with", |
| "num": null, |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "FIGREF9": { |
| "text": "Significance of differences (AdjN)", |
| "num": null, |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "TABREF1": { |
| "type_str": "table", |
| "num": null, |
| "html": null, |
| "text": "This is to a large part due to the fact that systematically overestimates the collocativity of low-frequency pairs, cf. section 4.3.whereas the t-test competes with log-likelihood, especially for larger values of \u00a6", |
| "content": "<table><tr><td colspan=\"2\">. Frequency leads \u00a1 and \u00a3 , and, for \u00a4 , comes close to the accuracy of t-test and to clearly better results than \u00a7 \u00a7 \u00a6 log-likelihood.</td></tr><tr><td colspan=\"2\">Adjective-Noun Combinations</td></tr><tr><td>Log-Likelihood</td><td>\u00a6 \u00a1 \u00a7 \u00a7 \u00a6 65.00% 42.80% \u00a7 \u00a7</td></tr><tr><td>t-Test</td><td>57.00% 42.00%</td></tr><tr><td colspan=\"2\">\u00a3 \u00a5 \u00a4 Mutual Information 23.00% 23.00% 36.00% 34.00%</td></tr><tr><td>Frequency</td><td>51.00% 41.20%</td></tr><tr><td>), the precision of is significantly lower than that of log-likelihood, 8 \u00a2 \u00a1</td><td/></tr></table>" |
| }, |
| "TABREF2": { |
| "type_str": "table", |
| "num": null, |
| "html": null, |
| "text": "Precision values for", |
| "content": "<table/>" |
| } |
| } |
| } |
| } |