| { |
| "paper_id": "P02-1030", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T09:31:12.862303Z" |
| }, |
| "title": "Scaling Context Space", |
| "authors": [ |
| { |
| "first": "James", |
| "middle": [ |
| "R" |
| ], |
| "last": "Curran", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "jamesc@cogsci.ed.ac.uk" |
| }, |
| { |
| "first": "Marc", |
| "middle": [], |
| "last": "Moens", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "Context is used in many NLP systems as an indicator of a term's syntactic and semantic function. The accuracy of the system is dependent on the quality and quantity of contextual information available to describe each term. However, the quantity variable is no longer fixed by limited corpus resources. Given fixed training time and computational resources, it makes sense for systems to invest time in extracting high quality contextual information from a fixed corpus. However, with an effectively limitless quantity of text available, extraction rate and representation size need to be considered. We use thesaurus extraction with a range of context extracting tools to demonstrate the interaction between context quantity, time and size on a corpus of 300 million words.", |
| "pdf_parse": { |
| "paper_id": "P02-1030", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "Context is used in many NLP systems as an indicator of a term's syntactic and semantic function. The accuracy of the system is dependent on the quality and quantity of contextual information available to describe each term. However, the quantity variable is no longer fixed by limited corpus resources. Given fixed training time and computational resources, it makes sense for systems to invest time in extracting high quality contextual information from a fixed corpus. However, with an effectively limitless quantity of text available, extraction rate and representation size need to be considered. We use thesaurus extraction with a range of context extracting tools to demonstrate the interaction between context quantity, time and size on a corpus of 300 million words.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Context plays an important role in many natural language tasks. For example, the accuracy of part of speech taggers or word sense disambiguation systems depends on the quality and quantity of contextual information these systems can extract from the training data. When predicting the sense of a word, for instance, the immediately preceding word is likely to be more important than the tenth previous word; similar observations can be made about POS taggers or chunkers. A crucial part of training these systems lies in extracting from the data high-quality contextual information, in the sense of defining contexts that are both accurate and correlated with the information (the POS tags, the word senses, the chunks) the system is trying to extract.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The quality of contextual information is often determined by the size of the training corpus: with less data available, extracting context information for any given phenomenon becomes less reliable. However, corpus size is no longer a limiting factor: whereas up to now people have typically worked with corpora of around one million words, it has become feasible to build much larger document collections; for example, Banko and Brill (2001) report on experiments with a one billion word corpus.", |
| "cite_spans": [ |
| { |
| "start": 420, |
| "end": 442, |
| "text": "Banko and Brill (2001)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "When using a much larger corpus and scaling the context space, there are, however, other trade-offs to take into consideration: the size of the corpus may make it unfeasible to train some systems because of efficiency issues or hardware costs; it may also result in an unmanageable expansion of the extracted context information, reducing the performance of the systems that have to make use of this information.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "This paper reports on experiments that try to establish some of the trade-offs between corpus size, processing time, hardware costs and the performance of the resulting systems. We report on experiments with a large corpus (around 300 million words). We trained a thesaurus extraction system with a range of context-extracting front-ends to demonstrate the interaction between context quality, extraction time and representation size.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Thesauri have traditionally been used in information retrieval tasks to expand words in queries with synonymous terms (e.g. Ruge, (1997) ). More re-cently, semantic resources have also been used in collocation discovery (Pearce, 2001 ), smoothing and model estimation (Brown et al., 1992; Clark and Weir, 2001 ) and text classification (Baker and Mc-Callum, 1998) . Unfortunately, thesauri are very expensive and time-consuming to produce manually, and tend to suffer from problems of bias, inconsistency, and lack of coverage. In addition, thesaurus compilers cannot keep up with constantly evolving language use and cannot afford to build new thesauri for the many subdomains that information extraction and retrieval systems are being developed for. There is a clear need for methods to extract thesauri automatically or tools that assist in the manual creation and updating of these semantic resources.", |
| "cite_spans": [ |
| { |
| "start": 124, |
| "end": 136, |
| "text": "Ruge, (1997)", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 220, |
| "end": 233, |
| "text": "(Pearce, 2001", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 268, |
| "end": 288, |
| "text": "(Brown et al., 1992;", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 289, |
| "end": 309, |
| "text": "Clark and Weir, 2001", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 336, |
| "end": 363, |
| "text": "(Baker and Mc-Callum, 1998)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Automatic Thesaurus Extraction", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Most existing work on thesaurus extraction and word clustering is based on the general observation that related terms will appear in similar contexts. The differences tend to lie in the way \"context\" is defined and in the way similarity is calculated. Most systems extract co-occurrence and syntactic information from the words surrounding the target term, which is then converted into a vector-space representation of the contexts that each target term appears in (Brown et al., 1992; Pereira et al., 1993; Ruge, 1997; Lin, 1998b) . Other systems take the whole document as the context and consider term co-occurrence at the document level (Crouch, 1988; Sanderson and Croft, 1999) . Once these contexts have been defined, these systems then use clustering or nearest neighbour methods to find similar terms.", |
| "cite_spans": [ |
| { |
| "start": 465, |
| "end": 485, |
| "text": "(Brown et al., 1992;", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 486, |
| "end": 507, |
| "text": "Pereira et al., 1993;", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 508, |
| "end": 519, |
| "text": "Ruge, 1997;", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 520, |
| "end": 531, |
| "text": "Lin, 1998b)", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 641, |
| "end": 655, |
| "text": "(Crouch, 1988;", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 656, |
| "end": 682, |
| "text": "Sanderson and Croft, 1999)", |
| "ref_id": "BIBREF18" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Automatic Thesaurus Extraction", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Finally, some systems extract synonyms directly without extracting and comparing contextual representations for each term. Instead, these systems recognise terms within certain linguistic patterns (e.g. X, Y and other Zs) which associate synonyms and hyponyms (Hearst, 1992; Caraballo, 1999) .", |
| "cite_spans": [ |
| { |
| "start": 260, |
| "end": 274, |
| "text": "(Hearst, 1992;", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 275, |
| "end": 291, |
| "text": "Caraballo, 1999)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Automatic Thesaurus Extraction", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Thesaurus extraction is a good task to use to experiment with scaling context spaces. The vectorspace model with nearest neighbour searching is simple, so we needn't worry about interactions between the contexts we select and a learning algorithm (such as independence of the features). But also, thesaurus extraction is a task where success has been limited when using small corpora (Grefenstette, 1994) ; corpora of the order of 300 million words have already been shown to be more successful at this task (Lin, 1998b) .", |
| "cite_spans": [ |
| { |
| "start": 384, |
| "end": 404, |
| "text": "(Grefenstette, 1994)", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 508, |
| "end": 520, |
| "text": "(Lin, 1998b)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Automatic Thesaurus Extraction", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Vector-space thesaurus extraction can be separated into two independent processes. The first step extracts the contexts from raw text and compiles them into a vector-space statistical description of the contexts each potential thesaurus term appears in.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments", |
| "sec_num": "3" |
| }, |
| { |
| "text": "We define a context relation as a tuple (w, r, w ) where w is a thesaurus term, occurring in relation type r, with another word w in the sentence. The type can be grammatical or the position of w in a context window: the relation (dog, direct-obj, walk) indicates that the term dog, was the direct object of the verb walk. Often we treat the tuple (r, w ) as a single unit and refer to it as an attribute of w. The context extraction systems used for these experiments are described in the following section.", |
| "cite_spans": [ |
| { |
| "start": 230, |
| "end": 253, |
| "text": "(dog, direct-obj, walk)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments", |
| "sec_num": "3" |
| }, |
| { |
| "text": "The second step in thesaurus extraction performs clustering or nearest-neighbour analysis to determine which terms are similar based on their context vectors. Our second component is similar to Grefenstette's SEXTANT system, which performs nearestneighbour calculations for each pair of potential thesaurus terms. For nearest-neighbour measurements we must define a function to judge the similarity between two context vectors (e.g. the cosine measure) and a function to combine the raw instance frequencies for each context relation into weighted vector components.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments", |
| "sec_num": "3" |
| }, |
| { |
| "text": "SEXTANT uses a generalisation of the Jaccard measure to measure similarity. The Jaccard measure is the cardinality ratio of the intersection and union of attribute sets (atts(w n ) is the attribute set for w n ):", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments", |
| "sec_num": "3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "| atts(w m ) \u2229 atts(w n )| | atts(w m ) \u222a atts(w n )|", |
| "eq_num": "(1)" |
| } |
| ], |
| "section": "Experiments", |
| "sec_num": "3" |
| }, |
| { |
| "text": "The generalised Jaccard measure allows each relation to have a significance weight (based on word, attribute and relation frequencies) associated with it:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments", |
| "sec_num": "3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "a\u2208atts(w m )\u222aatts(w n ) min(wgt(w m , a), wgt(w n , a)) a\u2208atts(w m )\u222aatts(w n ) max(wgt(w m , a), wgt(w n , a))", |
| "eq_num": "(2)" |
| } |
| ], |
| "section": "Experiments", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Grefenstette originally used the weighting function:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments", |
| "sec_num": "3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "wgt(w i , a j ) = log 2 ( f (w i , a j ) + 1) log 2 (n(a j ) + 1)", |
| "eq_num": "(3)" |
| } |
| ], |
| "section": "Experiments", |
| "sec_num": "3" |
| }, |
| { |
| "text": "where f (w i , a j ) is the frequency of the relation and n(a j ) is the number of different words a j appears in relations with.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Context Description W(L 1 R 1 ) one word to left or right W(L 1 ) one word to the left W(L 1,2 ) one or two words to the left W(L 1\u22123 ) one to three words to the left Table 1 : Window extractors However, we have found that using the t-test between the joint and independent distributions of a word and its attribute:", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 167, |
| "end": 174, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Name", |
| "sec_num": null |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "wgt(w i , a j ) = p(w i , a j ) \u2212 p(w i )p(a j ) p(w i )p(a j )", |
| "eq_num": "(4)" |
| } |
| ], |
| "section": "Name", |
| "sec_num": null |
| }, |
| { |
| "text": "gives superior performance (Curran and Moens, 2002) and is therefore used for our experiments.", |
| "cite_spans": [ |
| { |
| "start": 27, |
| "end": 51, |
| "text": "(Curran and Moens, 2002)", |
| "ref_id": "BIBREF8" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Name", |
| "sec_num": null |
| }, |
| { |
| "text": "We have experimented with a number of different systems for extracting the contexts for each word. These systems show a wide range in complexity of method and implementation, and hence development effort and execution time. The simplest method we implemented extracts the occurrence counts of words within a particular window surrounding the thesaurus term. These window extractors are very easy to implement and run very quickly. The window geometries used in this experiment are listed in Table 1 . Extractors marked with an asterisk, for example W(L 1 R 1 * ), do not distinguish (within the relation type) between different positions of the word w in the window.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 491, |
| "end": 498, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Context Extractors", |
| "sec_num": "4" |
| }, |
| { |
| "text": "At a greater level of complexity we have two shallow NLP systems which provide extra syntactic information in the extracted contexts. The first system is based on the syntactic relation extractor from SEXTANT with a different POS tagger and chunker. The SEXTANT-based extractor we developed uses a very simple Na\u00efve Bayes POS tagger and chunker. This is very simple to implement and is extremely fast since it optimises the tag selection locally at the current word rather than performing beam or Viterbi search over the entire sentence. After the raw text has been POS tagged and chunked, the SEXTANT relation extraction algorithm is run over the text. This consists of five passes over each sentence that associate each noun with the modifiers and verbs from the syntactic contexts that it appears in.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Context Extractors", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Sentences Words British National Corpus 6.2M 114M Reuters Corpus Vol 1 8.7M 193M The second shallow parsing extractor we used was the CASS parser (Abney, 1996) , which uses cascaded finite state transducers to produce a limited depth parse of POS tagged text. We used the output of the Na\u00efve Bayes POS tagger output as input to the CASS. The context relations used were extracted directly by the tuples program (using e8 demo grammar) included in the CASS distribution. The FST parsing algorithm is very efficient and so CASS also ran very quickly. The times reported below include the Na\u00efve Bayes POS tagging time.", |
| "cite_spans": [ |
| { |
| "start": 146, |
| "end": 159, |
| "text": "(Abney, 1996)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Corpus", |
| "sec_num": null |
| }, |
| { |
| "text": "The final, most sophisticated extractor used was the MINIPAR parser (Lin, 1998a) , which is a broadcoverage principle-based parser. The context relations used were extracted directly from the full parse tree. Although fast for a full parser, MINIPAR was no match for the simpler extractors.", |
| "cite_spans": [ |
| { |
| "start": 68, |
| "end": 80, |
| "text": "(Lin, 1998a)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Corpus", |
| "sec_num": null |
| }, |
| { |
| "text": "For this experiment we needed a large quantity of text which we could group into a range of corpus sizes. We combined the BNC and Reuters corpus to produce a 300 million word corpus. The respective sizes of each are shown in Table 2 . The sentences were randomly shuffled together to produce a single homogeneous corpus. This corpus was split into two 150M word corpora over which the main experimental results are averaged. We then created smaller corpora of size 1 2 down to 1 64 th of each 150M corpus. The next section describes the method of evaluating each thesaurus created by the combination of a given context extraction system and corpus size.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 225, |
| "end": 232, |
| "text": "Table 2", |
| "ref_id": "TABREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Corpus", |
| "sec_num": null |
| }, |
| { |
| "text": "For the purposes of evaluation, we selected 70 single word noun terms for thesaurus extraction. To avoid sample bias, the words were randomly selected from Wordnet such that they covered a range of values for the following word properties:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "5" |
| }, |
| { |
| "text": "occurrence frequency based on frequency counts from the Penn Treebank, BNC and Reuters;", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "5" |
| }, |
| { |
| "text": "number of senses based on the number of Wordnet synsets and Macquarie Thesaurus entries;", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "5" |
| }, |
| { |
| "text": "generality/specificity based on depth of the term in the Wordnet hierarchy;", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "5" |
| }, |
| { |
| "text": "abstractness/concreteness based on even distribution across all Wordnet subtrees. Table 3 shows some of the selected terms with frequency and synonym set data. For each term we extracted a thesaurus entry with 200 potential synonyms and their weighted Jaccard scores. The most difficult aspect of thesaurus extraction is evaluating the quality of the result. The simplest method of evaluation is direct comparison of the extracted thesaurus with a manually created gold standard (Grefenstette, 1994) . However on smaller corpora direct matching alone is often too coarsegrained and thesaurus coverage is a problem.", |
| "cite_spans": [ |
| { |
| "start": 479, |
| "end": 499, |
| "text": "(Grefenstette, 1994)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 82, |
| "end": 89, |
| "text": "Table 3", |
| "ref_id": "TABREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Our experiments use a combination of three thesauri available in electronic form: The Macquarie Thesaurus (Bernard, 1990), Roget's Thesaurus (Roget, 1911) , and the Moby Thesaurus (Ward, 1996) . Each thesaurus is structured differently: Roget's and Macquarie are topic ordered and the Moby thesaurus is head term ordered. Roget's is quite dated and has low coverage, and contains a deep hierarchy (depth up to seven) with terms grouped in 8696 small synonym sets at the leaves of the hierarchy. The Macquarie consists of 812 large topics (often in antonym related pairs), each of which is separated into 21174 small synonym sets. Roget's and the Macquarie provide sense distinctions by placing terms in multiple synonym sets. The Moby thesaurus consists of 30259 head terms and large synonym lists which conflate all the head term senses. The extracted thesaurus does not distinguish between different head senses. Therefore, we convert the Roget's and Macquarie thesaurus into head term ordered format by combining each small sense set that the head term appears in.", |
| "cite_spans": [ |
| { |
| "start": 123, |
| "end": 154, |
| "text": "Roget's Thesaurus (Roget, 1911)", |
| "ref_id": null |
| }, |
| { |
| "start": 180, |
| "end": 192, |
| "text": "(Ward, 1996)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "5" |
| }, |
| { |
| "text": "We create a gold standard thesaurus containing the union of the synonym lists from each thesaurus, giving a total of 23207 synonyms for the 70 terms. With these gold standard resources in place, it is possible to use precision and recall measures to calculate the performance of the thesaurus extraction systems. To help overcome the problems of coarsegrained direct comparisons we use three different types of measure to evaluate thesaurus quality:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "5" |
| }, |
| { |
| "text": "1. Direct Match (DIRECT) 2. Precision of the n top ranked synonyms (P(n))", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "5" |
| }, |
| { |
| "text": "A match is an extracted synonym that appears in the corresponding gold standard synonym list. The direct match score is the number of such matches for each term. Precision of the top n is the percentage of matches in the top n extracted synonyms. In these experiments, we calculate this for n = 1, 5, and 10. The inverse rank score is the sum of the inverse rank of each match. For example, if matching synonyms appear in the extracted synonym list at ranks 3, 5 and 28, then the inverse rank score is 1 3 + 1 5 + 1 28 = 0.569. The maximum inverse rank score is 5.878 for a synonym list of 200 terms. Inverse rank is a good measure of subtle differences in ranked results. Each measure is averaged over the extracted synonym lists for all 70 thesaurus terms.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Inverse Rank (INVR)", |
| "sec_num": "3." |
| }, |
| { |
| "text": "Since MINIPAR performs morphological analysis on the context relations we have added an existing morphological analyser (Minnen et al., 2000) to the other extractors. Table 4 shows the improvement gained by morphological analysis of the attributes and relations for the SEXTANT 150M corpus.", |
| "cite_spans": [ |
| { |
| "start": 120, |
| "end": 141, |
| "text": "(Minnen et al., 2000)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 167, |
| "end": 174, |
| "text": "Table 4", |
| "ref_id": "TABREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "6" |
| }, |
| { |
| "text": "The improvement in results is quite significant, as is the reduction in the representation space and number of unique context relations. The reduction in the number of terms is a result of coalescing the plural nouns with their corresponding singular nouns, which also reduces data sparseness problems. The remainder of the results use morphological analysis of both the words and attributes. Table 5 summarises the average results of applying all of the extraction systems to the two 150M word corpora. The first thing to note is the time spent extracting contextual information: MINIPAR takes significantly longer to run than the other extractors. Secondly, SEXTANT and MINI-PAR have quite similar results overall, but MINIPAR is slightly better across most measures. However, SEXTANT runs about 28 times faster than MINI-PAR. Also, MINIPAR extracts many more terms and relations with a much larger representation than SEXTANT. This is partly because MINIPAR extracts more types of relations from the parse tree The larger windows with low correlation between the thesaurus term and context, extract a massive context representation but the results are about 10% worse than the syntactic extractors. Overall the precision and recall are relatively poor. Poor recall is partly due to the gold standard containing some plurals and multi-word terms which account for about 25% of the synonyms. These have been retained because the MINIPAR and CASS systems are capable of identifying (at least some) multi-word terms.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 393, |
| "end": 400, |
| "text": "Table 5", |
| "ref_id": "TABREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "6" |
| }, |
| { |
| "text": "Given a fixed time period (of more than the four days MINIPAR takes) and a fixed 150M corpus we would probably still choose to use MINIPAR unless the representation was too big for our learning algorithm, since the thesaurus quality is slightly better. Table 6 shows what happens to thesaurus quality as we decrease the size of the corpus to 1 64 th of its original size (2.3M words) for SEXTANT. Halving the corpus results in a significant reduction for most of the measures. All five evaluation measures show the same log-linear dependence on the size of the corpus. Figure 1 shows the same trend for Inverse Rank evaluation of the MINIPAR thesaurus with a log-linear fitting the data points.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 253, |
| "end": 260, |
| "text": "Table 6", |
| "ref_id": "TABREF6" |
| }, |
| { |
| "start": 569, |
| "end": 577, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "6" |
| }, |
| { |
| "text": "We can use the same curve fitting to estimate the- saurus quality on larger corpora for three of the best extractors: SEXTANT, MINIPAR and W(L 1 R 1 ). Figure 2 does this with the direct match evaluation. The estimate indicates that MINIPAR will continue to be the best performer on direct matching. We then plot the direct match scores for the 300M word corpus to see how accurate our predictions are. The SEX-TANT system performs almost exactly as predicted and the other two slightly under-perform their predicted scores, thus the fitting is accurate enough to make reasonable predictions. Figure 2 is a graph for making engineering decisions in conjunction with the data in Figure 2 : Direct matches versus corpus size resources at an arbitrary point, e.g. the point where MINIPAR can process 75M words, we get a best direct match score of 23.5. However, we can get the same resultant accuracy by using SEXTANT on a corpus of 116M words or W(L 1 R 1 ) on a corpus of 240M words. From Figure 5 , extracting contexts from corpora of these sizes would take MINIPAR 37 hours, SEXTANT 2 hours and W(L 1 R 1 ) 12 minutes. Interpolation on Figure 3 predicts that the extraction would result in 10M unique relations from MINI-PAR and SEXTANT and 19M from W(L 1 R 1 ). ", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 152, |
| "end": 158, |
| "text": "Figure", |
| "ref_id": null |
| }, |
| { |
| "start": 593, |
| "end": 601, |
| "text": "Figure 2", |
| "ref_id": null |
| }, |
| { |
| "start": 678, |
| "end": 686, |
| "text": "Figure 2", |
| "ref_id": null |
| }, |
| { |
| "start": 988, |
| "end": 996, |
| "text": "Figure 5", |
| "ref_id": null |
| }, |
| { |
| "start": 1137, |
| "end": 1145, |
| "text": "Figure 3", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "6" |
| }, |
| { |
| "text": ") Sextant Minipar Cass W(L 1 R 1 ) W(L 1 R 1 *) W(L 1,2 )", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "6" |
| }, |
| { |
| "text": "Figure 3: Representation size versus corpus size MINIPAR terms, 200k SEXTANT terms and 600k W(L 1 R 1 ) terms.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "6" |
| }, |
| { |
| "text": "Given these values and the fact that the time complexity of most thesaurus extraction algorithms is at least linear in the number of unique relations and squared in the number of thesaurus terms, it seems SEXTANT may represent the best solution.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "6" |
| }, |
| { |
| "text": "With these size issues in mind, we finally consider some methods to limit the size of the context representation. Figure 4 : Thesaurus terms versus corpus size full 300M word corpus, but have size limits based on the 150M word corpus. The FIXED filter does not allow any object/attribute pairs to be added that were not extracted from the 150M word corpus. The LEX-ICON filter does not allow any objects to be added that were not extracted from the 150M word corpus. The > 1 and > 2 filters prune relations with a frequency of less than or equal to one or two. The FIXED and LEXICON filters show that counting over larger corpora does produce marginally better results. The > 1 and > 2 filters show that the many relations that occur infrequently do not contribute significantly to the vector comparisons and hence don't impact on the final results, even though they dramatically increase the representation size.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 114, |
| "end": 122, |
| "text": "Figure 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "6" |
| }, |
| { |
| "text": "It is a phenomenon common to many NLP tasks that the quality or accuracy of a system increases loglinearly with the size of the corpus. Banko and Brill, (2001) also found this trend for the task of confusion set disambiguation on corpora of up to one billion words. They demonstrated behaviour of different learning algorithms with very simple contexts on extremely large corpora. We have demonstrated the behaviour of a simple learning algorithm on much more complicated contextual information on very large corpora.", |
| "cite_spans": [ |
| { |
| "start": 136, |
| "end": 159, |
| "text": "Banko and Brill, (2001)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "7" |
| }, |
| { |
| "text": "Our experiments suggest that the existing methodology of evaluating systems on small corpora without reference to the execution time and representation size ignores important aspects of the evaluation of NLP tools.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "7" |
| }, |
| { |
| "text": "These experiments show that efficiently implementing and optimising the NLP tools used for context extraction is of crucial importance since the increased corpus sizes make execution speed an important evaluation factor when deciding between different learning algorithms for different tasks and corpora. These results also motivate further research into improving the asymptotic complexity of the learning algorithms used in NLP systems. In the new paradigm, it could well be that far simpler but scalable learning algorithms significantly outperform existing systems.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "7" |
| }, |
| { |
| "text": "Finally, the mass availability of online text resources should be taken on board. It is important that language engineers and computational linguists continue to try and find new unsupervised or (as Banko and Brill suggest) semi-supervised methods for tasks which currently rely on annotated data. It is also important to consider how information extracted by systems such as thesaurus extraction systems can be incorporated into tasks which use predominantly supervised techniques, e.g. in the form of class information for smoothing.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "7" |
| }, |
| { |
| "text": "We would like to extend this analysis to at least one billion words for at least the most successful methods and try other tools and parsers for extracting the contextual information. However, to do this we must look at methods of compressing the vectorspace model and approximating the full pair-wise comparison of thesaurus terms. We would also like to investigate how this thesaurus information can be used to improve the accuracy or generality of other NLP tasks.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "7" |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "We would like to thank Miles Osborne, Stephen Clark, Tara Murphy, and the anonymous reviewers for their comments on drafts of this paper. This research is supported by a Commonwealth scholarship and a Sydney University Travelling scholarship.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgements", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Partial parsing via finite-state cascades", |
| "authors": [ |
| { |
| "first": "Steve", |
| "middle": [], |
| "last": "Abney", |
| "suffix": "" |
| } |
| ], |
| "year": 1996, |
| "venue": "Journal of Natural Language Engineering", |
| "volume": "2", |
| "issue": "4", |
| "pages": "337--344", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Steve Abney. 1996. Partial parsing via finite-state cas- cades. Journal of Natural Language Engineering, 2(4):337-344, December.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Distributional clustering of words for text classification", |
| "authors": [ |
| { |
| "first": "L", |
| "middle": [], |
| "last": "", |
| "suffix": "" |
| }, |
| { |
| "first": "Douglas", |
| "middle": [], |
| "last": "Baker", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew", |
| "middle": [], |
| "last": "Mccallum", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "Proceedings of the 21st annual international ACM SI-GIR conference on Research and Development in Information Retrieval", |
| "volume": "", |
| "issue": "", |
| "pages": "24--28", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "L. Douglas Baker and Andrew McCallum. 1998. Distri- butional clustering of words for text classification. In Proceedings of the 21st annual international ACM SI- GIR conference on Research and Development in In- formation Retrieval, pages 96-103, Melbourne, Aus- tralia, 24-28 August.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Scaling to very very large corpora for natural language disambiguation", |
| "authors": [ |
| { |
| "first": "Michele", |
| "middle": [], |
| "last": "Banko", |
| "suffix": "" |
| }, |
| { |
| "first": "Eric", |
| "middle": [], |
| "last": "Brill", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the 39th annual meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "9--11", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Michele Banko and Eric Brill. 2001. Scaling to very very large corpora for natural language disambigua- tion. In Proceedings of the 39th annual meeting of the Association for Computational Linguistics, pages 26- 33, Toulouse, France, 9-11 July.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "The Macquarie Encyclopedic Thesaurus. The Macquarie Library", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [ |
| "L" |
| ], |
| "last": "John", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Bernard", |
| "suffix": "" |
| } |
| ], |
| "year": 1990, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "John R. L. Bernard, editor. 1990. The Macquarie Ency- clopedic Thesaurus. The Macquarie Library, Sydney, Australia.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Classbased n-gram models of natural language", |
| "authors": [ |
| { |
| "first": "F", |
| "middle": [], |
| "last": "Peter", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Brown", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [ |
| "Della" |
| ], |
| "last": "Vincent", |
| "suffix": "" |
| }, |
| { |
| "first": "Peter", |
| "middle": [ |
| "V" |
| ], |
| "last": "Pietra", |
| "suffix": "" |
| }, |
| { |
| "first": "Jennifer", |
| "middle": [ |
| "C" |
| ], |
| "last": "Desouza", |
| "suffix": "" |
| }, |
| { |
| "first": "Robert", |
| "middle": [ |
| "L" |
| ], |
| "last": "Lai", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Mercer", |
| "suffix": "" |
| } |
| ], |
| "year": 1992, |
| "venue": "Computational Linguistics", |
| "volume": "18", |
| "issue": "4", |
| "pages": "467--479", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Peter F. Brown, Vincent J. Della Pietra, Peter V. deSouza, Jennifer C. Lai, and Robert L. Mercer. 1992. Class- based n-gram models of natural language. Computa- tional Linguistics, 18(4):467-479, December.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Automatic construction of a hypernym-labeled noun hierarchy from text", |
| "authors": [ |
| { |
| "first": "Sharon", |
| "middle": [ |
| "A" |
| ], |
| "last": "Caraballo", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "Proceedings of the 37th annual meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "20--26", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sharon A. Caraballo. 1999. Automatic construction of a hypernym-labeled noun hierarchy from text. In Pro- ceedings of the 37th annual meeting of the Association for Computational Linguistics, pages 120-126, Col- lege Park, MD USA, 20-26 June.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Class-based probability estimation using a semantic hierarchy", |
| "authors": [ |
| { |
| "first": "Stephen", |
| "middle": [], |
| "last": "Clark", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Weir", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the Second Meeting of the North American Chapter of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "2--7", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Stephen Clark and David Weir. 2001. Class-based prob- ability estimation using a semantic hierarchy. In Pro- ceedings of the Second Meeting of the North American Chapter of the Association for Computational Linguis- tics, pages 95-102, Pittsburgh, PA USA, 2-7 June.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Construction of a dynamic thesaurus and its use for associated information retrieval", |
| "authors": [ |
| { |
| "first": "Carolyn", |
| "middle": [ |
| "J" |
| ], |
| "last": "Crouch", |
| "suffix": "" |
| } |
| ], |
| "year": 1988, |
| "venue": "Proceedings of the eleventh international conference on Research and Development in Information Retrieval", |
| "volume": "", |
| "issue": "", |
| "pages": "13--15", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Carolyn J. Crouch. 1988. Construction of a dynamic the- saurus and its use for associated information retrieval. In Proceedings of the eleventh international confer- ence on Research and Development in Information Re- trieval, pages 309-320, Grenoble, France, 13-15 June.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Improvements in automatic thesaurus extraction", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "James", |
| "suffix": "" |
| }, |
| { |
| "first": "Marc", |
| "middle": [], |
| "last": "Curran", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Moens", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "ACL-SIGLEX Workshop on Unsupervised Lexical Acquisition", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "James R. Curran and Marc Moens. 2002. Improve- ments in automatic thesaurus extraction. In ACL- SIGLEX Workshop on Unsupervised Lexical Acquisi- tion, Philadelphia, PA USA. (to appear).", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Explorations in Automatic Thesaurus Discovery", |
| "authors": [ |
| { |
| "first": "Gregory", |
| "middle": [], |
| "last": "Grefenstette", |
| "suffix": "" |
| } |
| ], |
| "year": 1994, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Gregory Grefenstette. 1994. Explorations in Automatic Thesaurus Discovery. Kluwer Academic Publishers, Boston, USA.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Automatic acquisition of hyponyms from large text corpora", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Marti", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Hearst", |
| "suffix": "" |
| } |
| ], |
| "year": 1992, |
| "venue": "Proceedings of the 14th international conference on Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "23--28", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Marti A. Hearst. 1992. Automatic acquisition of hy- ponyms from large text corpora. In Proceedings of the 14th international conference on Computational Lin- guistics, pages 539-545, Nantes, France, 23-28 July.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Dependency-based evaluation of MINIPAR", |
| "authors": [ |
| { |
| "first": "Dekang", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "Workshop on the Evaluation of Parsing Systems, Proceedings of the First International Conference on Language Resources and Evaluation", |
| "volume": "", |
| "issue": "", |
| "pages": "28--30", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dekang Lin. 1998a. Dependency-based evaluation of MINIPAR. In Workshop on the Evaluation of Parsing Systems, Proceedings of the First International Con- ference on Language Resources and Evaluation, pages 234-241, Granada, Spain, 28-30 May.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "An information-theoretic definition of similarity", |
| "authors": [ |
| { |
| "first": "Dekang", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "Proceedings of the Fifteen International Conference on Machine Learning", |
| "volume": "", |
| "issue": "", |
| "pages": "24--27", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dekang Lin. 1998b. An information-theoretic definition of similarity. In Proceedings of the Fifteen Interna- tional Conference on Machine Learning, pages 296- 304, Madison, WI USA, 24-27 July.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Robust applied morphological generation", |
| "authors": [ |
| { |
| "first": "Guido", |
| "middle": [], |
| "last": "Minnen", |
| "suffix": "" |
| }, |
| { |
| "first": "John", |
| "middle": [], |
| "last": "Carroll", |
| "suffix": "" |
| }, |
| { |
| "first": "Darren", |
| "middle": [], |
| "last": "Pearce", |
| "suffix": "" |
| } |
| ], |
| "year": 2000, |
| "venue": "Proceedings of the First International Natural Language Generation Conference", |
| "volume": "", |
| "issue": "", |
| "pages": "12--16", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Guido Minnen, John Carroll, and Darren Pearce. 2000. Robust applied morphological generation. In In Pro- ceedings of the First International Natural Language Generation Conference, pages 201-208, Mitzpe Ra- mon, Israel, 12-16 June.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Synonymy in collocation extraction", |
| "authors": [ |
| { |
| "first": "Darren", |
| "middle": [ |
| "Pearce" |
| ], |
| "last": "", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Workshop on WordNet and Other Lexical Resources: Applications, Extensions and Customizations", |
| "volume": "", |
| "issue": "", |
| "pages": "2--7", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Darren Pearce. 2001. Synonymy in collocation extrac- tion. In Workshop on WordNet and Other Lexical Resources: Applications, Extensions and Customiza- tions, (NAACL 2001), pages 41-46, Pittsburgh, PA USA, 2-7 June.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Distributional clustering of English words", |
| "authors": [ |
| { |
| "first": "Fernando", |
| "middle": [], |
| "last": "Pereira", |
| "suffix": "" |
| }, |
| { |
| "first": "Naftali", |
| "middle": [], |
| "last": "Tishby", |
| "suffix": "" |
| }, |
| { |
| "first": "Lillian", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| } |
| ], |
| "year": 1993, |
| "venue": "Proceedings of the 31st annual meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "22--26", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Fernando Pereira, Naftali Tishby, and Lillian Lee. 1993. Distributional clustering of English words. In Pro- ceedings of the 31st annual meeting of the Associ- ation for Computational Linguistics, pages 183-190, Columbus, Ohio USA, 22-26 June.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Thesaurus of English words and phrases. Longmans, Green and Co", |
| "authors": [ |
| { |
| "first": "Peter", |
| "middle": [], |
| "last": "Roget", |
| "suffix": "" |
| } |
| ], |
| "year": 1911, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Peter Roget. 1911. Thesaurus of English words and phrases. Longmans, Green and Co., London, UK.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Automatic detection of thesaurus relations for information retrieval applications", |
| "authors": [ |
| { |
| "first": "Gerda", |
| "middle": [], |
| "last": "Ruge", |
| "suffix": "" |
| } |
| ], |
| "year": 1997, |
| "venue": "Foundations of Computer Science: Potential -Theory -Cognition", |
| "volume": "1337", |
| "issue": "", |
| "pages": "499--506", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Gerda Ruge. 1997. Automatic detection of thesaurus re- lations for information retrieval applications. In Foun- dations of Computer Science: Potential -Theory - Cognition, Lecture Notes in Computer Science, vol- ume LNCS 1337, pages 499-506. Springer Verlag, Berlin, Germany.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Deriving concept hierarchies from text", |
| "authors": [ |
| { |
| "first": "Mark", |
| "middle": [], |
| "last": "Sanderson", |
| "suffix": "" |
| }, |
| { |
| "first": "Bruce", |
| "middle": [], |
| "last": "Croft", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "Proceedings of the 22nd annual international ACM SIGIR conference on Research and Development in Information Retrieval", |
| "volume": "", |
| "issue": "", |
| "pages": "15--19", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Mark Sanderson and Bruce Croft. 1999. Deriving con- cept hierarchies from text. In Proceedings of the 22nd annual international ACM SIGIR conference on Research and Development in Information Retrieval, pages 206-213, Berkeley, CA USA, 15-19 August.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "uris": null, |
| "type_str": "figure", |
| "num": null, |
| "text": "MINIPAR INVR scores versus corpus size" |
| }, |
| "FIGREF1": { |
| "uris": null, |
| "type_str": "figure", |
| "num": null, |
| "text": "indicates that extraction would result in 550k" |
| }, |
| "TABREF0": { |
| "text": "Training Corpora Statistics", |
| "type_str": "table", |
| "num": null, |
| "content": "<table/>", |
| "html": null |
| }, |
| "TABREF1": { |
| "text": "WordPTB Rank PTB # BNC # Reuters # Macquarie # WordNet # Min / Max WordNet subtree roots", |
| "type_str": "table", |
| "num": null, |
| "content": "<table><tr><td>company interest problem change house idea opinion radio star knowledge pants tightness</td><td>38 4076 52779 456580 138 919 37454 146043 418 622 56361 63333 681 406 35641 55081 896 223 47801 45651 1227 134 32754 13527 1947 78 9122 16320 2278 59 9046 20913 5130 29 8301 6586 5197 19 14580 2813 13264 5 429 282 30817 1 119 2020</td><td>8 12 4 8 10 10 4 2 11 3 3 5</td><td>9 12 3 10 12 5 6 3 7 1 2 3</td><td>3 / 6 3 / 8 3 / 7 2 / 12 3 / 6 3 / 7 4 / 8 6 / 8 4 / 8 1 / 1 6 / 9 4 / 5</td><td>entity, group, state abs., act, group, poss., state abs., psych., state abs., act, entity, event, phenom. act, entity, group entity, psych. abs., act, psych. entity abs., entity psych. entity abs., state</td></tr></table>", |
| "html": null |
| }, |
| "TABREF2": { |
| "text": "Examples of the 70 thesaurus evaluation terms with distribution information", |
| "type_str": "table", |
| "num": null, |
| "content": "<table><tr><td>Morph. Analysis</td><td>Space</td><td colspan=\"3\">Unique Terms DIRECT</td><td>P(1)</td><td>P(5)</td><td>P(10)</td><td>INVR</td></tr><tr><td>None Attributes Both</td><td colspan=\"2\">345Mb 14.70M 302Mb 13.17M 274Mb 12.08M</td><td>298k 298k 269k</td><td>20.33 20.65 23.74</td><td colspan=\"3\">32.5 % 36.9 % 33.6 % 32.0 % 37.6 % 32.5 % 64.5 % 47.0 % 39.0 %</td><td>1.37 1.36 1.86</td></tr></table>", |
| "html": null |
| }, |
| "TABREF3": { |
| "text": "", |
| "type_str": "table", |
| "num": null, |
| "content": "<table/>", |
| "html": null |
| }, |
| "TABREF4": { |
| "text": "For instance, if we fix the total time and computational", |
| "type_str": "table", |
| "num": null, |
| "content": "<table><tr><td>System</td><td colspan=\"3\">Space Relations Unique Terms DIRECT</td><td>P(1)</td><td>P(5)</td><td>P(10) INVR</td><td>Time</td></tr><tr><td colspan=\"2\">MINIPAR 399Mb 142.27M 16.62M SEXTANT 274Mb 53.07M 12.08M CASS 186Mb 50.63M 9.09M W(L 1 ) 117Mb 105.62M 7.04M W(L 1,2 ) 336Mb 206.02M 18.04M W(L 1,2 * ) 258Mb 206.02M 15.34M W(L 1\u22123 ) 570Mb 301.10M 30.62M W(L 1\u22123 * ) 388Mb 301.10M 22.86M W(L 1 R 1 ) 262Mb 211.24M 14.07M W(L 1 R 1 * ) 211Mb 211.24M 12.56M</td><td>914k 269k 204k 406k 440k 440k 444k 444k 435k 435k</td><td>24.55 23.75 20.20 20.60 21.30 20.75 20.50 19.85 22.40 20.90</td><td colspan=\"3\">61.5 % 46.5 % 40.5 % 64.5 % 47.0 % 39.0 % 48.5 % 38.5 % 32.5 % 51.5 % 40.0 % 32.5 % 58.5 % 44.5 % 36.5 % 55.0 % 41.5 % 35.5 % 60.0 % 43.5 % 37.0 % 48.5 % 39.5 % 33.5 % 62.0 % 44.5 % 37.0 % 54.5 % 42.5 % 34.5 %</td><td>1.85 1.85 1.51 1.56 1.71 1.64 1.69 1.53 1.76 1.64</td><td>4438.9m 159.0m 173.7m 6.8m 7.2m 6.8m 8.2m 8.2m 7.2m 7.2m</td></tr></table>", |
| "html": null |
| }, |
| "TABREF5": { |
| "text": "Average thesaurus quality results for different extraction systems", |
| "type_str": "table", |
| "num": null, |
| "content": "<table><tr><td>Corpus</td><td colspan=\"2\">Space Relations</td><td>Unique</td><td colspan=\"2\">Terms DIRECT</td><td>P(1)</td><td>P(5)</td><td>P(10)</td><td>INVR</td></tr><tr><td colspan=\"2\">150.0M 274Mb 75.0M 166Mb 37.5M 98Mb 18.8M 56Mb 9.4M 32Mb 4.7M 18Mb 2.3M 10Mb</td><td colspan=\"3\">53.07M 12.08M 268.94k 26.54M 7.38M 181.73k 13.27M 4.36M 120.48k 6.63M 2.54M 82.33k 3.32M 1.44M 55.55k 1.66M 0.82M 37.95k 0.83M 0.46M 25.97k</td><td>23.75 22.60 21.75 20.45 18.50 16.65 14.60</td><td colspan=\"3\">64.5 % 47.0 % 39.0 % 58.0 % 43.5 % 36.0 % 54.0 % 41.0 % 34.5 % 47.0 % 36.5 % 31.0 % 40.0 % 32.5 % 27.5 % 34.0 % 29.5 % 23.5 % 27.5 % 25.0 % 19.5 %</td><td>1.85 1.73 1.62 1.46 1.29 1.13 0.93</td></tr></table>", |
| "html": null |
| }, |
| "TABREF6": { |
| "text": "Average SEXTANT thesaurus quality results for different corpus sizes", |
| "type_str": "table", |
| "num": null, |
| "content": "<table><tr><td/><td>32</td><td/><td/><td/><td/><td/><td/><td/></tr><tr><td/><td>28</td><td/><td/><td/><td/><td/><td/><td/></tr><tr><td/><td>24</td><td/><td/><td/><td/><td/><td/><td/></tr><tr><td>Number of Direct Matches</td><td>12 16 20</td><td/><td/><td/><td/><td/><td/><td/></tr><tr><td/><td>8</td><td/><td/><td/><td/><td/><td/><td/></tr><tr><td/><td/><td/><td/><td/><td/><td/><td/><td>Sextant</td></tr><tr><td/><td>4</td><td/><td/><td/><td/><td/><td/><td>Minipar W(L 1 R 1 )</td></tr><tr><td/><td>0</td><td>0</td><td>50</td><td>100</td><td>150</td><td>200</td><td>250</td><td>300</td><td>350</td></tr><tr><td/><td/><td/><td/><td/><td colspan=\"2\">Number of words (millions)</td><td/><td/></tr></table>", |
| "html": null |
| }, |
| "TABREF7": { |
| "text": "shows the results of performing various kinds of filtering on the representation size. The FIXED and LEXICON filters run over the", |
| "type_str": "table", |
| "num": null, |
| "content": "<table><tr><td>System</td><td colspan=\"2\">Space Relations</td><td colspan=\"3\">Unique Terms DIRECT</td><td>P(1)</td><td>P(5)</td><td>P(10)</td><td>INVR</td></tr><tr><td colspan=\"2\">SEXTANT 300M SEXTANT 150M SEXTANT FIXED SEXTANT LEXICON 410Mb 431Mb 274Mb 244Mb SEXTANT > 1 149Mb SEXTANT > 2 88Mb</td><td colspan=\"2\">80.33M 20.41M 53.07M 12.08M 61.17M 10.74M 78.69M 18.09M 67.97M 6.63M 62.57M 3.93M</td><td>445k 269k 265k 264k 171k 109k</td><td colspan=\"3\">25.30 61.0 % 47.0 % 39.0 % 23.75 64.5 % 47.0 % 39.0 % 24.35 65.0 % 46.5 % 38.5 % 25.25 62.0 % 47.0 % 40.0 % 24.20 66.0 % 45.0 % 38.0 % 23.20 66.0 % 46.0 % 36.0 %</td><td>1.87 1.85 1.86 1.87 1.85 1.82</td></tr></table>", |
| "html": null |
| }, |
| "TABREF8": { |
| "text": "Thesaurus quality with relation filtering", |
| "type_str": "table", |
| "num": null, |
| "content": "<table><tr><td/><td>1800</td><td/><td/><td/><td/><td/><td/><td/></tr><tr><td/><td/><td/><td>Sextant</td><td/><td/><td/><td/><td/></tr><tr><td/><td>1600</td><td/><td>Minipar</td><td/><td/><td/><td/><td/></tr><tr><td/><td/><td/><td>Cass</td><td/><td/><td/><td/><td/></tr><tr><td>Number of thesaurus terms (millions)</td><td>400 600 800 1000 1200 1400</td><td/><td>W(L 1 R 1 )</td><td/><td/><td/><td/><td/></tr><tr><td/><td>200</td><td/><td/><td/><td/><td/><td/><td/></tr><tr><td/><td>0</td><td>0</td><td>50</td><td>100</td><td>150</td><td>200</td><td>250</td><td>300</td><td>350</td></tr><tr><td/><td/><td/><td/><td/><td colspan=\"2\">Number of words (millions)</td><td/><td/></tr></table>", |
| "html": null |
| } |
| } |
| } |
| } |