| { |
| "paper_id": "P01-1008", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T09:29:33.577923Z" |
| }, |
| "title": "Extracting Paraphrases from a Parallel Corpus", |
| "authors": [ |
| { |
| "first": "Regina", |
| "middle": [], |
| "last": "Barzilay", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Columbia University", |
| "location": { |
| "addrLine": "10027", |
| "settlement": "New York", |
| "region": "NY", |
| "country": "USA" |
| } |
| }, |
| "email": "regina@cs.columbia.edu" |
| }, |
| { |
| "first": "Kathleen", |
| "middle": [ |
| "R" |
| ], |
| "last": "Mckeown", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Columbia University", |
| "location": { |
| "addrLine": "10027", |
| "settlement": "New York", |
| "region": "NY", |
| "country": "USA" |
| } |
| }, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "While paraphrasing is critical both for interpretation and generation of natural language, current systems use manual or semi-automatic methods to collect paraphrases. We present an unsupervised learning algorithm for identification of paraphrases from a corpus of multiple English translations of the same source text. Our approach yields phrasal and single word lexical paraphrases as well as syntactic paraphrases.", |
| "pdf_parse": { |
| "paper_id": "P01-1008", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "While paraphrasing is critical both for interpretation and generation of natural language, current systems use manual or semi-automatic methods to collect paraphrases. We present an unsupervised learning algorithm for identification of paraphrases from a corpus of multiple English translations of the same source text. Our approach yields phrasal and single word lexical paraphrases as well as syntactic paraphrases.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Paraphrases are alternative ways to convey the same information. A method for the automatic acquisition of paraphrases has both practical and linguistic interest. From a practical point of view, diversity in expression presents a major challenge for many NLP applications. In multidocument summarization, identification of paraphrasing is required to find repetitive information in the input documents. In generation, paraphrasing is employed to create more varied and fluent text. Most current applications use manually collected paraphrases tailored to a specific application, or utilize existing lexical resources such as Word-Net (Miller et al., 1990) to identify paraphrases. However, the process of manually collecting paraphrases is time consuming, and moreover, the collection is not reusable in other applications. Existing resources only include lexical paraphrases; they do not include phrasal or syntactically based paraphrases.", |
| "cite_spans": [ |
| { |
| "start": 634, |
| "end": 655, |
| "text": "(Miller et al., 1990)", |
| "ref_id": "BIBREF19" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "From a linguistic point of view, questions concern the operative definition of paraphrases:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "what types of lexical relations and syntactic mechanisms can produce paraphrases? Many linguists (Halliday, 1985; de Beaugrande and Dressler, 1981) agree that paraphrases retain \"approximate conceptual equivalence\", and are not limited only to synonymy relations. But the extent of interchangeability between phrases which form paraphrases is an open question (Dras, 1999) . A corpus-based approach can provide insights on this question by revealing paraphrases that people use. This paper presents a corpus-based method for automatic extraction of paraphrases. We use a large collection of multiple parallel English translations of novels 1 . This corpus provides many instances of paraphrasing, because translations preserve the meaning of the original source, but may use different words to convey the meaning. An example of parallel translations is shown in Figure 1 . It contains two pairs of paraphrases: (\"burst into tears\", \"cried\") and (\"comfort\", \"console\").", |
| "cite_spans": [ |
| { |
| "start": 97, |
| "end": 113, |
| "text": "(Halliday, 1985;", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 114, |
| "end": 147, |
| "text": "de Beaugrande and Dressler, 1981)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 360, |
| "end": 372, |
| "text": "(Dras, 1999)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 862, |
| "end": 870, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Emma burst into tears and he tried to comfort her, saying things to make her smile. Emma cried, and he tried to console her, adorning his words with puns. Our method for paraphrase extraction builds upon methodology developed in Machine Translation (MT). In MT, pairs of translated sentences from a bilingual corpus are aligned, and occurrence patterns of words in two languages in the text are extracted and matched using correlation measures. However, our parallel corpus is far from the clean parallel corpora used in MT. The rendition of a literary text into another language not only includes the translation, but also restructuring of the translation to fit the appropriate literary style. This process introduces differences in the translations which are an intrinsic part of the creative process. This results in greater differences across translations than the differences in typical MT parallel corpora, such as the Canadian Hansards. We will return to this point later in Section 3.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Based on the specifics of our corpus, we developed an unsupervised learning algorithm for paraphrase extraction. During the preprocessing stage, the corresponding sentences are aligned. We base our method for paraphrasing extraction on the assumption that phrases in aligned sentences which appear in similar contexts are paraphrases. To automatically infer which contexts are good predictors of paraphrases, contexts surrounding identical words in aligned sentences are extracted and filtered according to their predictive power. Then, these contexts are used to extract new paraphrases. In addition to learning lexical paraphrases, the method also learns syntactic paraphrases, by generalizing syntactic patterns of the extracted paraphrases. Extracted paraphrases are then applied to the corpus, and used to learn new context rules. This iterative algorithm continues until no new paraphrases are discovered.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "A novel feature of our approach is the ability to extract multiple kinds of paraphrases: Identification of lexical paraphrases. In contrast to earlier work on similarity, our approach allows identification of multi-word paraphrases, in addition to single words, a challenging issue for corpus-based techniques. Extraction of morpho-syntactic paraphrasing rules. Our approach yields a set of paraphrasing patterns by extrapolating the syntactic and morphological structure of extracted paraphrases. This process relies on morphological information and a part-of-speech tagging. Many of the rules identified by the algorithm match those that have been described as productive paraphrases in the linguistic literature.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In the following sections, we provide an overview of existing work on paraphrasing, then we describe data used in this work, and detail our paraphrase extraction technique. We present re-sults of our evaluation, and conclude with a discussion of our results.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Many NLP applications are required to deal with the unlimited variety of human language in expressing the same information. So far, three major approaches of collecting paraphrases have emerged: manual collection, utilization of existing lexical resources and corpus-based extraction of similar words.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work on Paraphrasing", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Manual collection of paraphrases is usually used in generation (Iordanskaja et al., 1991; Robin, 1994) . Paraphrasing is an inevitable part of any generation task, because a semantic concept can be realized in many different ways. Knowledge of possible concept verbalizations can help to generate a text which best fits existing syntactic and pragmatic constraints. Traditionally, alternative verbalizations are derived from a manual corpus analysis, and are, therefore, application specific.", |
| "cite_spans": [ |
| { |
| "start": 63, |
| "end": 89, |
| "text": "(Iordanskaja et al., 1991;", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 90, |
| "end": 102, |
| "text": "Robin, 1994)", |
| "ref_id": "BIBREF22" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work on Paraphrasing", |
| "sec_num": "2" |
| }, |
| { |
| "text": "The second approach -utilization of existing lexical resources, such as WordNet -overcomes the scalability problem associated with an application specific collection of paraphrases. Lexical resources are used in statistical generation, summarization and question-answering. The question here is what type of WordNet relations can be considered as paraphrases. In some applications, only synonyms are considered as paraphrases (Langkilde and Knight, 1998) ; in others, looser definitions are used (Barzilay and Elhadad, 1997) . These definitions are valid in the context of particular applications; however, in general, the correspondence between paraphrasing and types of lexical relations is not clear. The same question arises with automatically constructed thesauri (Pereira et al., 1993; Lin, 1998) . While the extracted pairs are indeed similar, they are not paraphrases. For example, while \"dog\" and \"cat\" are recognized as the most similar concepts by the method described in (Lin, 1998) , it is hard to imagine a context in which these words would be interchangeable.", |
| "cite_spans": [ |
| { |
| "start": 426, |
| "end": 454, |
| "text": "(Langkilde and Knight, 1998)", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 496, |
| "end": 524, |
| "text": "(Barzilay and Elhadad, 1997)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 769, |
| "end": 791, |
| "text": "(Pereira et al., 1993;", |
| "ref_id": "BIBREF20" |
| }, |
| { |
| "start": 792, |
| "end": 802, |
| "text": "Lin, 1998)", |
| "ref_id": "BIBREF16" |
| }, |
| { |
| "start": 983, |
| "end": 994, |
| "text": "(Lin, 1998)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work on Paraphrasing", |
| "sec_num": "2" |
| }, |
| { |
| "text": "The first attempt to derive paraphrasing rules from corpora was undertaken by (Jacquemin et al., 1997) , who investigated morphological and syntactic variants of technical terms. While these rules achieve high accuracy in identifying term paraphrases, the techniques used have not been extended to other types of paraphrasing yet. Statistical techniques were also successfully used by (Lapata, 2001) to identify paraphrases of adjective-noun phrases. In contrast, our method is not limited to a particular paraphrase type.", |
| "cite_spans": [ |
| { |
| "start": 78, |
| "end": 102, |
| "text": "(Jacquemin et al., 1997)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 385, |
| "end": 399, |
| "text": "(Lapata, 2001)", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work on Paraphrasing", |
| "sec_num": "2" |
| }, |
| { |
| "text": "The corpus we use for identification of paraphrases is a collection of multiple English translations from a foreign source text. Specifically, we use literary texts written by foreign authors. Many classical texts have been translated more than once, and these translations are available on-line. In our experiments we used 5 books, among them, Flaubert's Madame Bovary, Andersen's Fairy Tales and Verne's Twenty Thousand Leagues Under the Sea. Some of the translations were created during different time periods and in different countries. In total, our corpus contains 11 translations 2 .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Data", |
| "sec_num": "3" |
| }, |
| { |
| "text": "At first glance, our corpus seems quite similar to parallel corpora used by researchers in MT, such as the Canadian Hansards. The major distinction lies in the degree of proximity between the translations. Analyzing multiple translations of the literary texts, critics (e.g. (Wechsler, 1998) ) have observed that translations \"are never identical\", and each translator creates his own interpretations of the text. Clauses such as \"adorning his words with puns\" and \"saying things to make her smile\" from the sentences in Figure 1 are examples of distinct translations. Therefore, a complete match between words of related sentences is impossible. This characteristic of our corpus is similar to problems with noisy and comparable corpora (Veronis, 2000) , and it prevents us from using methods developed in the MT community based on clean parallel corpora, such as (Brown et al., 1993) .", |
| "cite_spans": [ |
| { |
| "start": 275, |
| "end": 291, |
| "text": "(Wechsler, 1998)", |
| "ref_id": "BIBREF25" |
| }, |
| { |
| "start": 738, |
| "end": 753, |
| "text": "(Veronis, 2000)", |
| "ref_id": null |
| }, |
| { |
| "start": 865, |
| "end": 885, |
| "text": "(Brown et al., 1993)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 521, |
| "end": 529, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "The Data", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Another distinction between our corpus and parallel MT corpora is the irregularity of word matchings: in MT, no words in the source language are kept as is in the target language translation; for example, an English translation of 2 Free of copyright restrictions part of our corpus(9 translations) is available at http://www.cs.columbia.edu/\u02dcregina /par. a French source does not contain untranslated French fragments. In contrast, in our corpus the same word is usually used in both translations, and only sometimes its paraphrases are used, which means that word-paraphrase pairs will have lower co-occurrence rates than wordtranslation pairs in MT. For example, consider occurrences of the word \"boy\" in two translations of \"Madame Bovary\" -E. Marx-Aveling's translation and Etext's translation. The first text contains 55 occurrences of \"boy\", which correspond to 38 occurrences of \"boy\" and 17 occurrences of its paraphrases (\"son\", \"young fellow\" and \"youngster\"). This rules out using word translation methods based only on word co-occurrence counts.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Data", |
| "sec_num": "3" |
| }, |
| { |
| "text": "On the other hand, the big advantage of our corpus comes from the fact that parallel translations share many words, which helps the matching process. We describe below a method of paraphrase extraction, exploiting these features of our corpus.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Data", |
| "sec_num": "3" |
| }, |
| { |
| "text": "During the preprocessing stage, we perform sentence alignment. Sentences which are translations of the same source sentence contain a number of identical words, which serve as a strong clue to the matching process. Alignment is performed using dynamic programming (Gale and Church, 1991) with a weight function based on the number of common words in a sentence pair. This simple method achieves good results for our corpus, because 42% of the words in corresponding sentences are identical words on average. Alignment produces 44,562 pairs of sentences with 1,798,526 words. To evaluate the accuracy of the alignment process, we analyzed 127 sentence pairs from the algorithm's output. 120(94.5%) alignments were identified as correct alignments.", |
| "cite_spans": [ |
| { |
| "start": 264, |
| "end": 287, |
| "text": "(Gale and Church, 1991)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Preprocessing", |
| "sec_num": "4" |
| }, |
| { |
| "text": "We then use a part-of-speech tagger and chunker (Mikheev, 1997) to identify noun and verb phrases in the sentences. These phrases become the atomic units of the algorithm. We also record for each token its derivational root, using the CELEX (Baayen et al., 1993) database.", |
| "cite_spans": [ |
| { |
| "start": 48, |
| "end": 63, |
| "text": "(Mikheev, 1997)", |
| "ref_id": null |
| }, |
| { |
| "start": 241, |
| "end": 262, |
| "text": "(Baayen et al., 1993)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Preprocessing", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Given the aforementioned differences between translations, our method builds on similarity in the local context, rather than on global alignment. Consider the two sentences in Figure 2 .", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 176, |
| "end": 184, |
| "text": "Figure 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Method for Paraphrase Extraction", |
| "sec_num": "5" |
| }, |
| { |
| "text": "And finally, dazzlingly white, it shone high above them in the empty ? .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Method for Paraphrase Extraction", |
| "sec_num": "5" |
| }, |
| { |
| "text": "It appeared white and dazzling in the empty ? .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Method for Paraphrase Extraction", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Analyzing the contexts surrounding \" ? \"marked blanks in both sentences, one expects that they should have the same meaning, because they have the same premodifier \"empty\" and relate to the same preposition \"in\" (in fact, the first \" ? \" stands for \"sky\", and the second for \"heavens\"). Generalizing from this example, we hypothesize that if the contexts surrounding two phrases look similar enough, then these two phrases are likely to be paraphrases. The definition of the context depends on how similar the translations are. Once we know which contexts are good paraphrase predictors, we can extract paraphrase patterns from our corpus.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 2: Fragments of aligned sentences", |
| "sec_num": null |
| }, |
| { |
| "text": "Examples of such contexts are verb-object relations and noun-modifier relations, which were traditionally used in word similarity tasks from non-parallel corpora (Pereira et al., 1993; Hatzivassiloglou and McKeown, 1993) . However, in our case, more indirect relations can also be clues for paraphrasing, because we know a priori that input sentences convey the same information. For example, in sentences from Figure 3, the verbs \"ringing\" and \"sounding\" do not share identical subject nouns, but the modifier of both subjects \"Evening\" is identical. Can we conclude that identical modifiers of the subject imply verb similarity? To address this question, we need a way to identify contexts that are good predictors for paraphrasing in a corpus.", |
| "cite_spans": [ |
| { |
| "start": 162, |
| "end": 184, |
| "text": "(Pereira et al., 1993;", |
| "ref_id": "BIBREF20" |
| }, |
| { |
| "start": 185, |
| "end": 220, |
| "text": "Hatzivassiloglou and McKeown, 1993)", |
| "ref_id": "BIBREF8" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 2: Fragments of aligned sentences", |
| "sec_num": null |
| }, |
| { |
| "text": "People said \"The Evening Noise is sounding, the sun is setting.\" \"The evening bell is ringing,\" people used to say.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 2: Fragments of aligned sentences", |
| "sec_num": null |
| }, |
| { |
| "text": "To find \"good\" contexts, we can analyze all contexts surrounding identical words in the pairs of aligned sentences, and use these contexts to learn new paraphrases. This provides a basis for a bootstrapping mechanism. Starting with identical words in aligned sentences as a seed, we can incrementally learn the \"good\" contexts, and in turn use them to learn new paraphrases. Identical words play two roles in this process: first, they are used to learn context rules; second, identical words are used in application of these rules, because the rules contain information about the equality of words in context. This method of co-training has been previously applied to a variety of natural language tasks, such as word sense disambiguation (Yarowsky, 1995) , lexicon construction for information extraction (Riloff and Jones, 1999) , and named entity classification (Collins and Singer, 1999) . In our case, the co-training process creates a binary classifier, which predicts whether a given pair of phrases makes a paraphrase or not.", |
| "cite_spans": [ |
| { |
| "start": 739, |
| "end": 755, |
| "text": "(Yarowsky, 1995)", |
| "ref_id": "BIBREF26" |
| }, |
| { |
| "start": 806, |
| "end": 830, |
| "text": "(Riloff and Jones, 1999)", |
| "ref_id": "BIBREF21" |
| }, |
| { |
| "start": 865, |
| "end": 891, |
| "text": "(Collins and Singer, 1999)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 3: Fragments of aligned sentences", |
| "sec_num": null |
| }, |
| { |
| "text": "Our model is based on the DLCoTrain algorithm proposed by (Collins and Singer, 1999) , which applies a co-training procedure to decision list classifiers for two independent sets of features. In our case, one set of features describes the paraphrase pair itself, and another set of features corresponds to contexts in which paraphrases occur. These features and their computation are described below.", |
| "cite_spans": [ |
| { |
| "start": 58, |
| "end": 84, |
| "text": "(Collins and Singer, 1999)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Figure 3: Fragments of aligned sentences", |
| "sec_num": null |
| }, |
| { |
| "text": "Our paraphrase features include lexical and syntactic descriptions of the paraphrase pair. The lexical feature set consists of the sequence of tokens for each phrase in the paraphrase pair; the syntactic feature set consists of a sequence of part-of-speech tags where equal words and words with the same root are marked. For example, the value of the syntactic feature for the pair (\"the vast chimney\", \"the chimney\") is (\"DT\u00a2 JJ NN\u00a3 \", \"DT\u00a2 NN\u00a3 \"), where indices indicate word equalities. We believe that this feature can be useful for two reasons: first, we expect that some syntactic categories can not be paraphrased in another syntactic category. For example, a determiner is unlikely to be a paraphrase of a verb. Second, this description is able to capture regularities in phrase level paraphrasing. In fact, a similar representation was used by (Jacquemin et al., 1997) to describe term variations.", |
| "cite_spans": [ |
| { |
| "start": 853, |
| "end": 877, |
| "text": "(Jacquemin et al., 1997)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Feature Extraction", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "The contextual feature is a combination of the left and right syntactic contexts surrounding actual known paraphrases. There are a num-ber of context representations that can be considered as possible candidates: lexical n-grams, POS-ngrams and parse tree fragments. The natural choice is a parse tree; however, existing parsers perform poorly in our domain 3 . Partof-speech tags provide the required level of abstraction, and can be accurately computed for our data. The left (right) context is a sequence of part-of-speech tags of \u00a4 words, occurring on the left (right) of the paraphrase. As in the case of syntactic paraphrase features, tags of identical words are marked. For example, when \u00a4 \u00a6 \u00a5 \u00a7 , the contextual feature for the paraphrase pair (\"comfort\", \"console\") from Figure 1 sentences is left\u00a2 =\"VB\u00a2 TO\u00a3 \", (\"tried to\"), left\u00a3 =\"VB\u00a2 TO\u00a3 \", (\"tried to\"), right\u00a2 =\"PRP$\u00a8,\u00a9 \", (\"her,\") right context$\u00a3 =\"PRP$\u00a8,\u00a9 \", (\"her,\"). In the next section, we describe how the classifiers for contextual and paraphrasing features are co-trained.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 780, |
| "end": 788, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Feature Extraction", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "Our co-training algorithm has three stages: initialization, training of the contextual classifier and training of the paraphrasing classifiers.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The co-training algorithm", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "Initialization Words which appear in both sentences of an aligned pair are used to create the initial \"seed\" rules. Using identical words, we create a set of positive paraphrasing examples, such as word\u00a2 =tried, word\u00a3 =tried. However, training of the classifier demands negative examples as well; in our case it requires pairs of words in aligned sentences which are not paraphrases of each other. To find negative examples, we match identical words in the alignment against all different words in the aligned sentence, assuming that identical words can match only each other, and not any other word in the aligned sentences. For example, \"tried\" from the first sentence in Figure 1 does not correspond to any other word in the second sentence but \"tried\". Based on this observation, we can derive negative examples such as word\u00a2 =tried, word\u00a3 =Emma and word\u00a2 =tried, word\u00a3 =console. Given a pair of identical words from two sentences of length \u00a4 and , the algorithm produces one positive ex- 3 To the best of our knowledge all existing statistical parsers are trained on WSJ or similar type of corpora. In the experiments we conducted, their performance significantly degraded on our corpus -literary texts. ample and \u00a4 ! \" # ! n egative examples. Training of the contextual classifier Using this initial seed, we record contexts around positive and negative paraphrasing examples. From all the extracted contexts we must identify the ones which are strong predictors of their category. Following (Collins and Singer, 1999) , filtering is based on the strength of the context and its frequency. in our experiments) with the highest frequency and strength higher than the predefined threshold of 95%. Examples of selected context rules are shown in Figure 4 .", |
| "cite_spans": [ |
| { |
| "start": 993, |
| "end": 994, |
| "text": "3", |
| "ref_id": null |
| }, |
| { |
| "start": 1498, |
| "end": 1524, |
| "text": "(Collins and Singer, 1999)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 674, |
| "end": 682, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| }, |
| { |
| "start": 1749, |
| "end": 1757, |
| "text": "Figure 4", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "The co-training algorithm", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "The parameter of the contextual classifier is a context length. In our experiments we found that a maximal context length of three produces best results. We also observed that for some rules a shorter context works better. Therefore, when recording contexts around positive and negative examples, we record all the contexts with length smaller or equal to the maximal length.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The co-training algorithm", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "Because our corpus consists of translations of several books, created by different translators, we expect that the similarity between translations varies from one book to another. This implies that contextual rules should be specific to a particular pair of translations. Therefore, we train the contextual classifier for each pair of translations separately. Training of the paraphrasing classifier Context rules extracted in the previous stage are then applied to the corpus to derive a new set of pairs of positive and negative paraphrasing examples. Applications of the rule performed by searching sentence pairs for subsequences which match the left and right parts of the contextual rule, and are less than S tokens apart. For example, applying the first rule from Figure 4 to sentences from Figure 1 yields the paraphrasing pair (\"comfort\", \"console\"). Note that in the original seed set, the left and right contexts were separated by one token. This stretch in rule application allows us to extract multi-word paraphrases.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 771, |
| "end": 779, |
| "text": "Figure 4", |
| "ref_id": "FIGREF1" |
| }, |
| { |
| "start": 798, |
| "end": 806, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "The co-training algorithm", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "For each extracted example, paraphrasing rules are recorded and filtered in a similar manner as contextual rules. Examples of lexical and syntactic paraphrasing rules are shown in Figure 5 and in Figure 6 . After extracted lexical and syntactic paraphrases are applied to the corpus, the contextual classifier is retrained. New paraphrases not only add more positive and negative instances to the contextual classifier, but also revise contextual rules for known instances based on new paraphrase information. The iterative process is terminated when no new paraphrases are discovered or the number of iterations exceeds a predefined threshold.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 180, |
| "end": 188, |
| "text": "Figure 5", |
| "ref_id": null |
| }, |
| { |
| "start": 196, |
| "end": 204, |
| "text": "Figure 6", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "The co-training algorithm", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "Our algorithm produced 9483 pairs of lexical paraphrases and 25 morpho-syntactic rules. To evaluate the quality of produced paraphrases, we picked at random 500 paraphrasing pairs from the lexical paraphrases produced by our algorithm. These pairs were used as test data and also to evaluate whether humans agree on paraphrasing judgments. The judges were given a page of guidelines, defining paraphrase as \"approximate conceptual equivalence\". The main dilemma in designing the evaluation is whether to include the context: should the human judge see only a paraphrase pair or should a pair of sentences containing these paraphrases also be given? In a similar MT task -evaluation of word-to-word translation -context is usually included (Melamed, 2001) . Although paraphrasing is considered to be context dependent, there is no agreement on the extent. To evaluate the influence of context on paraphrasing judgments, we performed two experiments -with and without context. First, the human judge is given a paraphrase pair without context, and after the judge entered his answer, he is given the same pair with its surrounding context. Each context was evaluated by two judges (other than the authors). The agreement was measured using the Kappa coefficient (Siegel and Castellan, 1988) . Complete agreement between judges would correspond to K equals ; if there is no agreement among judges, then K equals H . The judges agreement on the paraphrasing judgment without context was", |
| "cite_spans": [ |
| { |
| "start": 739, |
| "end": 754, |
| "text": "(Melamed, 2001)", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 1260, |
| "end": 1288, |
| "text": "(Siegel and Castellan, 1988)", |
| "ref_id": "BIBREF23" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The results", |
| "sec_num": "6" |
| }, |
| { |
| "text": "U \u00a5 H 9 V X W 9 Y", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The results", |
| "sec_num": "6" |
| }, |
| { |
| "text": "which is substantial agreement (Landis and Koch, 1977) . The first judge found 439(87.8%) pairs as correct paraphrases, and the second judge -426(85.2%). Judgments with context have even higher agreement (\u00d9\u00a5 F H 9 V X a 9 b", |
| "cite_spans": [ |
| { |
| "start": 31, |
| "end": 54, |
| "text": "(Landis and Koch, 1977)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The results", |
| "sec_num": "6" |
| }, |
| { |
| "text": ") , and judges identified 459(91.8%) and 457(91.4%) pairs as correct paraphrases.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The results", |
| "sec_num": "6" |
| }, |
| { |
| "text": "The recall of our method is a more problematic issue. The algorithm can identify paraphrasing relations only between words which occurred in our corpus, which of course does not cover all English tokens. Furthermore, direct comparison with an electronic thesaurus like WordNet is impossible, because it is not known a priori which lexical relations in WordNet can form paraphrases. Thus, we can not evaluate recall. We hand-evaluated the coverage, by asking a human judges to extract paraphrases from 50 sentences, and then counted how many of these paraphrases where predicted by our algorithm. From 70 paraphrases extracted by human judge, 48(69%) were identified as paraphrases by our algorithm.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The results", |
| "sec_num": "6" |
| }, |
| { |
| "text": "In addition to evaluating our system output through precision and recall, we also compared our results with two other methods. The first of these was a machine translation technique for deriving bilingual lexicons (Melamed, 2001) including detection of non-compositional compounds 4 . We did this evaluation on 60% of the full dataset; this is the portion of the data which is publicly available. Our system produced 6,826 word pairs from this data and Melamed provided the top 6,826 word pairs resulting from his system on this data. We randomly extracted 500 pairs each from both sets of output. Of the 500 pairs produced by our system, 354(70.8%) were single word pairs and 146(29.2%) were multi-word paraphrases, while the majority of pairs produced by Melamed's system were single word pairs (90%). We mixed this output and gave the resulting, randomly ordered 1000 pairs to six evaluators, all of whom were native speakers. Each evaluator provided judgments on 500 pairs without context. Precision for our system was 71.6% and for Melamed's was 52.7%. This increased precision is a clear advantage of our approach and shows that machine translation techniques cannot be used without modification for this task, particularly for producing multi-word paraphrases. There are three caveats that should be noted; Melamed's system was run without changes for this new task of paraphrase extraction and his system does not use chunk segmentation, he ran the system for three days of computation and the result may be improved with more running time since it makes incremental improvements on subsequent rounds, and finally, the agreement between human judges was lower than in our previous experiments. We are currently exploring whether the information produced by the two different systems may be combined to improve the performance of either system alone.", |
| "cite_spans": [ |
| { |
| "start": 214, |
| "end": 229, |
| "text": "(Melamed, 2001)", |
| "ref_id": "BIBREF17" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The results", |
| "sec_num": "6" |
| }, |
| { |
| "text": "Another view on the extracted paraphrases can be derived by comparing them with the Word-Net thesaurus. This comparison provides us with 4 The equivalences that were identical on both sides were removed from the output quantitative evidence on the types of lexical relations people use to create paraphrases. We selected 112 paraphrasing pairs which occurred at least 20 times in our corpus and such that the words comprising each pair appear in WordNet. The 20 times cutoff was chosen to ensure that the identified pairs are general enough and not idiosyncratic. We use the frequency threshold to select paraphrases which are not tailored to one context. Examples of paraphrases and their WordNet relations are shown in Figure 7 . Only 40(35%) paraphrases are synonyms, 36(32%) are hyperonyms, 20(18%) are siblings in the hyperonym tree, 11(10%) are unrelated, and the remaining 5% are covered by other relations. These figures quantitatively validate our intuition that synonymy is not the only source of paraphrasing. One of the practical implications is that using synonymy relations exclusively to recognize paraphrasing limits system performance.", |
| "cite_spans": [ |
| { |
| "start": 137, |
| "end": 138, |
| "text": "4", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 721, |
| "end": 729, |
| "text": "Figure 7", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "The results", |
| "sec_num": "6" |
| }, |
| { |
| "text": "Synonyms: (rise, stand up), (hot, warm) Hyperonyms: (landlady, hostess), (reply, say) Siblings: (city, town), (pine, fir) Unrelated: (sick, tired), (next, then) Figure 7 : Lexical paraphrases extracted by the algorithm.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 161, |
| "end": 169, |
| "text": "Figure 7", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "The results", |
| "sec_num": "6" |
| }, |
| { |
| "text": "In this paper, we presented a method for corpusbased identification of paraphrases from multiple English translations of the same source text. We showed that a co-training algorithm based on contextual and lexico-syntactic features of paraphrases achieves high performance on our data. The wide range of paraphrases extracted by our algorithm sheds light on the paraphrasing phenomena, which has not been studied from an empirical perspective.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions and Future work", |
| "sec_num": "7" |
| }, |
| { |
| "text": "Future work will extend this approach to extract paraphrases from comparable corpora, such as multiple reports from different news agencies about the same event or different descriptions of a disease from the medical literature. This extension will require using a more selective alignment technique (similar to that of (Hatzivassiloglou et al., 1999) ). We will also investigate a more powerful representation of contextual features. Fortunately, statistical parsers produce reliable results on news texts, and therefore can be used to improve context representation. This will allow us to extract macro-syntactic paraphrases in addition to local paraphrases which are currently produced by the algorithm.", |
| "cite_spans": [ |
| { |
| "start": 320, |
| "end": 351, |
| "text": "(Hatzivassiloglou et al., 1999)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions and Future work", |
| "sec_num": "7" |
| }, |
| { |
| "text": "Foreign sources are not used in our experiment.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "This work was partially supported by a Louis Morin scholarship and by DARPA grant N66001-00-1-8919 under the TIDES program. We are grateful to Dan Melamed for providing us with the output of his program. We thank Noemie Elhadad, Mike Collins, Michael Elhadad and Maria Lapata for useful discussions.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgments", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "The CELEX Lexical Database(CD-ROM). Linguistic Data Consortium", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [ |
| "H" |
| ], |
| "last": "Baayen", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Piepenbrock", |
| "suffix": "" |
| }, |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Van Rijn", |
| "suffix": "" |
| } |
| ], |
| "year": 1993, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "R. H. Baayen, R. Piepenbrock, and H. van Rijn, editors. 1993. The CELEX Lexical Database(CD-ROM). Lin- guistic Data Consortium, University of Pennsylvania.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Using lexical chains for text summarization", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Barzilay", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Elhadad", |
| "suffix": "" |
| } |
| ], |
| "year": 1997, |
| "venue": "Proceedings of the ACL Workshop on Intelligent Scalable Text Summarization", |
| "volume": "", |
| "issue": "", |
| "pages": "10--17", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "R. Barzilay and M. Elhadad. 1997. Using lexical chains for text summarization. In Proceedings of the ACL Workshop on Intelligent Scalable Text Summarization, pages 10-17, Madrid, Spain, August.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "The mathematics of statistical machine translation: Parameter estimation", |
| "authors": [ |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Brown", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [ |
| "Della" |
| ], |
| "last": "Pietra", |
| "suffix": "" |
| }, |
| { |
| "first": "V", |
| "middle": [ |
| "Della" |
| ], |
| "last": "Pietra", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Mercer", |
| "suffix": "" |
| } |
| ], |
| "year": 1993, |
| "venue": "Computational Linguistics", |
| "volume": "19", |
| "issue": "2", |
| "pages": "263--311", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "P. Brown, S. Della Pietra, V. Della Pietra, and R. Mercer. 1993. The mathematics of statistical machine transla- tion: Parameter estimation. Computational Linguistics, 19(2):263-311.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Unsupervised models for named entity classification", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Collins", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [], |
| "last": "Singer", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "proceedings of the Joint SIGDAT Conference on Empirical Methods in Natural Language Processing and Very Large Corpora", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "M. Collins and Y. Singer. 1999. Unsupervised models for named entity classification. In proceedings of the Joint SIGDAT Conference on Empirical Methods in Natural Language Processing and Very Large Corpora.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Introduction to Text Linguistics", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "De Beaugrande", |
| "suffix": "" |
| }, |
| { |
| "first": "W", |
| "middle": [ |
| "V" |
| ], |
| "last": "Dressler", |
| "suffix": "" |
| } |
| ], |
| "year": 1981, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "R. de Beaugrande and W. V. Dressler. 1981. Introduction to Text Linguistics. Longman, New York, NY.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Tree Adjoining Grammar and the Reluctant Paraphrasing of Text", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Dras", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "M. Dras. 1999. Tree Adjoining Grammar and the Reluctant Paraphrasing of Text. Ph.D. thesis, Macquarie Univer- sity, Australia.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "A program for aligning sentences in bilingual corpora", |
| "authors": [ |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Gale", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [ |
| "W" |
| ], |
| "last": "Church", |
| "suffix": "" |
| } |
| ], |
| "year": 1991, |
| "venue": "Proceedings of the 29th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "1--8", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "W. Gale and K. W. Church. 1991. A program for align- ing sentences in bilingual corpora. In Proceedings of the 29th Annual Meeting of the Association for Computa- tional Linguistics, pages 1-8.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "An introduction to functional grammar", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Halliday", |
| "suffix": "" |
| } |
| ], |
| "year": 1985, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "M. Halliday. 1985. An introduction to functional grammar. Edward Arnold, UK.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Towards the automatic identification of adjectival scales: Clustering adjectives according to their meaning", |
| "authors": [ |
| { |
| "first": "V", |
| "middle": [], |
| "last": "Hatzivassiloglou", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [ |
| "R" |
| ], |
| "last": "Mckeown", |
| "suffix": "" |
| } |
| ], |
| "year": 1993, |
| "venue": "Proceedings of the 31rd Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "172--182", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "V. Hatzivassiloglou and K.R. McKeown. 1993. Towards the automatic identification of adjectival scales: Clustering adjectives according to their meaning. In Proceedings of the 31rd Annual Meeting of the Association for Compu- tational Linguistics, pages 172-182.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Detecting text similarity over short passages: Exploring linguistic feature combinations via machine learning", |
| "authors": [ |
| { |
| "first": "V", |
| "middle": [], |
| "last": "Hatzivassiloglou", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Klavans", |
| "suffix": "" |
| }, |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Eskin", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "proceedings of the Joint SIGDAT Conference on Empirical Methods in Natural Language Processing and Very Large Corpora", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "V. Hatzivassiloglou, J. Klavans, and E. Eskin. 1999. Detect- ing text similarity over short passages: Exploring linguis- tic feature combinations via machine learning. In pro- ceedings of the Joint SIGDAT Conference on Empirical Methods in Natural Language Processing and Very Large Corpora.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Natural language Generation in Artificial Intelligence and Computational Linguistics, chapter 11", |
| "authors": [ |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Iordanskaja", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Kittredge", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Polguere", |
| "suffix": "" |
| } |
| ], |
| "year": 1991, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "L. Iordanskaja, R. Kittredge, and A. Polguere, 1991. Natural language Generation in Artificial Intelligence and Com- putational Linguistics, chapter 11. Kluwer Academic Publishers.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Expansion of multi-word terms for indexing and retrieval using morphology and syntax", |
| "authors": [ |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Jacquemin", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Klavans", |
| "suffix": "" |
| }, |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Tzoukermann", |
| "suffix": "" |
| } |
| ], |
| "year": 1997, |
| "venue": "proceedings of the 35th", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "C. Jacquemin, J. Klavans, and E. Tzoukermann. 1997. Ex- pansion of multi-word terms for indexing and retrieval using morphology and syntax. In proceedings of the 35th", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Annual Meeting of the ACL", |
| "authors": [], |
| "year": null, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "24--31", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Annual Meeting of the ACL, pages 24-31, Madrid, Spain, July. ACL.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "The measurement of observer agreement for categorical data", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [ |
| "R" |
| ], |
| "last": "Landis", |
| "suffix": "" |
| }, |
| { |
| "first": "G", |
| "middle": [ |
| "G" |
| ], |
| "last": "Koch", |
| "suffix": "" |
| } |
| ], |
| "year": 1977, |
| "venue": "Biometrics", |
| "volume": "33", |
| "issue": "", |
| "pages": "159--174", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "J.R. Landis and G.G. Koch. 1977. The measurement of observer agreement for categorical data. Biometrics, 33:159-174.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Generation that exploits corpus-based statistical knowledge", |
| "authors": [ |
| { |
| "first": "I", |
| "middle": [], |
| "last": "Langkilde", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Knight", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "proceedings of the COLING-ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "I. Langkilde and K. Knight. 1998. Generation that exploits corpus-based statistical knowledge. In proceedings of the COLING-ACL.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "A corpus-based account of regular polysemy: The case of context-sensitive adjectives", |
| "authors": [ |
| { |
| "first": "Maria", |
| "middle": [], |
| "last": "Lapata", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the 2nd Meeting of the NAACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Maria Lapata. 2001. A corpus-based account of regular pol- ysemy: The case of context-sensitive adjectives. In Pro- ceedings of the 2nd Meeting of the NAACL, Pittsburgh, PA.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Automatic retrieval and clustering of similar words", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "proceedings of the COLING-ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "768--774", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "D. Lin. 1998. Automatic retrieval and clustering of similar words. In proceedings of the COLING-ACL, pages 768- 774.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Empirical Methods for Exploiting Parallel Texts", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Melamed", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Melamed. 2001. Empirical Methods for Exploiting Parallel Texts. MIT press.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "1997. the ltg part of speech tagger", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Mikheev", |
| "suffix": "" |
| } |
| ], |
| "year": null, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "A. Mikheev. 1997. the ltg part of speech tagger. University of Edinburgh.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Introduction to WordNet: An on-line lexical database", |
| "authors": [ |
| { |
| "first": "G", |
| "middle": [ |
| "A" |
| ], |
| "last": "Miller", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Beckwith", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Fellbaum", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Gross", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [ |
| "J" |
| ], |
| "last": "Miller", |
| "suffix": "" |
| } |
| ], |
| "year": 1990, |
| "venue": "International Journal of Lexicography (special issue)", |
| "volume": "3", |
| "issue": "4", |
| "pages": "235--245", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "G.A. Miller, R. Beckwith, C. Fellbaum, D. Gross, and K.J. Miller. 1990. Introduction to WordNet: An on-line lexi- cal database. International Journal of Lexicography (spe- cial issue), 3(4):235-245.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "Distributional clustering of english words", |
| "authors": [ |
| { |
| "first": "F", |
| "middle": [], |
| "last": "Pereira", |
| "suffix": "" |
| }, |
| { |
| "first": "N", |
| "middle": [], |
| "last": "Tishby", |
| "suffix": "" |
| }, |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| } |
| ], |
| "year": 1993, |
| "venue": "proceedings of the 30th Annual Meeting of the ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "183--190", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "F. Pereira, N. Tishby, and L. Lee. 1993. Distributional clus- tering of english words. In proceedings of the 30th An- nual Meeting of the ACL, pages 183-190. ACL.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "Learning Dictionaries for Information Extraction by Multi-level Boot-strapping", |
| "authors": [ |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Riloff", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Jones", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "Proceedings of the Sixteenth National Conference on Artificial Intelligence", |
| "volume": "", |
| "issue": "", |
| "pages": "1044--1049", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "E. Riloff and R. Jones. 1999. Learning Dictionaries for Information Extraction by Multi-level Boot-strapping. In Proceedings of the Sixteenth National Conference on Artificial Intelligence, pages 1044-1049. The AAAI Press/MIT Press.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "Revision-Based Generation of Natural Language Summaries Providing Historical Background: Corpus-Based Analysis, Design, Implementation, and Evaluation", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Robin", |
| "suffix": "" |
| } |
| ], |
| "year": 1994, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "J. Robin. 1994. Revision-Based Generation of Natural Language Summaries Providing Historical Background: Corpus-Based Analysis, Design, Implementation, and Evaluation. Ph.D. thesis, Department of Computer Sci- ence, Columbia University, NY.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "Non Parametric Statistics for Behavioral Sciences", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Siegel", |
| "suffix": "" |
| }, |
| { |
| "first": "N", |
| "middle": [ |
| "J" |
| ], |
| "last": "Castellan", |
| "suffix": "" |
| } |
| ], |
| "year": 1988, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "S. Siegel and N.J. Castellan. 1988. Non Parametric Statis- tics for Behavioral Sciences. McGraw-Hill.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "Parallel Text Processing: Alignment and Use of Translation Corpora", |
| "authors": [], |
| "year": 2000, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "J. Veronis, editor. 2000. Parallel Text Processing: Align- ment and Use of Translation Corpora. Kluwer Academic Publishers.", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "Performing Without a Stage: The Art of Literary Translation", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Wechsler", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "R. Wechsler. 1998. Performing Without a Stage: The Art of Literary Translation. Catbird Press.", |
| "links": null |
| }, |
| "BIBREF26": { |
| "ref_id": "b26", |
| "title": "Unsupervised word sense disambiguation rivaling supervised methods", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Yarowsky", |
| "suffix": "" |
| } |
| ], |
| "year": 1995, |
| "venue": "Proceedings of the 33rd Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "189--196", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "D. Yarowsky. 1995. Unsupervised word sense disambigua- tion rivaling supervised methods. In Proceedings of the 33rd Annual Meeting of the Association for Computa- tional Linguistics, pages 189-196.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "type_str": "figure", |
| "text": "Two English translations of the French sentence from Flaubert's \"Madame Bovary\"", |
| "num": null, |
| "uris": null |
| }, |
| "FIGREF1": { |
| "type_str": "figure", |
| "text": "leftI = (VBP TOI ) rightI = (PRP$Q ,) leftQ = (VBP TOI ) rightQ = (PRP$Q ,) leftI = (WRBP NNI ) rightI = (NNQ IN) leftQ = (WRBP NNI ) rightQ = (NNQ IN) leftI = (VBP ) rightI = (JJI ) leftQ = (VBP ) rightQ = (JJI ) leftI = (IN NNP ) rightI = (NNQ INR ) leftQ = (NNP ,) rightQ = (NNQ INR ) Example of context rules extracted by the algorithm.", |
| "num": null, |
| "uris": null |
| }, |
| "FIGREF2": { |
| "type_str": "figure", |
| "text": "NNP POS NNI )T (NNI IN DT NNP )King's son son of the king (", |
| "num": null, |
| "uris": null |
| }, |
| "FIGREF3": { |
| "type_str": "figure", |
| "text": "Morpho-Syntactic patterns extracted by the algorithm. Lower indices denote token equivalence, upper indices denote root equivalence.(countless, lots of) (repulsion, aversion) (undertone, low voice) (shrubs, bushes) (refuse, say no) (dull tone, gloom) (sudden appearance, apparition) Lexical paraphrases extracted by the algorithm.", |
| "num": null, |
| "uris": null |
| } |
| } |
| } |
| } |