| { |
| "paper_id": "2016", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T01:04:01.735863Z" |
| }, |
| "title": "A language-independent LESK based approach to Word Sense Disambiguation", |
| "authors": [ |
| { |
| "first": "Tommaso", |
| "middle": [], |
| "last": "Petrolito", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "University of Pisa", |
| "location": { |
| "country": "Italy" |
| } |
| }, |
| "email": "tommasouni@gmail.com" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "This paper describes a languageindependent LESK based approach to Word Sense Disambiguation (WSD), involving also Vector Space Models applied to the Distributional Semantics Hypotesis. In particular this approach tries to solve some issues that come up with less-resourced languages. The approach also addresses the inadequacy of the Most Frequent Sense (MFS) heuristics to fit specific domain corpora.", |
| "pdf_parse": { |
| "paper_id": "2016", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "This paper describes a languageindependent LESK based approach to Word Sense Disambiguation (WSD), involving also Vector Space Models applied to the Distributional Semantics Hypotesis. In particular this approach tries to solve some issues that come up with less-resourced languages. The approach also addresses the inadequacy of the Most Frequent Sense (MFS) heuristics to fit specific domain corpora.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "This language independent approach to WSD, even if in a very early stage of development, tries to solve two main problems.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "1. Variable quality of glosses and examples (the solution would be to use glosses and examples for the aligned synsets in several languages, we will explain how).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "2. Weakness of Most Frequent Sense heuristics for domain corpora (or even general corpora that, for some reasons, are not so similar to the corpus on which the frequencies were calculated), but also lack of synset annotated corpora for several non-English languages (the solution would involve Space Vector Models, we will explain how).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "We use Wordnet (WN) resources (Miller et al., 1990; Miller, 1995; Fellbaum, 1998 ) (synsets glosses and examples) from a specific standpoint: preferring to avoid the usage of monolingual resources, even though the specific task does not involve cross-lingual WSD on aligned parallel corpora.", |
| "cite_spans": [ |
| { |
| "start": 30, |
| "end": 51, |
| "text": "(Miller et al., 1990;", |
| "ref_id": "BIBREF16" |
| }, |
| { |
| "start": 52, |
| "end": 65, |
| "text": "Miller, 1995;", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 66, |
| "end": 80, |
| "text": "Fellbaum, 1998", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "It has to be pointed out that the approach is being considered 'unsupervised': it does not rely on semantic annotation, although lemmatization and PoS-tagging are taken into account.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In most cases the quality of lexical resources is very variable, even though some languages have good resources, as of course English and, for instance, Italian with both MultiWordnet (Pianta et al., 2002) and ItalWordnet (Roventini et al., 2000) .", |
| "cite_spans": [ |
| { |
| "start": 184, |
| "end": 205, |
| "text": "(Pianta et al., 2002)", |
| "ref_id": "BIBREF22" |
| }, |
| { |
| "start": 222, |
| "end": 246, |
| "text": "(Roventini et al., 2000)", |
| "ref_id": "BIBREF25" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "An example is given with the dog/c\u00e2ine (first synset) glosses and examples 1 in Table 1 . It is evident that the English synset has a richer gloss.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 80, |
| "end": 87, |
| "text": "Table 1", |
| "ref_id": "TABREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Assuming a WSD approach involving overlap counts, the English words Canis, wolf, breeds will be counted in; as for the Romanian words Animal, mamifer, carnivor (all IS-A relations), their English lemmas would be reached in any case in an Expanded Gloss implementation.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Anyway, paz\u0203 and v\u00e2n\u0203toare ('guarding' and 'hunting') would be useful for the same task.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In the counterexample given in Table 2 , the Romanian gloss is evidently richer than the English one, in particular using a WSD overlapping algorithm that is able to count on asisten\u0163\u0203, social\u0203, \u00eentre\u0163inerea, b\u0203tr\u00e2nilor ('assistance', 'social', 'maintenance', 'elders') and so on.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 31, |
| "end": 38, |
| "text": "Table 2", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In general, it can be noticed how variable the quality is for different corpora and for different synsets.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Anyway, usually English WN provides the best and richest set of examples for a given synset.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "This variability in quality is observable also concerning the coverage of different WNs 2 (Bond and Foster, 2013 ).", |
| "cite_spans": [ |
| { |
| "start": 90, |
| "end": 112, |
| "text": "(Bond and Foster, 2013", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "1 For a quick series of examples of this kind, just have a look on multilingual aligned synsets on the MultiWordnet Interface (Ranieri et al., 2004) .", |
| "cite_spans": [ |
| { |
| "start": 126, |
| "end": 148, |
| "text": "(Ranieri et al., 2004)", |
| "ref_id": "BIBREF23" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "See http://multiwordnet.fbk.eu/online/ multiwordnet.php 2 See http://compling.hss.ntu.edu. sg/omw/ and http://globalwordnet.org/ wordnets-in-the-world/ for an overview.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Gloss dog,domestic_dog, EN a member of the genus Canis (probably descended Canis_familiaris/1 from the common wolf) that has been domesticated by man since prehistoric times; occurs in many breeds; c\u00e2ine/1 RO Animal mamifer carnivor domesticit, folosit pentru paz\u0203, v\u00e2n\u0203toare etc.. Kilgarriff and Rosenzweig (2000a) , Kilgarriff and Rosenzweig (2000b) , Banerjee and Pedersen (2002) and Basile et al. (2014) . Our approach is an adaptation that takes into account the issues about glosses and examples quality.", |
| "cite_spans": [ |
| { |
| "start": 282, |
| "end": 315, |
| "text": "Kilgarriff and Rosenzweig (2000a)", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 318, |
| "end": 351, |
| "text": "Kilgarriff and Rosenzweig (2000b)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 354, |
| "end": 382, |
| "text": "Banerjee and Pedersen (2002)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 387, |
| "end": 407, |
| "text": "Basile et al. (2014)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lang", |
| "sec_num": null |
| }, |
| { |
| "text": "In particular we try to gain advantage from the usage of better resources available in other languages.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lang", |
| "sec_num": null |
| }, |
| { |
| "text": "A first baseline attempt was tested here, due to time constraints: relying on English glosses and examples with non-English target corpora. For future work, a more complex adaptation will be attempted, trying to take advantage of glosses and examples in several languages at once. This kind of approach leads to two main issues. Either trying to use many glosses and examples from WNs in several languages or trying to use glosses and examples just from the Princeton Wordnet (PWN) working on a non-English corpus, the issue arises of how to compare the contexts of the target words with the glosses and examples of their candidate synsets.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lang", |
| "sec_num": null |
| }, |
| { |
| "text": "The problem needs to be addressed, for instance, if the contexts of the target words are in a given language (not English) that is not compatible with an overlapping approach involving words from glosses and examples in many different languages or even just with the English ones.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lang", |
| "sec_num": null |
| }, |
| { |
| "text": "Second, the widely used Most Frequent Sense (MFS) heuristics (Gale et al., 1992; Miller et al., 1994; Kohomban and Lee, 2005) , easily implementable in English by choosing the first synset for the given lemma, cannot be used when working with other languages, as the synset ordering does not mirror sense frequency statistics.", |
| "cite_spans": [ |
| { |
| "start": 61, |
| "end": 80, |
| "text": "(Gale et al., 1992;", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 81, |
| "end": 101, |
| "text": "Miller et al., 1994;", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 102, |
| "end": 125, |
| "text": "Kohomban and Lee, 2005)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lang", |
| "sec_num": null |
| }, |
| { |
| "text": "Even working on English, MFS' usefulness varies accordingly to the similarity of the target corpus to SemCor (Mihalcea, 1998) , concerning the topic(s) and the granularity of meanings.", |
| "cite_spans": [ |
| { |
| "start": 109, |
| "end": 125, |
| "text": "(Mihalcea, 1998)", |
| "ref_id": "BIBREF14" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lang", |
| "sec_num": null |
| }, |
| { |
| "text": "Also this issue needs a proper solution and some help can come from Vector Space Models (VSMs) applied to the Distributional Hypothesis (DH) (Harris, 1951; Turney et al., 2010) of Semantics implementing Distributional Semantic Spaces (DSS).", |
| "cite_spans": [ |
| { |
| "start": 141, |
| "end": 155, |
| "text": "(Harris, 1951;", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 156, |
| "end": 176, |
| "text": "Turney et al., 2010)", |
| "ref_id": "BIBREF27" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lang", |
| "sec_num": null |
| }, |
| { |
| "text": "This approach is organized in two disambiguation steps.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Methodology", |
| "sec_num": "2" |
| }, |
| { |
| "text": "The first (focused on quality) is based mainly on a kind of LESK adapted in the languageindependent perspective discussed above and involves WN glosses and examples.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Methodology", |
| "sec_num": "2" |
| }, |
| { |
| "text": "The second (focused on quantity) is based on VSMs and follows the assumption, coherent with the Distributional Hypothesis, that the neighbours of the target word in the Semantic Space are semantically related (in paradigmatic relations) with the target word.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Methodology", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Both these two disambiguation steps will be discussed in this section.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Methodology", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Our idea consists in counting the overlaps in couples of candidate-synset-bag-of-lemmas and context-bag-of-lemmas.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Then, the candidate synset for which the count is higher is chosen.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Let us assume that we use an Italian sentence, but we want to rely on English synsets glosses and examples (we will explain later why we would want to do that).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Let us take the Italian sentence:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Il cane abbaia spesso quando fa la guardia ai suoi giocattoli o al suo cibo \"The dog often barks when guarding its toys or food\"", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Given cane (\"dog\") as our target lemma and n, 'noun' as part of speech, the algorithm has to fulfill the following steps:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "1. Find all the Italian synsets associated to the given lemma and part of speech that are aligned to the English WordNet.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "2. For each candidate synset, build a 'bag of lemmas' by retrieving all content words found in the English gloss and example(s) and lemmatizing them.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "3. For each sentence (in this case the current sentence containing cane), build a context bag of lemmas by taking English glosses and examples of the English synsets aligned to the Italian synsets of the words in the sentence (lemma and part of speech annotations are assumed to be there).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "To avoid a computational nightmare (and maybe also to avoid noise), only unambiguous lemmas and lemmas with a number of synsets less than an upper bound, previously defined, will be taken into account as sources of synset-glosses and synset-examples.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "The synset for which the overlapping between the two bags is bigger is the chosen one.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "With this approach we want to show that, theoretically, one can benefit from the semantic information available in different languages to help solve the ambiguity, even though the task doesn't start off as multilingual.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "This means that theoretically we can disambiguate an Italian text using information from a WN in any language. Now, let us suppose to use at once pairs of English bags (as explained above) and other pairs of bags of lemmas, built in the same way, but taken from WNs of other languages.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "So we will have for each synset of cane a bag with lemmas from each language (separately).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Similarly, for the words in the sentence there will be a bag of lemmas for each language.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Let us take one 'monolingual' group at the time. Each bag-of-lemmas pair (one from the candidate synset, one from the sentence words) will have an overlapping score. We can take into account all the scores, for example by summing them then choose the synset that has the higher total score.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Why should all this improve the results? Let us suppose to include Romanian WN in these group of wordnets and try to disambiguate cane in the same Italian sentence seen above:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Il cane abbaia spesso quando fa la guardia ai suoi giocattoli o al suo cibo \"The dog often barks when guarding its toys or food\"", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "We point out that dog.n.01 and cane.n.01 (respectively the English and Italian first synsets for the Italian lemma cane) have glosses and examples with no mention to 'hunting' or 'guarding', while the gloss of the Romanian synset (c\u00e2ine.n.01) refers to both.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "The context word guardia ('guard') would be exploited much better by using the Romanian WN than by using the Italian one, even though the language of the text is Italian.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "The same thing could happen with English (or any other language) texts about dogs in which 'guarding' and 'hunting' words are not exploited by a monolingual LESK approach.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "This case is an evidence of how a multilingual approach, involving comparisons between the bags for the candidate synsets and for the context in several languages, could enhance overlapping counts and lead to a better synset selection.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "We have provided an example showing that this approach can be applied also by building many sub-bags in distinct languages (and this was the full original idea): for each synset existing in English, Italian and Romanian (for example) a list containing the three monolingual bags can be built and the synset-scores can take into account the overlapping in all the languages (summing the overlapping scores together), taking advantage from eventual better quality (or even just few lucky occurring keywords) in the glosses and examples in other languages.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Language independent LESK algorithm", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "For the future, a more complex and representative scoring measure will be defined, maybe taking into account the good example provided by Basile et al. (2014) based on different weights for lemmas.", |
| "cite_spans": [ |
| { |
| "start": 138, |
| "end": 158, |
| "text": "Basile et al. (2014)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Candidate synsets scoring", |
| "sec_num": "2.1.1" |
| }, |
| { |
| "text": "In the current version, due to time constraints, each synset gains a very simple score equal to the number of lemmas shared by the candidatesynset-bag and the context-synsets-bag (that is the union of the single synset-bags occurring in the sentence).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Candidate synsets scoring", |
| "sec_num": "2.1.1" |
| }, |
| { |
| "text": "Only one specific customization is added to this naive scoring approach: unambiguous lemmas in the context have double weight (so their overlapping will be counted twice).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Candidate synsets scoring", |
| "sec_num": "2.1.1" |
| }, |
| { |
| "text": "Here we show a baseline experiment exploiting only English glosses and examples on an Italian target corpus.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "If we set a configuration that takes context lemmas from words linked to a certain number of synsets (up to 6), this algorithm tags correctly the 36.17% of words in the Italian MultiSemCor (Pianta and Bentivogli, 2003; ).", |
| "cite_spans": [ |
| { |
| "start": 201, |
| "end": 218, |
| "text": "Bentivogli, 2003;", |
| "ref_id": "BIBREF21" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "If we use it to remove the wrong synsets it works much better: removing, for each target word, synsets with score lower than max_score/2, 65% of words still have right synsets in the remaining set of synsets.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "As for the second issue, concerning the Most Frequent heuristics, VSMs could provide a big help.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Paradigmatic relations algorithm", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "In particular, while the first disambiguation step focuses on the specificity of meanings observed in the specific contexts, a help from distributional quantities would focus on the frequencies of cooccurrences, thus providing a frequency based heuristics.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Paradigmatic relations algorithm", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "So, while the LESK based approach is contextdependent (so it will select different synsets for different usages of the same lemma in different contexts), the highest frequency heuristics would just help by pushing for the only one synset (always the same) that is the most frequent for the given lemma (independently whether observed in different contexts) in the corpus on which the frequencies have been measured.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Paradigmatic relations algorithm", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "A way to reproduce that kind of heuristics, even for languages with lack in synsets-annotated corpora 3 (Petrolito and Bond, 2014) (even well resourced languages as Italian cannot provide such resources for corpora other than SemCor), could be implemented as a WSD algorithm involving a Distributional Semantic Space.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Paradigmatic relations algorithm", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "An example is provided by (McCarthy et al., 2004) . McCarthy et al. (2004) use a thesaurus, acquired from automatically parsed text, based on the method of Lin (1998) , in order to find the predominant sense of a target word.", |
| "cite_spans": [ |
| { |
| "start": 26, |
| "end": 49, |
| "text": "(McCarthy et al., 2004)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 52, |
| "end": 74, |
| "text": "McCarthy et al. (2004)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 156, |
| "end": 166, |
| "text": "Lin (1998)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Paradigmatic relations algorithm", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "This thesaurus provides, through distributional similarity scores, the nearest neighbours to each target word. Then they use the WordNet similarity package (Patwardhan and Pedersen, 2003) to obtain semantic similarity measures to weight the contribution that each neighbour gives to the various senses of the target word.", |
| "cite_spans": [ |
| { |
| "start": 156, |
| "end": 187, |
| "text": "(Patwardhan and Pedersen, 2003)", |
| "ref_id": "BIBREF19" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Paradigmatic relations algorithm", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Here we do something similar, but we specifically exploit paradigmatic relations.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Paradigmatic relations algorithm", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "1. In the DSS, neighbour words with high cosine similarity share the same contexts and are therefore supposed to be in paradigmatic relation.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Paradigmatic relations algorithm", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "2. Also through WordNet we can infer words in paradigmatic relation with the target word, such as hypernyms, hyponyms, cohyponyms, synonyms and antonyms.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Paradigmatic relations algorithm", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Also this method consists of measuring the overlapping between bags of lemmas, as for the algorithm described previously.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Paradigmatic relations algorithm", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "The first bag contains the N (chosen arbitrarily) neighbours of the target lemma in the Semantic Space.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Paradigmatic relations algorithm", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Both the target lemma and the neighbours are combinations lemma-PoS (the lemma and PoS information is assumed to be there).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Paradigmatic relations algorithm", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "The second bag contains lemmas taken through an exploration of the paradigmatic relations in the WN ontology for the candidate synset.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Paradigmatic relations algorithm", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "So for the candidate synset the following will be taken: lemmas, antonyms of lemmas, hyper-nyms lemmas, hyponyms lemmas, cohyponyms (hyponyms of the hypernyms) lemmas.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Paradigmatic relations algorithm", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Actually, also in this bag (as for the one of the neighbours) instead of simple lemmas, we have combinations of lemma-PoS (in this case the information is obviously provided because we are taking lemmas from WN synsets).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Paradigmatic relations algorithm", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Then the synset with maximum score among the overlapping values between the Semantic Space 'neighbourhood' and the paradigmatic-relationsbag is selected.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Paradigmatic relations algorithm", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Also in this case the score is a simple count of the intersection between the two bags.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Candidate synsets scoring", |
| "sec_num": "2.2.1" |
| }, |
| { |
| "text": "Also this algorithm on its own is not achieving good performances, only a 34.5% of correct annotations on the Italian SemCor.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "2.2.2" |
| }, |
| { |
| "text": "All the experiments have been done on the Italian MultiSemCor (Pianta and Bentivogli, 2003; corpus, already sense-tagged.", |
| "cite_spans": [ |
| { |
| "start": 62, |
| "end": 91, |
| "text": "(Pianta and Bentivogli, 2003;", |
| "ref_id": "BIBREF21" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data Set", |
| "sec_num": "3" |
| }, |
| { |
| "text": "SemCor is the perfect data set for this task, as it is the first case of corpus annotated with WN synsets and it is available in various languages (English, Italian, Romanian and Japanese). The Italian MultiSemCor contains 14,144 sentences and 261,283 tokens, 119,802 of which are annotated with senses.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data Set", |
| "sec_num": "3" |
| }, |
| { |
| "text": "The availability in a good number of languages makes MultiSemCor a good resource to try this language-independent approach.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data Set", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Also the NTU-Multilingual Corpus (NTU-MC) (Tan and Bond, 2011) could be a perfect resource for this kind of experiments.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data Set", |
| "sec_num": "3" |
| }, |
| { |
| "text": "NTU-MC is a corpus designed to be multilingual from the start. It contains parallel text in eight languages: English, Mandarin Chinese, Japanese, Indonesian, Korean, Arabic, Vietnamese and Thai.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data Set", |
| "sec_num": "3" |
| }, |
| { |
| "text": "The two algorithms have been implemented as Python scripts importing the NLTK (Bird, 2006) WN Interface and the Gensim (\u0158eh\u016f\u0159ek and Sojka, 2010) word2vec (Mikolov et al., 2013) library.", |
| "cite_spans": [ |
| { |
| "start": 78, |
| "end": 90, |
| "text": "(Bird, 2006)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 154, |
| "end": 176, |
| "text": "(Mikolov et al., 2013)", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Implementation and Evaluation", |
| "sec_num": "4" |
| }, |
| { |
| "text": "At first, the two algorithms were implemented separately, achieving the results discussed in Subsection 2.1.2 and Subsection 2.2.2, then the two algorithms have been implemented together in sequence.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Implementation and Evaluation", |
| "sec_num": "4" |
| }, |
| { |
| "text": "The first algorithm has been used for a first disambiguation step excluding candidate synsets with scores lower than the 50% of the maximum, then the second algorithm has been applied taking into account only the remaining candidate synsets (provided by the first step of disambiguation) instead of considering all the possible synsets.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Implementation and Evaluation", |
| "sec_num": "4" |
| }, |
| { |
| "text": "When the candidate with higher score in the paradigmatic relations algorithm differs from the one with higher score in the LESK based one, the two scores are normalized in a minimummaximum 0-1, range and the candidate synset with the highest average is chosen.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Implementation and Evaluation", |
| "sec_num": "4" |
| }, |
| { |
| "text": "The results have improved a lot achieving an encouraging result: 38.67% of the content words have been correctly annotated, with a maximum number of 6 synsets for the context words.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Implementation and Evaluation", |
| "sec_num": "4" |
| }, |
| { |
| "text": "There is reason to hope that some further attempts based on the approach described in this paper will lead to significant improvements in languageindependent WSD.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Future Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "A first improvement will be exploiting the disambiguated glosses at least for English, as most of the English glosses are disambiguated.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Future Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "A second improvement will be the extension of the LESK based algorithm with other languages; considered that many glosses are translations of English, we should focus on Merge WNs (Dutch, Polish, etc) in particular.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Future Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "To do that it will be useful to extend NLTK multilingual support: the .definition() and .examples() methods of WN synsets would be much more useful for tasks like this by exploiting a lang attribute.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Future Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "A third improvement will be a further development of scores definitions and a complete testing of parameters like: for the first algorithm, the lower bound for the candidate synsets to be saved and passed to the second step of disambiguation and the upper bound for the number of synsets of the context words; for the second algorithm, the number of neighbours or even try to include the approach defined by McCarthy et al. (2004) .", |
| "cite_spans": [ |
| { |
| "start": 408, |
| "end": 430, |
| "text": "McCarthy et al. (2004)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Future Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "See http://globalwordnet.org/wordnetannotated-corpora/", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "An adapted lesk algorithm for word sense disambiguation using wordnet", |
| "authors": [ |
| { |
| "first": "Satanjeev", |
| "middle": [], |
| "last": "Banerjee", |
| "suffix": "" |
| }, |
| { |
| "first": "Ted", |
| "middle": [], |
| "last": "Pedersen", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "Computational linguistics and intelligent text processing", |
| "volume": "", |
| "issue": "", |
| "pages": "136--145", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Satanjeev Banerjee and Ted Pedersen. 2002. An adapted lesk algorithm for word sense disambigua- tion using wordnet. In Computational linguis- tics and intelligent text processing, pages 136-145. Springer.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "An enhanced lesk word sense disambiguation algorithm through a distributional semantic model", |
| "authors": [ |
| { |
| "first": "Pierpaolo", |
| "middle": [], |
| "last": "Basile", |
| "suffix": "" |
| }, |
| { |
| "first": "Annalina", |
| "middle": [], |
| "last": "Caputo", |
| "suffix": "" |
| }, |
| { |
| "first": "Giovanni", |
| "middle": [], |
| "last": "Semeraro", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of COLING", |
| "volume": "", |
| "issue": "", |
| "pages": "1591--1600", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Pierpaolo Basile, Annalina Caputo, and Giovanni Se- meraro. 2014. An enhanced lesk word sense dis- ambiguation algorithm through a distributional se- mantic model. In Proceedings of COLING, pages 1591-1600.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Exploiting parallel texts in the creation of multilingual semantically annotated resources: the multisemcor corpus", |
| "authors": [ |
| { |
| "first": "Luisa", |
| "middle": [], |
| "last": "Bentivogli", |
| "suffix": "" |
| }, |
| { |
| "first": "Emanuele", |
| "middle": [], |
| "last": "Pianta", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Natural Language Engineering", |
| "volume": "11", |
| "issue": "3", |
| "pages": "247--261", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Luisa Bentivogli and Emanuele Pianta. 2005. Ex- ploiting parallel texts in the creation of multilingual semantically annotated resources: the multisemcor corpus. Natural Language Engineering, 11(3):247- 261.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Multisemcor: an english italian aligned corpus with a shared inventory of senses", |
| "authors": [ |
| { |
| "first": "Luisa", |
| "middle": [], |
| "last": "Bentivogli", |
| "suffix": "" |
| }, |
| { |
| "first": "Emanuele", |
| "middle": [], |
| "last": "Pianta", |
| "suffix": "" |
| }, |
| { |
| "first": "Marcello", |
| "middle": [], |
| "last": "Ranieri", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Proceedings of the Meaning Workshop", |
| "volume": "90", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Luisa Bentivogli, Emanuele Pianta, and Marcello Ranieri. 2005. Multisemcor: an english italian aligned corpus with a shared inventory of senses. In Proceedings of the Meaning Workshop, volume 90.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Nltk: the natural language toolkit", |
| "authors": [ |
| { |
| "first": "Steven", |
| "middle": [], |
| "last": "Bird", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of the COLING/ACL on Interactive presentation sessions", |
| "volume": "", |
| "issue": "", |
| "pages": "69--72", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Steven Bird. 2006. Nltk: the natural language toolkit. In Proceedings of the COLING/ACL on Interactive presentation sessions, pages 69-72. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Linking and extending an open multilingual wordnet", |
| "authors": [ |
| { |
| "first": "Francis", |
| "middle": [], |
| "last": "Bond", |
| "suffix": "" |
| }, |
| { |
| "first": "Ryan", |
| "middle": [], |
| "last": "Foster", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "ACL (1)", |
| "volume": "", |
| "issue": "", |
| "pages": "1352--1362", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Francis Bond and Ryan Foster. 2013. Linking and ex- tending an open multilingual wordnet. In ACL (1), pages 1352-1362.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "A method for disambiguating word senses in a large corpus", |
| "authors": [ |
| { |
| "first": "William", |
| "middle": [ |
| "A" |
| ], |
| "last": "Gale", |
| "suffix": "" |
| }, |
| { |
| "first": "Kenneth", |
| "middle": [ |
| "W" |
| ], |
| "last": "Church", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Yarowsky", |
| "suffix": "" |
| } |
| ], |
| "year": 1992, |
| "venue": "Computers and the Humanities", |
| "volume": "26", |
| "issue": "5-6", |
| "pages": "415--439", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "William A. Gale, Kenneth W. Church, and David Yarowsky. 1992. A method for disambiguating word senses in a large corpus. Computers and the Humanities, 26(5-6):415-439.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Methods in structural linguistics", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Zellig", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Harris", |
| "suffix": "" |
| } |
| ], |
| "year": 1951, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Zellig S Harris. 1951. Methods in structural linguis- tics.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "English senseval: Report and results", |
| "authors": [ |
| { |
| "first": "Adam", |
| "middle": [], |
| "last": "Kilgarriff", |
| "suffix": "" |
| }, |
| { |
| "first": "Joseph", |
| "middle": [], |
| "last": "Rosenzweig", |
| "suffix": "" |
| } |
| ], |
| "year": 2000, |
| "venue": "LREC", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Adam Kilgarriff and Joseph Rosenzweig. 2000a. En- glish senseval: Report and results. In LREC.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Framework and results for english senseval", |
| "authors": [ |
| { |
| "first": "Adam", |
| "middle": [], |
| "last": "Kilgarriff", |
| "suffix": "" |
| }, |
| { |
| "first": "Joseph", |
| "middle": [], |
| "last": "Rosenzweig", |
| "suffix": "" |
| } |
| ], |
| "year": 2000, |
| "venue": "Computers and the Humanities", |
| "volume": "34", |
| "issue": "", |
| "pages": "15--48", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Adam Kilgarriff and Joseph Rosenzweig. 2000b. Framework and results for english senseval. Com- puters and the Humanities, 34(1-2):15-48.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Learning semantic classes for word sense disambiguation", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Upali", |
| "suffix": "" |
| }, |
| { |
| "first": "Wee", |
| "middle": [], |
| "last": "Kohomban", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Sun Lee", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Proceedings of the 43rd Annual Meeting on Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "34--41", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Upali S Kohomban and Wee Sun Lee. 2005. Learn- ing semantic classes for word sense disambiguation. In Proceedings of the 43rd Annual Meeting on Asso- ciation for Computational Linguistics, pages 34-41. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Automatic retrieval and clustering of similar words", |
| "authors": [ |
| { |
| "first": "Dekang", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "Proceedings of the 36th Annual Meeting of the Association for Computational Linguistics and 17th International Conference on Computational Linguistics", |
| "volume": "2", |
| "issue": "", |
| "pages": "768--774", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dekang Lin. 1998. Automatic retrieval and cluster- ing of similar words. In Proceedings of the 36th Annual Meeting of the Association for Computa- tional Linguistics and 17th International Conference on Computational Linguistics-Volume 2, pages 768- 774. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Finding predominant word senses in untagged text", |
| "authors": [ |
| { |
| "first": "Diana", |
| "middle": [], |
| "last": "Mccarthy", |
| "suffix": "" |
| }, |
| { |
| "first": "Rob", |
| "middle": [], |
| "last": "Koeling", |
| "suffix": "" |
| }, |
| { |
| "first": "Julie", |
| "middle": [], |
| "last": "Weeds", |
| "suffix": "" |
| }, |
| { |
| "first": "John", |
| "middle": [], |
| "last": "Carroll", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proceedings of the 42nd Annual Meeting on Association for Computational Linguistics, page 279. Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Diana McCarthy, Rob Koeling, Julie Weeds, and John Carroll. 2004. Finding predominant word senses in untagged text. In Proceedings of the 42nd Annual Meeting on Association for Computational Linguis- tics, page 279. Association for Computational Lin- guistics.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Semcor semantically tagged corpus", |
| "authors": [ |
| { |
| "first": "Rada", |
| "middle": [], |
| "last": "Mihalcea", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Rada Mihalcea. 1998. Semcor semantically tagged corpus. Unpublished manuscript.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Efficient estimation of word representations in vector space", |
| "authors": [ |
| { |
| "first": "Tomas", |
| "middle": [], |
| "last": "Mikolov", |
| "suffix": "" |
| }, |
| { |
| "first": "Kai", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Greg", |
| "middle": [], |
| "last": "Corrado", |
| "suffix": "" |
| }, |
| { |
| "first": "Jeffrey", |
| "middle": [], |
| "last": "Dean", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1301.3781" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tomas Mikolov, Kai Chen, Greg Corrado, and Jef- frey Dean. 2013. Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Introduction to wordnet: An on-line lexical database*", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "George", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Miller", |
| "suffix": "" |
| }, |
| { |
| "first": "Christiane", |
| "middle": [], |
| "last": "Beckwith", |
| "suffix": "" |
| }, |
| { |
| "first": "Derek", |
| "middle": [], |
| "last": "Fellbaum", |
| "suffix": "" |
| }, |
| { |
| "first": "Katherine", |
| "middle": [ |
| "J" |
| ], |
| "last": "Gross", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Miller", |
| "suffix": "" |
| } |
| ], |
| "year": 1990, |
| "venue": "International journal of lexicography", |
| "volume": "3", |
| "issue": "4", |
| "pages": "235--244", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "George A Miller, Richard Beckwith, Christiane Fell- baum, Derek Gross, and Katherine J Miller. 1990. Introduction to wordnet: An on-line lexical database*. International journal of lexicography, 3(4):235-244.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Using a semantic concordance for sense identification", |
| "authors": [ |
| { |
| "first": "G", |
| "middle": [], |
| "last": "Miller", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Chodorow", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Landes", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Leacock", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Thomas", |
| "suffix": "" |
| } |
| ], |
| "year": 1994, |
| "venue": "Proceedings of the Human Language Technology Workshop", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "G. Miller, M. Chodorow, S. Landes, C. Leacock, and R. Thomas. 1994. Using a semantic concordance for sense identification. In In Proceedings of the Hu- man Language Technology Workshop.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Wordnet: a lexical database for english", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "George", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Miller", |
| "suffix": "" |
| } |
| ], |
| "year": 1995, |
| "venue": "Communications of the ACM", |
| "volume": "38", |
| "issue": "11", |
| "pages": "39--41", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "George A Miller. 1995. Wordnet: a lexical database for english. Communications of the ACM, 38(11):39-41.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "The cpan wordnet:: similarity package", |
| "authors": [ |
| { |
| "first": "Siddharth", |
| "middle": [], |
| "last": "Patwardhan", |
| "suffix": "" |
| }, |
| { |
| "first": "Ted", |
| "middle": [], |
| "last": "Pedersen", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Siddharth Patwardhan and Ted Pedersen. 2003. The cpan wordnet:: similarity package.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "A survey of wordnet annotated corpora", |
| "authors": [ |
| { |
| "first": "Tommaso", |
| "middle": [], |
| "last": "Petrolito", |
| "suffix": "" |
| }, |
| { |
| "first": "Francis", |
| "middle": [], |
| "last": "Bond", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 7th Global WordNet Conference", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tommaso Petrolito and Francis Bond. 2014. A survey of wordnet annotated corpora. In Proceedings of the 7th Global WordNet Conference (GWC 2014).", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "Translation as annotation", |
| "authors": [ |
| { |
| "first": "Emanuele", |
| "middle": [], |
| "last": "Pianta", |
| "suffix": "" |
| }, |
| { |
| "first": "Luisa", |
| "middle": [], |
| "last": "Bentivogli", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Proceedings of the AI* IA 2003 Workshop \"Topics and Perspectives of Natural Language Processing in Italy", |
| "volume": "", |
| "issue": "", |
| "pages": "40--48", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Emanuele Pianta and Luisa Bentivogli. 2003. Trans- lation as annotation. In Proceedings of the AI* IA 2003 Workshop \"Topics and Perspectives of Natu- ral Language Processing in Italy, pages 40-48. Cite- seer.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "Multiwordnet: developing an aligned multilingual database", |
| "authors": [ |
| { |
| "first": "Emanuele", |
| "middle": [], |
| "last": "Pianta", |
| "suffix": "" |
| }, |
| { |
| "first": "Luisa", |
| "middle": [], |
| "last": "Bentivogli", |
| "suffix": "" |
| }, |
| { |
| "first": "Christian", |
| "middle": [], |
| "last": "Girardi", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "Proceedings of the first international conference on global WordNet", |
| "volume": "152", |
| "issue": "", |
| "pages": "55--63", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Emanuele Pianta, Luisa Bentivogli, and Christian Gi- rardi. 2002. Multiwordnet: developing an aligned multilingual database. In Proceedings of the first in- ternational conference on global WordNet, volume 152, pages 55-63.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "Browsing multilingual information with the multisemcor web interface", |
| "authors": [ |
| { |
| "first": "Marcello", |
| "middle": [], |
| "last": "Ranieri", |
| "suffix": "" |
| }, |
| { |
| "first": "Emanuele", |
| "middle": [], |
| "last": "Pianta", |
| "suffix": "" |
| }, |
| { |
| "first": "Luisa", |
| "middle": [], |
| "last": "Bentivogli", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proceedings of the LREC 2004 Satellite Workshop on The Amazing Utility of Parallel and Comparable Corpora", |
| "volume": "", |
| "issue": "", |
| "pages": "38--41", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Marcello Ranieri, Emanuele Pianta, and Luisa Ben- tivogli. 2004. Browsing multilingual information with the multisemcor web interface. In Proceed- ings of the LREC 2004 Satellite Workshop on The Amazing Utility of Parallel and Comparable Cor- pora, pages 38-41. Citeseer.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "Software Framework for Topic Modelling with Large Corpora", |
| "authors": [ |
| { |
| "first": "Petr", |
| "middle": [], |
| "last": "Radim\u0159eh\u016f\u0159ek", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Sojka", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the LREC 2010 Workshop on New Challenges for NLP Frameworks", |
| "volume": "", |
| "issue": "", |
| "pages": "45--50", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Radim\u0158eh\u016f\u0159ek and Petr Sojka. 2010. Software Framework for Topic Modelling with Large Cor- pora. In Proceedings of the LREC 2010 Workshop on New Challenges for NLP Frameworks, pages 45- 50, Valletta, Malta, May. ELRA. http://is. muni.cz/publication/884893/en.", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "Italwordnet: a large semantic database for italian", |
| "authors": [ |
| { |
| "first": "Adriana", |
| "middle": [], |
| "last": "Roventini", |
| "suffix": "" |
| }, |
| { |
| "first": "Antonietta", |
| "middle": [], |
| "last": "Alonge", |
| "suffix": "" |
| }, |
| { |
| "first": "Nicoletta", |
| "middle": [], |
| "last": "Calzolari", |
| "suffix": "" |
| }, |
| { |
| "first": "Bernardo", |
| "middle": [], |
| "last": "Magnini", |
| "suffix": "" |
| }, |
| { |
| "first": "Francesca", |
| "middle": [], |
| "last": "Bertagna", |
| "suffix": "" |
| } |
| ], |
| "year": 2000, |
| "venue": "LREC", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Adriana Roventini, Antonietta Alonge, Nicoletta Cal- zolari, Bernardo Magnini, and Francesca Bertagna. 2000. Italwordnet: a large semantic database for italian. In LREC.", |
| "links": null |
| }, |
| "BIBREF26": { |
| "ref_id": "b26", |
| "title": "Building and annotating the linguistically diverse ntu-mc (ntumultilingual corpus)", |
| "authors": [ |
| { |
| "first": "Liling", |
| "middle": [], |
| "last": "Tan", |
| "suffix": "" |
| }, |
| { |
| "first": "Francis", |
| "middle": [], |
| "last": "Bond", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "PACLIC", |
| "volume": "", |
| "issue": "", |
| "pages": "362--371", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Liling Tan and Francis Bond. 2011. Building and annotating the linguistically diverse ntu-mc (ntu- multilingual corpus). In PACLIC, pages 362-371. Citeseer.", |
| "links": null |
| }, |
| "BIBREF27": { |
| "ref_id": "b27", |
| "title": "From frequency to meaning: Vector space models of semantics", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Peter", |
| "suffix": "" |
| }, |
| { |
| "first": "Patrick", |
| "middle": [], |
| "last": "Turney", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Pantel", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Journal of artificial intelligence research", |
| "volume": "37", |
| "issue": "1", |
| "pages": "141--188", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Peter D Turney, Patrick Pantel, et al. 2010. From frequency to meaning: Vector space models of se- mantics. Journal of artificial intelligence research, 37(1):141-188.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "TABREF0": { |
| "content": "<table><tr><td colspan=\"3\">: Synset gloss comparison (EN:dog,domestic_dog,Canis_familiaris/1 -RO:c\u00e2ine/1)</td></tr><tr><td>Synset</td><td>Lang</td><td>Gloss</td></tr><tr><td>home,nursing_home,</td><td>EN</td><td>an institution where people are cared for;</td></tr><tr><td>rest_home</td><td/><td/></tr><tr><td>azil</td><td>RO</td><td>Institu\u0163ie de asisten\u0163\u0203 social\u0203 pentru \u00eentre\u0163inerea b\u0203tr\u00e2nilor,</td></tr><tr><td/><td/><td>infirmilor, copiilor orfani etc.</td></tr></table>", |
| "text": "", |
| "type_str": "table", |
| "html": null, |
| "num": null |
| }, |
| "TABREF1": { |
| "content": "<table><tr><td>: Synset gloss comparison (EN:home,nursing_home,rest_home -RO:azil)</td></tr><tr><td>Many LESK-inspired algorithms have been pre-</td></tr><tr><td>sented; see for instance</td></tr></table>", |
| "text": "", |
| "type_str": "table", |
| "html": null, |
| "num": null |
| } |
| } |
| } |
| } |