| { |
| "paper_id": "2020", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T12:14:27.107151Z" |
| }, |
| "title": "Character Alignment in Morphologically Complex Translation Sets for Related Languages", |
| "authors": [ |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Gasser", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Indiana University", |
| "location": {} |
| }, |
| "email": "gasser@indiana.edu" |
| }, |
| { |
| "first": "Nazareth", |
| "middle": [ |
| "Amlesom" |
| ], |
| "last": "Kifle", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "\u00d8stfold University College", |
| "location": {} |
| }, |
| "email": "nazareth.a.kifle@hiof.no" |
| }, |
| { |
| "first": "Binyam", |
| "middle": [], |
| "last": "Ephrem", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Addis Ababa University", |
| "location": {} |
| }, |
| "email": "binephrem@gmail.com" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "For languages with complex morphology, word-to-word translation is a task with various potential applications, for example, in information retrieval, language instruction, and dictionary creation, as well as in machine translation. In this paper, we confine ourselves to the subtask of character alignment for the particular case of families of related languages with very few resources for most or all members. There are many such families; we focus on the subgroup of Semitic languages spoken in Ethiopia and Eritrea. We begin with an adaptation of the familiar alignment algorithms behind statistical machine translation, modifying them as appropriate for our task. We show how character alignment can reveal morphological, phonological, and orthographic correspondences among related languages.", |
| "pdf_parse": { |
| "paper_id": "2020", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "For languages with complex morphology, word-to-word translation is a task with various potential applications, for example, in information retrieval, language instruction, and dictionary creation, as well as in machine translation. In this paper, we confine ourselves to the subtask of character alignment for the particular case of families of related languages with very few resources for most or all members. There are many such families; we focus on the subgroup of Semitic languages spoken in Ethiopia and Eritrea. We begin with an adaptation of the familiar alignment algorithms behind statistical machine translation, modifying them as appropriate for our task. We show how character alignment can reveal morphological, phonological, and orthographic correspondences among related languages.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Languages such as Turkish, Arabic, and Swahili are characterized by a wide variety of productive ways in which morphemes are combined to form words. In these morphologically complex languages (MCLs) -and there are thousands of them-many more specific word forms are possible than in languages such as English and Chinese. This very high type-to-token ratio results in a severe data sparsity problem, leading the field of machine translation (MT) to begin to take morphology seriously once such languages were considered (El-Kahlout et al., 2019) . As far as we know, however, no one has investigated the translation of individual words between MCLs. Why would such a task be worth pursuing? First, in MCLs, words may correspond to entire phrases in languages such as English or Chinese. Thus when translating from, say, Arabic to Turkish, the capacity to translate individual nouns or verbs, segmented into their constituent morphemes, could play a significant role. We suspect this will be especially true for translation between closely related MCLs. In such cases, syntactic differences may be limited, with word order and long-distance dependencies largely preserved, while the within-word differences may remain challenging. Second, focusing on word translation pairs may reveal other potentially useful information, especially for related languages, such as language distance or correspondences between characters or phonemes. Finally, a system that translates complex words can support applications such as crosslinguistic information retrieval, dictionary creation, and language instruction.", |
| "cite_spans": [ |
| { |
| "start": 520, |
| "end": 545, |
| "text": "(El-Kahlout et al., 2019)", |
| "ref_id": "BIBREF7" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this paper, we consider the more general problem of translation among a set of words in related MCLs. As an illustration, see the words in column (b) of Figure1, each a translation of 'I was listening' in one of four closely related Romance languages of the Iberian Peninsula. By examining multiple related languages, we eventually hope to be able to capitalize on inter-relationships among the languages. We are specifically interested in the word translation task for languages with few computational resources.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Our task is related to other work in MT. We could, for example, approach the translation of pairs of words within a set using one of the state-of-the-art neural sequence-to-sequence methods, which have also been shown to be applicable to other morphological tasks (Cotterell et al., 2018; McCarthy et al., 2019) . However, we may be just as interested in the details of what is learned -where the boundaries between morphemes are, what morphophonological rules apply -as we are in the performance itself, and it may be difficult or impossible to extract this knowledge from a neural system. For now, we will look for inspiration in earlier MT work, specifically variants of statistical MT (SMT) and leave comparison with neural approaches to future work.", |
| "cite_spans": [ |
| { |
| "start": 264, |
| "end": 288, |
| "text": "(Cotterell et al., 2018;", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 289, |
| "end": 311, |
| "text": "McCarthy et al., 2019)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background", |
| "sec_num": "1" |
| }, |
| { |
| "text": "A further related area of work, which actually resembles our task more closely, concerns morphological paradigms (Erdmann et al., 2020; Jin et al., 2020; Cotterell et al., 2018; McCarthy et al., 2019) . A partial example of a Spanish paradigm appears in column (a) of Figure 1. Of particular interest to us is the \"paradigm cell filling\" problem, that of learning the inflections that comprise a paradigm given all forms for a set of lemmas (Durrett and DeNero, 2013 ).", |
| "cite_spans": [ |
| { |
| "start": 113, |
| "end": 135, |
| "text": "(Erdmann et al., 2020;", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 136, |
| "end": 153, |
| "text": "Jin et al., 2020;", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 154, |
| "end": 177, |
| "text": "Cotterell et al., 2018;", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 178, |
| "end": 200, |
| "text": "McCarthy et al., 2019)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 441, |
| "end": 466, |
| "text": "(Durrett and DeNero, 2013", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background", |
| "sec_num": "1" |
| }, |
| { |
| "text": "If we compare word translations in a set of related languages with cells in a morphological paradigm within one language, the relationship between our task and the paradigm cell filling problem becomes clear. However, this is only the case for the translation of words with similar roots. Figure 1 illustrates this point with an example paradigm from Spanish and two example translation sets from four Romance languages. 1 When the roots are related, as in the translation set in column (b), the set resembles a paradigm; when they aren't (or aren't obviously), as in the translation set in column (c), only the affixes resemble one another. We will be focusing on the case where roots are similar; thus we will need to address the question of how to distinguish such sets from those like the one in column (c). Our longer-term goal is translation: given a word in one language, generate a plausible word in one or more of the other related training languages. In this paper, however, we focus on the subtask that Durrett and Denero (2013) begin with in their paradigm cell filling research, that of aligning the characters of the words in a translation set.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 289, |
| "end": 297, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Background", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The rest of the paper is organized as follows. In Section 2, after an introduction to the Ethio-Eritrean Semitic (EES) languages, we define our task in more detail, looking specifically at possible types of morphological relationships between related languages, with focus on the EES languages. Next, in Section 3, we discuss how some cross-linguistic morphological relationships can be captured in terms of character alignment and show how character alignment within word translation pairs in closely related languages can be be constrained. Next, in Section 4, we outline our approach to character alignment within word translation sets, in particular, how we adapt SMT word alignment methods to handle character-level alignment. Then, in Section 5, we describe our experiments applying the method to data from the Semitic languages. In Section 5.4, we discuss results, both in terms of alignment performance and in terms of useful information that the alignments provide. Finally, in Section 6, we conclude, looking forward to the next steps in the long-term project of the translation of morphologically complex words.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The languages of the world differ greatly with respect to how much morphological elaboration they make available to speakers (Stump, 2017) . What is interesting for our purposes (and those of this workshop) is the apparent tendency for languages within a closely related family to agree in morphological complexity, though not necessarily in the morphological details.", |
| "cite_spans": [ |
| { |
| "start": 125, |
| "end": 138, |
| "text": "(Stump, 2017)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Semitic languages of Ethiopia and Eritrea", |
| "sec_num": "2" |
| }, |
| { |
| "text": "In this paper we focus on one of these families, the Ethio-Eritrean Semitic (EES) languages, consisting of the subgroup of Semitic languages spoken in Ethiopia and Eritrea (Demeke, 2001 ). There is no general agreement on the number of EES languages, but, judged by mutual intelligibility, there are between 10 and 15. Of these, two, Tigrinya (hereafter, Ti) and Tigre (hereafter, Te), are spoken in Eritrea, by roughly 85% of the population, and the remainder, in addition to Ti, are spoken in Ethiopia, as native languages of roughly 38% of the population. These include Amharic (hereafter Am), the official language of the Federal Government of Ethiopia. The EES languages are divided into northern and southern subgroups, with Ge'ez, Ti, and Te in the northern subgroup, Am and the remaining languages in the southern subgroup.", |
| "cite_spans": [ |
| { |
| "start": 151, |
| "end": 185, |
| "text": "Ethiopia and Eritrea (Demeke, 2001", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Semitic languages of Ethiopia and Eritrea", |
| "sec_num": "2" |
| }, |
| { |
| "text": "The languages exhibit many features common to other Semitic languages, especially very complex verbal morphology, including subject agreement suffixes and prefixes, object agreement suffixes, and the templatic morphology that Semitic languages are famous for. Verb roots consist of sequences of consonants, organized in a number of root classes. Verb stems consist of patterns of particular vowel and consonant gemination that are overlaid on the consonantal roots. For an Am, Ti, and Te example, see Figure 3a .", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 501, |
| "end": 510, |
| "text": "Figure 3a", |
| "ref_id": "FIGREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Morphology", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "The EES languages are written using the Ge'ez writing system, an abugida system in which each character represents either a consonant-vowel sequence (the onset and nucleus of a syllable) or a single consonant (the coda of a syllable). The characters are usually presented in a table, in which each row contains the characters for a single consonant. Figure 2 shows three rows from this table, the symbols representing the consonants /l/, /b/, and /g/ followed by vowels or alone. In addition to their realization as a single consonant, the characters in the sixth column can also represent the consonant followed by the epenthetic vowel [1]. 2 For example, the Am word b@l1bb 'by heart' is written \u1260\u120d\u1265. Of the two sixth-column characters, the first, \u120d, is pronounced with a following vowel, while the second, \u1265, is not. This example also illustrates the only significant deviation of the orthography from the phonological form of a word: consonant gemination (length) is not indicated. In our experiments we work with both the orthographic representations of words and a phonemic transcription in which gemination is indicated (marked with \":\"). Because the orthography obscures the boundaries between consonants and vowels, it may also obscure the boundaries between morphemes. An example is shown in Figure 3a . Circles surround affixes; rectangles surround stems. In multiple cases, morpheme boundaries are between consonants and vowels, so a single Ge'ez character spans two morphemes. For example, the final character in each word, \u1260 /b@/, includes the last stem consonant /b/ and the perfective 3sm suffix /@/. We include gemination in our phonemic representation because of its role in EES verb morphology as well as its importance for speech synthesis in the languages.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 350, |
| "end": 358, |
| "text": "Figure 2", |
| "ref_id": "FIGREF1" |
| }, |
| { |
| "start": 1302, |
| "end": 1311, |
| "text": "Figure 3a", |
| "ref_id": "FIGREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Orthography", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Like most African languages, EES languages suffer from a lack of resources. With the exception of Am and Ti (Abate et al., 2018), corpora in the languages are very limited, in some cases non-existent. However, we do have the advantage of bilingual root lexica (Am-Ti, Ti-Te) and dedicated morphological analyzers and generators for three of the languages, Am, Ti, and Te, within the HornMorpho project (Gasser, 2011) . 3 HornMorpho consists of weighted finite-state transducers constructed by hand. Using HornMorpho and our bilingual root lexica, we can generate data for the morphological translation task. ", |
| "cite_spans": [ |
| { |
| "start": 402, |
| "end": 416, |
| "text": "(Gasser, 2011)", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 419, |
| "end": 420, |
| "text": "3", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Previous work", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "It is beyond the scope of this paper to discuss morphological complexity (Stump, 2017) . We base our consideration of morphology in the remainder of the paper on the structure of verbs in EES languages and on general notions familiar from basic morphological research. We assume that for each word in a MCL, there is a stem, based on a root; zero or more affixes; and possibly instances of reduplication. For each of these elements in a word, there may be a correspondence within the word's translation in another language. There may be similarities in the order of the morphemes, their forms, and the morphophonological processes that take place at their boundaries. Some of these correspondences are illustrated in Figure 3a for an Am-Ti-Te translation set. The arrows indicate corresponding morphemes, with the phonetic relationships relatively clear, except in the case of the first prefix. Within the stem, the root consonants appear in boldface. Though this may not be apparent, the Am and Ti roots in this case are systematically related: the Ti consonant /\u00e8/ corresponds regularly to zero in Am, and the Ti consonant /s'/ corresponds often (though not always) to /t'/.", |
| "cite_spans": [ |
| { |
| "start": 73, |
| "end": 86, |
| "text": "(Stump, 2017)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 717, |
| "end": 726, |
| "text": "Figure 3a", |
| "ref_id": "FIGREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Morphological alignment and character alignment", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Clearly translation between words such as those in Figure 3a can benefit from recognizing the relationships indicated by the arrows in the figure. Alignment is the process of determining such relationships.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 51, |
| "end": 60, |
| "text": "Figure 3a", |
| "ref_id": "FIGREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Morphological alignment and character alignment", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Alignment-identifying corresponding elements in source and target-is fundamental to the training of MT systems; the progress in neural MT resulting from the inclusion of attention mechanisms is just the most recent example (Bahdanau et al., 2015; Vaswani et al., 2017) . Within SMT, alignment is a separate step, and because everything else depends on the quality of the alignment, a great deal of emphasis is given to alignment mechanisms (Och and Ney, 2003) .", |
| "cite_spans": [ |
| { |
| "start": 223, |
| "end": 246, |
| "text": "(Bahdanau et al., 2015;", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 247, |
| "end": 268, |
| "text": "Vaswani et al., 2017)", |
| "ref_id": null |
| }, |
| { |
| "start": 440, |
| "end": 459, |
| "text": "(Och and Ney, 2003)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Morphological alignment and character alignment", |
| "sec_num": "3" |
| }, |
| { |
| "text": "How is character alignment constrained for word translation pairs in closely related languages? With rare exceptions (see below), we can expect it to be monotonic. That is, for word translation pair s and t, if the character in position i in s is associated with the character in position j in t, then the character in position i + 1 in s is either (1) not associated with any character in t or (2) associated with a character in position k in t such that k >= j. This is illustrated for the Am-Ti pair in Figure 3b . Informally, monotonicity means that none of the association arrows (as in the upper diagram in Figure 3b ) cross one another. Given this constraint, we can represent character alignments in the simpler format shown in the lower diagram in the figure, with aligned characters above one another.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 506, |
| "end": 515, |
| "text": "Figure 3b", |
| "ref_id": "FIGREF3" |
| }, |
| { |
| "start": 613, |
| "end": 622, |
| "text": "Figure 3b", |
| "ref_id": "FIGREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Morphological alignment and character alignment", |
| "sec_num": "3" |
| }, |
| { |
| "text": "What would it mean for the monotonicity constraint to be violated for word pairs in related languages? As far as we can tell, the only situation where this could occur is one in which one of the languages has undergone a diachronic or synchronic metathesis process, by which a pair of characters has swapped their positions. For example, imagine a word gabla in language 1, which translates as the word galba in language 2. While we are not aware of any such cases, they are certainly possible, but in this paper we restrict ourselves to the computationally simpler, and obviously much more common, monotonic cases.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Morphological alignment and character alignment", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Morphemes can of course consist of multiple characters, and a morpheme in one language can correspond to a morpheme of a different length in another; there are several examples in Figure 3a . However, as in Durrett and DeNero's (2013) approach to paradigm cell filling, we begin with the alignment of single characters, constraining the alignment to be one-to-one in both directions. As can be seen from the alignment shown in Figure 3b , these one-to-one character alignments usually do not represent morpheme correspondences, but they do represent a step towards the identification of corresponding morphemes.", |
| "cite_spans": [ |
| { |
| "start": 207, |
| "end": 234, |
| "text": "Durrett and DeNero's (2013)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 180, |
| "end": 189, |
| "text": "Figure 3a", |
| "ref_id": "FIGREF3" |
| }, |
| { |
| "start": 427, |
| "end": 436, |
| "text": "Figure 3b", |
| "ref_id": "FIGREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Morphological alignment and character alignment", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Following character alignment, Durrett and DeNero merge neighboring aligned characters into candidate morphemes. This step will be the topic of the next phase of the project and a future paper.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Morphological alignment and character alignment", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Durrett and DeNero (2013) realize the alignment step in their paradigm cell filling system as an iterated edit-distance algorithm. However, the word translation problem does not lend itself to this approach because of the number of associated characters that are not identical. We applied Durrett and DeNero's algorithm to our task but were only able to achieve reasonable results by modifying it to include explicit character categories, such as consonants and vowels, affecting the probabilities of the edit operations. For example, we treated the substitution of a vowel for a vowel as more likely than the substitution of a vowel for a consonant. In general, we cannot expect to have this kind of knowledge of character categories for arbitrary orthographic systems, so it is desirable to develop a less knowledge-intensive approach to alignment. Since the early SMT work makes no assumptions at all about the categories of source and target words, we looked to this work for inspiration in designing an approach to our task that makes no assumptions about the categories of source and target characters.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Approach", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Nearly all SMT research relies on the original insights from Brown et al. (1993) . Alignment is implemented through a set of parameters whose values are determined by maximizing the likelihood of a training corpus of sentence pairs. The maximization is performed using the expectation maximization (EM) algorithm (Dempster et al., 1977) . In the expectation (E) step of each iteration, a set of \"expected counts\" of co-occurring source and target words within the corpus is calculated, weighted by the current values of the parameters. In the maximization (M) step of each iteration, new values of the parameters are calculated on the basis of the expected counts from the E-step.", |
| "cite_spans": [ |
| { |
| "start": 61, |
| "end": 80, |
| "text": "Brown et al. (1993)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 313, |
| "end": 336, |
| "text": "(Dempster et al., 1977)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "IBM word alignment models", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "In the simplest SMT alignment model (IBM Model 1), the only parameters involved are word-word translation probabilities. The goodness of an alignment of the words in a pair of sentences depends only on the probabilities that the aligned words are translations of one another, not on position in the sentences. There is the additional assumption that each word in the source sentence can be aligned with at most one word in the target sentence. For our word translation task, we start with IBM Model 1, with words replaced by characters. That is, for each pair of source and target language characters, there is a character association probability, representing the probability of the source character given the target character. Each source character also has a probability of being \"deleted,\" that is, of no association at all in the target word. However, we start with a bit more knowledge here than in the case of sentence translation. Specifically characters that are identical should have relatively high association probabilities. Thus we initialize the probabilities with a high value for identical characters (a meta-parameter that we set to 0.5) and equal low probabilities for all other character pairs. As with IBM Model 1 for sentence translation, we make the assumption that each source word character is associated with at most one target word character. In addition, we assume the inverse constraint as well: that each target word character is associated with at most one source word character. During the E-step of this simple version of our model, for each word pair in the corpus, for each combination of source and target language character, we count the ways they can be aligned (if they occur at all in the words), weighted by their current association probabilities. During the M-step we recalculate the probabilities based on the counts. The E-and M-steps are repeated until the parameters converge within some criterion.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Character association probabilities", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "IBM Model 1 is clearly inadequate for sentence translation because a target language word is equally likely to be aligned with a source language word wherever it appears in the sentence; the order of the words is irrelevant. Succeeding models in the IBM series include additional parameters to handle position, as well as relative sentence length.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Position and length effects", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "We also add additional parameters to handle position, but because our task is more constrained than the sentence translation task, we diverge here from the IBM models. Specifically, given our monotonicity constraint, we expect a character at a particular relative position within the source word to be associated with a character at a similar relative position with the target word. For example, in the pair of words in Figure 3b , even though the Am l and the Ti y are not equivalent, they should have a relatively high probability of association because of their positions within the word.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 420, |
| "end": 429, |
| "text": "Figure 3b", |
| "ref_id": "FIGREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Position and length effects", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "To operationalize this notion, we view an alignment as either \"left-justified\" or \"right-justified,\" as proceeding from one or the other end of the word. Thus rather than relative distance within the words, we are actually comparing their absolute distances from one end of the word or the other. We include a set of parameters representing the probability of a left-justified alignment given the difference in the lengths of the two words. That is, for each of the possible differences in word length (there are about 20 of these in a typical dataset for our languages), the system is trained on a left justification probability (the right justification probability is just one minus this value).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Position and length effects", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "During the E-step in this modified version of the model, the expected counts for each word pair depend not only on the current association probabilities for each character pair but also on the positions of the associated characters and on the left justification probabilities. That is, separate counts are calculated starting at the left and right ends of the word pairs, each weighted by the justification parameters. In addition, expected counts are also calculated for left justified and right justified alignments of each word pair, to use in updating the left justification parameters during the M-step.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Position and length effects", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Finally, the number of characters in a source word that are associated with characters in the target word depends not only on the difference in the lengths of the words but also on the number of target characters that are \"deleted,\" that is, not associated with any source language characters. For example, consider the example in Figure 3b . The source word has nine characters, the target word ten characters. The number of deleted characters in the source word not only depends on the length difference (\u22121) but also on the likelihood that target characters are deleted. In particular the Ti phoneme /\u00e8/ is nearly always deleted. Thus we include a separate set of target character deletion probabilities, which are also trained during EM.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 331, |
| "end": 340, |
| "text": "Figure 3b", |
| "ref_id": "FIGREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Position and length effects", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "In summary, we define three sets of parameters governing alignments over word translation pairs and learn the values of the parameters that maximize the likelihood of the training set: (1) character association parameters, (2) left justification parameters, and (3) target character deletion parameters. We train the system using the EM algorithm. During the E-step, expected counts are calculated for each of the parameter types over the whole training set using the current parameter values as weights. During the M-step, new values for each of the parameters are calculated on the basis of the counts from the Estep. Following training, we align a given pair of words by searching for the alignment with the highest probability given the parameter values.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Position and length effects", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Where would the sets of word translations for training the parameters come from? In the most general case, we are interested in languages with little or no available data, but for related languages there is often a community of multilinguals, especially when one of the languages is a lingua franca or an official regional or national language. Thus one source of data would be elicitation of word translations from multilinguals. In other cases, we might benefit from limited corpora. For the purposes of this paper, however, we wish to show only that the framework is viable, so we work with a set of languages for which we have the morphological and lexical resources necessary to generate the data automatically. Below, we describe the generation of the datasets in detail.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Possible data sources", |
| "sec_num": "4.4" |
| }, |
| { |
| "text": "Applying the method to a language with few or no resources at all, through a combination of elicitation of translated words from multilinguals or word alignment in limited corpora, will wait for future work.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Possible data sources", |
| "sec_num": "4.4" |
| }, |
| { |
| "text": "The goal of the experiments reported on below is to investigate whether our EM alignment algorithm (1) results in character alignments that reflect the correspondences between affixes and root consonants and (2) yields parameters that embody the known phonetic relationships between the languages.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments", |
| "sec_num": "5" |
| }, |
| { |
| "text": "We generate the data for the experiments automatically using our multilingual root lexica and the morphological generators available in HornMorpho (Gasser, 2011) . Because our initial assumption is that alignment will succeed only for words with similar roots, using our root lexica for the languages, we first extract a set of verbs in the three languages with the same or similar meanings that have the same or similar roots. 4 The result is a set of 42 roots for each of the three languages. 5 For each of the roots, we generate a set of word forms, yielding all combinations of a set of morphological properties that are common to the three languages. These properties are shown in Table 1 . This results in 1488 word translation sets. We reserve 298 of these for validation and testing, yielding a training set of 1190 translation sets. Each translation set appears in an orthographic form and a phonemic form. We also isolated a separate set of 42 dissimilar root pairs for Am and Ti (see below for the motivation), for example, the pair Am <\u010d'fr> Ti <sQsQ> 'dance'. For each of these roots, we generated word forms using the same set of morphological properties. The result is a separate dissimilar training set consisting of 1488 word translation pairs and a \"mixed\" training set of 2,380 pairs consisting of both similar and dissimilar pairs.", |
| "cite_spans": [ |
| { |
| "start": 147, |
| "end": 161, |
| "text": "(Gasser, 2011)", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 428, |
| "end": 429, |
| "text": "4", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 686, |
| "end": 693, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Training and test data", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "The number of distinct characters in the orthographic data sets is as follows: Am-72, Ti-76, Te-57. The phonemic data sets include separate characters for the geminated and labialized forms of consonants; for example, k, k:, k w , and k w : all appear in the Am data as \"characters\". The number of distinct characters in the phonemic data sets is as follows: Am-47, Ti-40, Te-39.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Training and test data", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "For all conditions, we trained eight iterations of the EM algorithm. Following training, we aligned the test items on the basis of the learned parameters and evaluated the alignments on the basis of the constraints described below.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Training", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "We first trained the Am-to-Ti and Ti-to-Te translation pairs, separately for the orthographic and phonemic training sets. Next, we trained each of these translation pairs in the opposite direction. Then, inspired by symmetrization techniques applied to alignment in SMT (Och and Ney, 2003) , we aligned the test set on the basis of the union and the intersection of alignments based on the parameters learned in the two directions. In the union case, characters were considered aligned if they would have been aligned in either of the two directions. In the intersection case, characters were considered aligned if they would have been aligned in both of the two directions.", |
| "cite_spans": [ |
| { |
| "start": 266, |
| "end": 289, |
| "text": "SMT (Och and Ney, 2003)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Training", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "Finally, we attempted to address the artificiality of our training set by training on the \"mixed\" training set described above. For the mixed condition, the test set was the same one used for the similar roots condition.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Training", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "We annotated each phonemic and orthographic test set of 149 word pairs by hand, based on our knowledge of the morphology of the languages; see the upper portion of Figure 4 . For the phonemic translation pairs, we indicate specific character-to-character alignments for root consonants, where these are present in both words, and for affixes, when there are clear correspondences between the characters. In cases where corresponding affixes do not clearly map onto one another, we constrain the characters in the two affixes to overlap. That is, each character within the shorter affix should align with some character in the longer affix. For the orthographic translation pairs, we indicate corresponding root characters (sharing a consonant but not necessarily a vowel), where these are present in both words. For affixes, annotation is as for the phonemic pairs. There are no constraints associated with affixes in one language that correspond to nothing in the other language. The upper part of Figure 4 shows an example, the Am-Ti translation pair from the previous figures, translating 'which was not washed'. Root consonant correspondences are indicated by arrows, affix correspondences by boxes. There are seven alignment constraints in the phonemic pair, five constraints in the orthographic pair. We evaluated the performance of the system for both recall, the proportion of alignment constraints that were satisfied, and precision, the proportion of aligned characters that satisfied the constraints. The lower part of Figure 4 shows two alignments for the phonemic and two for the orthographic pair in the upper part of the figure. Correct alignments are indicated with solid lines, incorrect alignments with dashed lines. The upper alignments are like those we would expect early in training; only identical characters are aligned. (Recall that the character association probabilities for identical characters are initialized with much higher values than those for other character pairs.) Note that precision is high in this case: 0.75 for the phonemic pair, 1.0 for the orthographic pair. Because many correct alignments are missed, however, recall is low: 0.43 for the phonemic pair, 0.2 for the orthographic pair. The lower alignments are of a sort we would expect later in training. In both the phonemic and orthographic pairs, the system is confused because of the similarities (based on the trained parameters) of characters that should not be aligned. Recall is 0.86 for the phonemic example, 0.4 for the orthographic example. Precision is 0.67 for the phonemic example, 0.4 for the orthographic example.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 164, |
| "end": 172, |
| "text": "Figure 4", |
| "ref_id": "FIGREF4" |
| }, |
| { |
| "start": 999, |
| "end": 1007, |
| "text": "Figure 4", |
| "ref_id": "FIGREF4" |
| }, |
| { |
| "start": 1530, |
| "end": 1538, |
| "text": "Figure 4", |
| "ref_id": "FIGREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "5.3" |
| }, |
| { |
| "text": "Results of the experiments are shown in Table 2 . From these results, we may notice the following points.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 40, |
| "end": 47, |
| "text": "Table 2", |
| "ref_id": "TABREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results and discussion", |
| "sec_num": "5.4" |
| }, |
| { |
| "text": "First, note the relatively high baseline precision scores, especially for the orthographic data. This is not surprising, given alignments based almost entirely on character identity (see Figure 4) .", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 187, |
| "end": 196, |
| "text": "Figure 4)", |
| "ref_id": "FIGREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results and discussion", |
| "sec_num": "5.4" |
| }, |
| { |
| "text": "Second, as we might expect, we see somewhat higher recall and precision for the more closely related languages, Ti and Te, than for Am and Ti. This suggests that the alignment mechanism might function as a measure of morphological relatedness between languages or dialects.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results and discussion", |
| "sec_num": "5.4" |
| }, |
| { |
| "text": "Third, clearly both recall and precision improve a great deal with training; this can be also be seen in the uniformly increasing F-scores. Since the system starts with no knowledge of character similarity, this is significant in and of itself. Recall improves in all four cases when alignments are formed from the union of the alignments in the two directions (\"L1 \u222a \u2194 L2\" in the table). Somewhat surprisingly, we do not see the same sort of improvement with precision, which we would expect to benefit from the intersection of the alignments (\"L1 \u2229 \u2194 L2\" in the table). Fourth, with dissimilar roots included along with similar roots in training (\"mixed\" in the table, reported only for the union of the alignments and for Am and Ti translation pairs), recall is not as high as with only similar roots, but there is still significant improvement over the baseline. Of the expected correspondences between root consonants and affix segments, 0.894 are correctly aligned. Even with as many dissimilar as similar roots, the alignment algorithm is not confused.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results and discussion", |
| "sec_num": "5.4" |
| }, |
| { |
| "text": "To further investigate the value of what has been learned, we looked at the trained character association probabilities, which should reflect the morphological and phonological relationships between characters in the two languages. Focusing on the Am-Ti case, we first isolated those character pairs with character association probabilities of 0.01 or greater. We then took the intersection of the set of such characters found in the two translation directions. We performed this analysis for both the phonemic and orthographic conditions. Not surprisingly, Am characters tend to be associated with the identical Ti characters. Other intriguing relationships are also found. Because gemination operates quite differently in the two languages, ten of the consonants in the phonemic conditions are associated with their geminated variants (recall that these take the form of different characters in the data). In addition, in both the phonemic and orthographic conditions, the Am phoneme /t'/ is clearly associated with the Ti phoneme /s'/, a frequent and familiar correspondence between the languages. Similarly, the Am phonemes k and k' correspond, when ungeminated, to fricative versions of these phonemes in Ti, and again these correspondences are found, in both the phonemic and orthographic conditions. Finally, in the orthographic condition, 27 of the characters are associated with other characters in the same row in the Ge'ez character table, that is, with characters representing the same consonant followed by different vowels (see Figure 2 ).", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 1542, |
| "end": 1550, |
| "text": "Figure 2", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results and discussion", |
| "sec_num": "5.4" |
| }, |
| { |
| "text": "The main contributions of this paper are (1) the introduction of a new task, that of translation of sets of words in morphologically complex, closely related languages and (2) the demonstration that a simple alignment algorithm, based on work within SMT, is a promising first step in approaching this task. In particular, we have shown that the parameters that are learned during EM training support alignments that reveal (1) correspondences between affixes and root consonants in pairs of EES languages and (2) relationships between orthographic or phonetic segments that we know to be valid.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "6" |
| }, |
| { |
| "text": "Of course we have not yet shown to what extent alignment actually supports translation of morphologically complex words. To that end, our next step will be to proceed roughly as in the work of Durrett and DeNero (2013) , extracting a set of transformation rules based on the alignments and then using a discriminative sequence model to learn the contexts for the rules.", |
| "cite_spans": [ |
| { |
| "start": 193, |
| "end": 218, |
| "text": "Durrett and DeNero (2013)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "6" |
| }, |
| { |
| "text": "Without the benefit of context, there is a much greater potential for translation ambiguity for words than for sentences. Here and elsewhere in this paper, we assume that \"translation of a word\" refers to a translation in some possible context.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "Linguists disagree on whether this vowel represents a phoneme, and its status may vary between the languages. 3 Te has been added to HornMorpho for the purpose of this project.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "Phonetic \"similarity\" here takes into account known relationships among the languages, for example, the fact that the root consonants h, \u00e8, P, and Q in Ti and Te correspond to zero in Am roots. Of course these are relationships that must be discovered by the alignment mechanism.5 More precisely, each \"root\" is a combination of a consonantal root, its root category, and one of the set of derivational categories that characterize EES verbs, for example, passive-reflexive or transitive. An example is the Ti \"root\" <lPk>:A:ps 'be sent', consisting of root consonant sequence lPk in root class A and in the passive-reflexive derivational category.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Parallel corpora for bi-directional statistical machine translation for seven Ethiopian language pairs", |
| "authors": [ |
| { |
| "first": "Michael", |
| "middle": [ |
| "Melese" |
| ], |
| "last": "Solomon Teferra Abate", |
| "suffix": "" |
| }, |
| { |
| "first": "Martha", |
| "middle": [ |
| "Yifiru" |
| ], |
| "last": "Woldeyohannis", |
| "suffix": "" |
| }, |
| { |
| "first": "Million", |
| "middle": [], |
| "last": "Tachbelie", |
| "suffix": "" |
| }, |
| { |
| "first": "Solomon", |
| "middle": [], |
| "last": "Meshesha", |
| "suffix": "" |
| }, |
| { |
| "first": "Wondwossen", |
| "middle": [], |
| "last": "Atinafu", |
| "suffix": "" |
| }, |
| { |
| "first": "Yaregal", |
| "middle": [], |
| "last": "Mulugeta", |
| "suffix": "" |
| }, |
| { |
| "first": "Hafte", |
| "middle": [], |
| "last": "Assabie", |
| "suffix": "" |
| }, |
| { |
| "first": "Biniyam", |
| "middle": [], |
| "last": "Abera", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Ephrem", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Tewodros Abebe, Wondimagegnehu Tsegaye, Amanuel Lemma, Tsegaye Andargie, and Seifedin Shifaw", |
| "volume": "", |
| "issue": "", |
| "pages": "83--90", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Solomon Teferra Abate, Michael Melese Woldeyohannis, Martha Yifiru Tachbelie, Million Meshesha, Solomon Atinafu, Wondwossen Mulugeta, Yaregal Assabie, Hafte Abera, Biniyam Ephrem, Tewodros Abebe, Wondim- agegnehu Tsegaye, Amanuel Lemma, Tsegaye Andargie, and Seifedin Shifaw. 2018. Parallel corpora for bi-directional statistical machine translation for seven Ethiopian language pairs. In Proceedings of the First Workshop on Linguistic Resources for Natural Language Processing, pages 83-90, Santa Fe, NM, USA.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Neural machine translation by jointly learning to align and translate", |
| "authors": [ |
| { |
| "first": "Dzmitry", |
| "middle": [], |
| "last": "Bahdanau", |
| "suffix": "" |
| }, |
| { |
| "first": "Kyunghyun", |
| "middle": [], |
| "last": "Cho", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "3rd International Conference on Learning Representations", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural machine translation by jointly learning to align and translate. In Yoshua Bengio and Yann LeCun, editors, 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "The mathematics of statistical machine translation: Parameter estimation", |
| "authors": [ |
| { |
| "first": "F", |
| "middle": [], |
| "last": "Peter", |
| "suffix": "" |
| }, |
| { |
| "first": "Stephen", |
| "middle": [ |
| "A Della" |
| ], |
| "last": "Brown", |
| "suffix": "" |
| }, |
| { |
| "first": "Vincent", |
| "middle": [ |
| "J" |
| ], |
| "last": "Pietra", |
| "suffix": "" |
| }, |
| { |
| "first": "Robert", |
| "middle": [ |
| "L" |
| ], |
| "last": "Della Pietra", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Mercer", |
| "suffix": "" |
| } |
| ], |
| "year": 1993, |
| "venue": "Computational Linguistics", |
| "volume": "19", |
| "issue": "2", |
| "pages": "263--311", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Peter F. Brown, Stephen A. Della Pietra, Vincent J. Della Pietra, and Robert L. Mercer. 1993. The mathematics of statistical machine translation: Parameter estimation. Computational Linguistics, 19(2):263-311.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "The CoNLL-SIGMORPHON 2018 shared task: Universal morphological reinflection", |
| "authors": [ |
| { |
| "first": "Ryan", |
| "middle": [], |
| "last": "Cotterell", |
| "suffix": "" |
| }, |
| { |
| "first": "Christo", |
| "middle": [], |
| "last": "Kirov", |
| "suffix": "" |
| }, |
| { |
| "first": "John", |
| "middle": [], |
| "last": "Sylak-Glassman", |
| "suffix": "" |
| }, |
| { |
| "first": "G\u00e9raldine", |
| "middle": [], |
| "last": "Walther", |
| "suffix": "" |
| }, |
| { |
| "first": "Ekaterina", |
| "middle": [], |
| "last": "Vylomova", |
| "suffix": "" |
| }, |
| { |
| "first": "Arya", |
| "middle": [ |
| "D" |
| ], |
| "last": "Mccarthy", |
| "suffix": "" |
| }, |
| { |
| "first": "Katharina", |
| "middle": [], |
| "last": "Kann", |
| "suffix": "" |
| }, |
| { |
| "first": "Sabrina", |
| "middle": [ |
| "J" |
| ], |
| "last": "Mielke", |
| "suffix": "" |
| }, |
| { |
| "first": "Garrett", |
| "middle": [], |
| "last": "Nicolai", |
| "suffix": "" |
| }, |
| { |
| "first": "Miikka", |
| "middle": [], |
| "last": "Silfverberg", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Yarowsky", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the CoNLL-SIGMORPHON 2018 Shared Task: Universal Morphological Reinflection", |
| "volume": "", |
| "issue": "", |
| "pages": "1--27", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ryan Cotterell, Christo Kirov, John Sylak-Glassman, G\u00e9raldine Walther, Ekaterina Vylomova, Arya D. McCarthy, Katharina Kann, Sabrina J. Mielke, Garrett Nicolai, Miikka Silfverberg, David Yarowsky, Jason Eisner, and Mans Hulden. 2018. The CoNLL-SIGMORPHON 2018 shared task: Universal morphological reinflection. In Proceedings of the CoNLL-SIGMORPHON 2018 Shared Task: Universal Morphological Reinflection, pages 1-27, Brussels, October. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "The Ethio-Semitic languages (re-examining the classification)", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Girma", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Demeke", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Journal of Ethiopian Studies", |
| "volume": "34", |
| "issue": "2", |
| "pages": "57--93", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Girma A. Demeke. 2001. The Ethio-Semitic languages (re-examining the classification). Journal of Ethiopian Studies, 34(2):57-93.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Maximum likelihood from incomplete data via the EM algorithm", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [ |
| "P" |
| ], |
| "last": "Dempster", |
| "suffix": "" |
| }, |
| { |
| "first": "N", |
| "middle": [ |
| "M" |
| ], |
| "last": "Laird", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [ |
| "B" |
| ], |
| "last": "Rubin", |
| "suffix": "" |
| } |
| ], |
| "year": 1977, |
| "venue": "Journal of the Royal Statistical Society. Series B (Methodological)", |
| "volume": "39", |
| "issue": "1", |
| "pages": "1--38", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "A. P. Dempster, N. M. Laird, and D. B. Rubin. 1977. Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society. Series B (Methodological), 39(1):1-38.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Supervised learning of complete morphological paradigms", |
| "authors": [ |
| { |
| "first": "Greg", |
| "middle": [], |
| "last": "Durrett", |
| "suffix": "" |
| }, |
| { |
| "first": "John", |
| "middle": [], |
| "last": "Denero", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "1185--1195", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Greg Durrett and John DeNero. 2013. Supervised learning of complete morphological paradigms. In Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 1185-1195, Atlanta, Georgia, June. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Translating between morphologically rich languages: An Arabic-to-Turkish machine translation system", |
| "authors": [ |
| { |
| "first": "Emre", |
| "middle": [], |
| "last": "Ilknur Durgar El-Kahlout", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Bekta\u015f", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the Fourth Arabic Natural Language Processing Workshop", |
| "volume": "", |
| "issue": "", |
| "pages": "158--166", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ilknur Durgar El-Kahlout, Emre Bekta\u015f, Naime \u015eeyma Erdem, and Hamza Kaya. 2019. Translating between mor- phologically rich languages: An Arabic-to-Turkish machine translation system. In Proceedings of the Fourth Arabic Natural Language Processing Workshop, pages 158-166.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "The paradigm discovery problem", |
| "authors": [ |
| { |
| "first": "Alexander", |
| "middle": [], |
| "last": "Erdmann", |
| "suffix": "" |
| }, |
| { |
| "first": "Micha", |
| "middle": [], |
| "last": "Elsner", |
| "suffix": "" |
| }, |
| { |
| "first": "Shijie", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Ryan", |
| "middle": [], |
| "last": "Cotterell", |
| "suffix": "" |
| }, |
| { |
| "first": "Nizar", |
| "middle": [], |
| "last": "Habash", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "7778--7790", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alexander Erdmann, Micha Elsner, Shijie Wu, Ryan Cotterell, and Nizar Habash. 2020. The paradigm discovery problem. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 7778-7790.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "HornMorpho: a system for morphological processing of Amharic, Oromo, and Tigrinya", |
| "authors": [ |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Gasser", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the Conference on Human Language Technology for Development", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Michael Gasser. 2011. HornMorpho: a system for morphological processing of Amharic, Oromo, and Tigrinya. In Proceedings of the Conference on Human Language Technology for Development, Alexandria, Egypt.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Unsupervised morphological paradigm completion", |
| "authors": [ |
| { |
| "first": "Huimin", |
| "middle": [], |
| "last": "Jin", |
| "suffix": "" |
| }, |
| { |
| "first": "Liwei", |
| "middle": [], |
| "last": "Cai", |
| "suffix": "" |
| }, |
| { |
| "first": "Yihui", |
| "middle": [], |
| "last": "Peng", |
| "suffix": "" |
| }, |
| { |
| "first": "Xia", |
| "middle": [], |
| "last": "", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "6696--6707", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Huimin Jin, Liwei Cai, Yihui Peng, and Xia. 2020. Unsupervised morphological paradigm completion. In Pro- ceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 6696-6707.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "The SIGMORPHON 2019 shared task: Morphological analysis in context and cross-lingual transfer for inflection", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Arya", |
| "suffix": "" |
| }, |
| { |
| "first": "Ekaterina", |
| "middle": [], |
| "last": "Mccarthy", |
| "suffix": "" |
| }, |
| { |
| "first": "Shijie", |
| "middle": [], |
| "last": "Vylomova", |
| "suffix": "" |
| }, |
| { |
| "first": "Chaitanya", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Lawrence", |
| "middle": [], |
| "last": "Malaviya", |
| "suffix": "" |
| }, |
| { |
| "first": "Garrett", |
| "middle": [], |
| "last": "Wolf-Sonkin", |
| "suffix": "" |
| }, |
| { |
| "first": "Christo", |
| "middle": [], |
| "last": "Nicolai", |
| "suffix": "" |
| }, |
| { |
| "first": "Miikka", |
| "middle": [], |
| "last": "Kirov", |
| "suffix": "" |
| }, |
| { |
| "first": "Sabrina", |
| "middle": [ |
| "J" |
| ], |
| "last": "Silfverberg", |
| "suffix": "" |
| }, |
| { |
| "first": "Jeffrey", |
| "middle": [], |
| "last": "Mielke", |
| "suffix": "" |
| }, |
| { |
| "first": "Ryan", |
| "middle": [], |
| "last": "Heinz", |
| "suffix": "" |
| }, |
| { |
| "first": "Mans", |
| "middle": [], |
| "last": "Cotterell", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Hulden", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 16th Workshop on Computational Research in Phonetics, Phonology, and Morphology", |
| "volume": "", |
| "issue": "", |
| "pages": "229--244", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Arya D. McCarthy, Ekaterina Vylomova, Shijie Wu, Chaitanya Malaviya, Lawrence Wolf-Sonkin, Garrett Nico- lai, Christo Kirov, Miikka Silfverberg, Sabrina J. Mielke, Jeffrey Heinz, Ryan Cotterell, and Mans Hulden. 2019. The SIGMORPHON 2019 shared task: Morphological analysis in context and cross-lingual transfer for inflection. In Proceedings of the 16th Workshop on Computational Research in Phonetics, Phonology, and Morphology, pages 229-244, Florence, Italy, August. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "A systematic comparison of various statistical alignment models", |
| "authors": [ |
| { |
| "first": "Josef", |
| "middle": [], |
| "last": "Franz", |
| "suffix": "" |
| }, |
| { |
| "first": "Hermann", |
| "middle": [], |
| "last": "Och", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Ney", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Computational Linguistics", |
| "volume": "29", |
| "issue": "1", |
| "pages": "19--51", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Franz Josef Och and Hermann Ney. 2003. A systematic comparison of various statistical alignment models. Computational Linguistics, 29(1):19-51.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "The nature and dimensions of complexity in morphology", |
| "authors": [ |
| { |
| "first": "Gregory", |
| "middle": [], |
| "last": "Stump", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Annual Review of Linguistics", |
| "volume": "3", |
| "issue": "", |
| "pages": "65--83", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Gregory Stump. 2017. The nature and dimensions of complexity in morphology. Annual Review of Linguistics, 3:65-83.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "num": null, |
| "text": "Paradigm (Spanish); word translation sets (Spanish, Portuguese, Catalan, Galician)", |
| "type_str": "figure", |
| "uris": null |
| }, |
| "FIGREF1": { |
| "num": null, |
| "text": "Three rows from the table of Ge'ez characters.", |
| "type_str": "figure", |
| "uris": null |
| }, |
| "FIGREF2": { |
| "num": null, |
| "text": "(a) Morpheme alignment, phonemic and orthographic. 1-RELATIVE, 2-NEGATION, 3-PASSIVE, 4-STEM(wash), 5-PERFECTIVE, SBJ=3RD.MAS.SG. (b) Monotonic character alignment (Am, Ti only), diagrammed in two ways.", |
| "type_str": "figure", |
| "uris": null |
| }, |
| "FIGREF3": { |
| "num": null, |
| "text": "Alignment within an Am-Ti-Te translation set. Gloss: 'which was not washed.'", |
| "type_str": "figure", |
| "uris": null |
| }, |
| "FIGREF4": { |
| "num": null, |
| "text": "Alignment constraints (above), possible parameter-based alignments (below) for an Am-Ti translation pair.", |
| "type_str": "figure", |
| "uris": null |
| }, |
| "TABREF1": { |
| "text": ".405 0.906 0.865 0.959 0.552 0.932 0.581 0.925 0.650 0.813 0.614 0.865 Ti\u2192Am 0.405 0.926 0.870 0.976 0.553 0.950 0.496 0.893 0.625 0.824 0.553 0.857 Am", |
| "content": "<table><tr><td/><td/><td/><td/><td colspan=\"2\">Orthographic</td><td/><td/><td/><td/><td colspan=\"2\">Phonemic</td><td/></tr><tr><td/><td/><td colspan=\"2\">Recall</td><td colspan=\"2\">Precision</td><td colspan=\"2\">F-score</td><td colspan=\"2\">Recall</td><td colspan=\"2\">Precision</td><td colspan=\"2\">F-score</td></tr><tr><td/><td colspan=\"2\">Base</td><td>Train</td><td>Base</td><td>Train</td><td>Base</td><td>Train</td><td>Base</td><td>Train</td><td>Base</td><td>Train</td><td>Base</td><td>Train</td></tr><tr><td colspan=\"2\">Am\u2192Ti 0\u222a \u2194Ti</td><td>-</td><td>0.960</td><td>-</td><td>0.976</td><td>-</td><td>0.968</td><td>-</td><td>0.940</td><td>-</td><td>0.821</td><td>-</td><td>0.876</td></tr><tr><td>Am \u2194Ti \u2229</td><td/><td>-</td><td>0.896</td><td>-</td><td>0.961</td><td>-</td><td>0.927</td><td>-</td><td>0.925</td><td>-</td><td>0.813</td><td>-</td><td>0.865</td></tr><tr><td>Am \u2194Ti \u222a (mixed)</td><td/><td>-</td><td>-</td><td>-</td><td>-</td><td>-</td><td>-</td><td>-</td><td>0.894</td><td>-</td><td>0.770</td><td>-</td><td>0.827</td></tr><tr><td>Ti\u2192Te</td><td colspan=\"7\">0.484 0.973 0.868 0.979 0.621 0.976</td><td colspan=\"6\">0.659 0.976 0.773 0.864 0.711 0.917</td></tr><tr><td>Te\u2192Ti</td><td colspan=\"7\">0.515 0.974 0.880 0.977 0.650 0.975</td><td colspan=\"6\">0.756 0.981 0.797 0.858 0.776 0.915</td></tr><tr><td>Ti \u2194Te \u222a</td><td/><td>-</td><td>0.981</td><td>-</td><td>0.980</td><td>-</td><td>0.980</td><td>-</td><td>0.989</td><td>-</td><td>0.856</td><td>-</td><td>0.918</td></tr><tr><td>Ti \u2194Te \u2229</td><td/><td>-</td><td>0.973</td><td>-</td><td>0.979</td><td>-</td><td>0.976</td><td>-</td><td>0.976</td><td>-</td><td>0.864</td><td>-</td><td>0.917</td></tr></table>", |
| "type_str": "table", |
| "num": null, |
| "html": null |
| }, |
| "TABREF2": { |
| "text": "Experimental results.", |
| "content": "<table/>", |
| "type_str": "table", |
| "num": null, |
| "html": null |
| } |
| } |
| } |
| } |