| { |
| "paper_id": "2020", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T15:33:53.379802Z" |
| }, |
| "title": "Morphological Disambiguation of South S\u00e1mi with FSTs and Neural Networks", |
| "authors": [ |
| { |
| "first": "Mika", |
| "middle": [], |
| "last": "H\u00e4m\u00e4l\u00e4inen", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "UiT The Arctic University of Norway Finland", |
| "location": { |
| "country": "Norway" |
| } |
| }, |
| "email": "mika.hamalainen@helsinki.fi" |
| }, |
| { |
| "first": "Linda", |
| "middle": [], |
| "last": "Wiechetek", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "UiT The Arctic University of Norway Finland", |
| "location": { |
| "country": "Norway" |
| } |
| }, |
| "email": "linda.wiechetek@uit.no" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "We present a method for conducting morphological disambiguation for South S\u00e1mi, which is an endangered language. Our method uses an FST-based morphological analyzer to produce an ambiguous set of morphological readings for each word in a sentence. These readings are disambiguated with a Bi-RNN model trained on the related North S\u00e1mi UD Treebank and some synthetically generated South S\u00e1mi data. The disambiguation is done on the level of morphological tags ignoring word forms and lemmas; this makes it possible to use North S\u00e1mi training data for South S\u00e1mi without the need for a bilingual dictionary or aligned word embeddings. Our approach requires only minimal resources for South S\u00e1mi, which makes it usable and applicable in the contexts of any other endangered language as well.", |
| "pdf_parse": { |
| "paper_id": "2020", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "We present a method for conducting morphological disambiguation for South S\u00e1mi, which is an endangered language. Our method uses an FST-based morphological analyzer to produce an ambiguous set of morphological readings for each word in a sentence. These readings are disambiguated with a Bi-RNN model trained on the related North S\u00e1mi UD Treebank and some synthetically generated South S\u00e1mi data. The disambiguation is done on the level of morphological tags ignoring word forms and lemmas; this makes it possible to use North S\u00e1mi training data for South S\u00e1mi without the need for a bilingual dictionary or aligned word embeddings. Our approach requires only minimal resources for South S\u00e1mi, which makes it usable and applicable in the contexts of any other endangered language as well.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "S\u00e1mi languages are a part of the Uralic language family, and like many other Uralic languages they are endangered. The languages of this family are synthetic, meaning that they exhibit a great deal of inflectional and derivational morphology making their processing with computational means far from trivial. In this paper, we present a method for morphological disambiguation of South S\u00e1mi (ISO 639-3 code sma) by using a morphological FST (finite-state transducer) analyzer and a Bi-RNN (bi-directional recurrent neural network) trained on North S\u00e1mi (ISO 639-3 code sme) data and synthetically generated South S\u00e1mi data. The disambiguation process takes in all the morphological readings produced by the FST and uses the neural network to pick the contextually correct disambiguated reading. North and South are not direct neighbors in the dialect continuum, but share a big part of the lexicon and many grammatical features like an elaborate case system, non-finite clause constructions, a large amount of verbal and nominal derivations. However, they have a number of distinctions in lexicon, morphology and syntax. One of the important differences is the omission of the copula verb in South S\u00e1mi, but not or less so in North S\u00e1mi. The typical word order is SOV (subject-object-verb) in South S\u00e1mi, and SVO (subject-verb-object) in North S\u00e1mi. The case system is slightly different as well. South S\u00e1mi distinguishes between inessive (place) and elative (source) case (Bergsland, 1994) . In North S\u00e1mi, this is synthesized in one morpho-syntactic case, called locative case. In addition to the aforementioned differences, also the homonymies are not the same. In North S\u00e1mi, regular noun homonymies are genitive/accusative and comitative singular/locative plural. In South S\u00e1mi, on the other hand, they are illative plural/accusative plural and essive (underspecified as regards number)/inessive plural/comitative singular. Even in the context of morphologically rich languages, a simple POS (part-of-speech) tagging is often not enough as it only reduces some of the ambiguity, and is not enough for lemmatization, for instance. Then again, without lemmati-zation and the small amount of data available for these languages, modern NLP methods such as word embeddings cannot be as reliably used as in the case of majority languages. South S\u00e1mi, with its estimated number of 500 speakers, is categorized as severely endangered by UNESCO (Moseley, 2010) and is spoken in Norway and Sweden. The language is spoken in Norway and Sweden and its bilingual users frequently face bigger challenges regarding literacy in the lesser used language than in the majority language due to reduced access to language arenas (Outakoski, 2013; Lindgren et al., 2016) . The central tools used for disambiguation of S\u00e1mi languages are finite state transducers and Constraint Grammars. Constraint Grammar is a rule-based formalism for writing disambiguation and syntactic annotation grammars (Karlsson, 1990; Karlsson et al., 1995) . Constraint Grammar relies on a bottom-up analysis of running text. Possible but unlikely analyses are discarded step by step with the help of morpho-syntactic context. The vislcg3 implementation 1 is used in particular. South S\u00e1mi has several Constraint Grammars including a morpho-syntactic disambiguator, a shallow syntactic analyzer, and a dependency analyzer (Antonsen et al., 2010; Antonsen and Trosterud, 2011) . Antonsen and Trosterud (2011) use a fairly small Constraint Grammar (115 rules) for South S\u00e1mi part of speech (POS) and lemma disambiguation, resulting in a precision of 0.87 and a recall of 0.98 for full morpho-syntactic disambiguation. While these are very good results with a comparatively small workload, they require the work of a linguist with knowledge of the language or a linguist and a language expert in addition. However, we want to show how grammatical tools can be built in the absence of these. Whereas our paper deals with South S\u00e1mi disambiguation, the main purpose of this work is to demonstrate that a disambiguator can be built with relatively few resources based on a morpho-syntactically related language. This is", |
| "cite_spans": [ |
| { |
| "start": 1473, |
| "end": 1490, |
| "text": "(Bergsland, 1994)", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 2441, |
| "end": 2456, |
| "text": "(Moseley, 2010)", |
| "ref_id": null |
| }, |
| { |
| "start": 2713, |
| "end": 2730, |
| "text": "(Outakoski, 2013;", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 2731, |
| "end": 2753, |
| "text": "Lindgren et al., 2016)", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 2976, |
| "end": 2992, |
| "text": "(Karlsson, 1990;", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 2993, |
| "end": 3015, |
| "text": "Karlsson et al., 1995)", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 3381, |
| "end": 3404, |
| "text": "(Antonsen et al., 2010;", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 3405, |
| "end": 3434, |
| "text": "Antonsen and Trosterud, 2011)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 3437, |
| "end": 3466, |
| "text": "Antonsen and Trosterud (2011)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "Gos d\u00e1ppe lea m\u00e1ddi? 'Where is the South here?'", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sentence", |
| "sec_num": null |
| }, |
| { |
| "text": "FST output ['gos+Adv+Subqst', 'gos+Adv'] , ['d\u00e1ppe+Adv'] , ['leat+V+IV+Ind+Prs+Sg3'] , ['m\u00e1ddat+V+TV+Imprt+Du2', 'm\u00e1ddat+V+TV+PrsPrc', 'm\u00e1ddi+N+Sg+Nom'] Table 1 : An example of the training data useful, not only in the wider context of S\u00e1mi languages, but also for other endangered languages as it provides the language community quickly with much-needed resources while there are children -the future speakers -learning the language. Our approach follows the previously established ideology for using FSTs together with neural networks to solve the problem of disambiguation (Ens et al., 2019) .", |
| "cite_spans": [ |
| { |
| "start": 11, |
| "end": 40, |
| "text": "['gos+Adv+Subqst', 'gos+Adv']", |
| "ref_id": null |
| }, |
| { |
| "start": 43, |
| "end": 56, |
| "text": "['d\u00e1ppe+Adv']", |
| "ref_id": null |
| }, |
| { |
| "start": 59, |
| "end": 84, |
| "text": "['leat+V+IV+Ind+Prs+Sg3']", |
| "ref_id": null |
| }, |
| { |
| "start": 87, |
| "end": 112, |
| "text": "['m\u00e1ddat+V+TV+Imprt+Du2',", |
| "ref_id": null |
| }, |
| { |
| "start": 113, |
| "end": 134, |
| "text": "'m\u00e1ddat+V+TV+PrsPrc',", |
| "ref_id": null |
| }, |
| { |
| "start": 135, |
| "end": 152, |
| "text": "'m\u00e1ddi+N+Sg+Nom']", |
| "ref_id": null |
| }, |
| { |
| "start": 576, |
| "end": 594, |
| "text": "(Ens et al., 2019)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 153, |
| "end": 160, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Sentence", |
| "sec_num": null |
| }, |
| { |
| "text": "Parallel texts have been used to deal with morphological tagging in the context of low-resourced languages (Buys and Botha, 2016) . They use aligned parallel sentences to train their their Wsabie-based model to tag the low-resource language based on the morphological tags of a more resourced language sentences in the training data. A limitation of this approach is that the morphological relatedness of the high-resource and low-resource languages has to be high. Andrews et al. (Andrews et al., 2017) have proposed a method for POS (part of speech) tagging of low-resource languages. They use a bilingual dictionary between a lowresource and high-resource language. In addition, their system requires monolingual data for building cross-lingual word embeddings. The resulting POS tagger is trained on an LSTM neural network, and their approach performs consistently better than the other approaches on the benchmarks they report. Lim et al. (Lim et al., 2018) present an approach for syntactic parsing of Komi-Zyrian and North S\u00e1mi data using multilingual word embeddings. They use pre-trained wordembeddings of two high-resource languages; Finnish and Russian. Then they train monolingual word-embeddings for the low-resource languages from small corpora. They project these individual word embeddings into a single space by using bilingual dictionaries for alignment. The parser was implemented as an LSTM based model, and its performance is higher for POS tagging than for syntactic parsing. The most important finding for our purposes is that including a related high-resource language improves the accuracy of their method. DsDs (Plank and Agi\u0107, 2018 ) is a neural network based part-of-speech tagger intended to be used in the context of low-resource languages. Their core idea is to use a bi-LSTM model to project POS tags from one language to another with the help of lexical information and word embeddings. Their experiments in a low-resource setting reveal that including word embeddings can boost the model, but lexical information can also help to a smaller degree. The scope of a great part of the related work is limited to POS tagging. Nevertheless, the morphologically rich Uralic languages call for a more full blown morphological disambiguation than a mere POS tagging in order to make higher-level NLP tools usable for these languages. Moreover, our approach cannot count on the existence of high-quality bilingual dictionaries between morphologically similar languages nor aligned word embeddings, as such resources are not easily available for endangered languages.", |
| "cite_spans": [ |
| { |
| "start": 107, |
| "end": 129, |
| "text": "(Buys and Botha, 2016)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 481, |
| "end": 503, |
| "text": "(Andrews et al., 2017)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 944, |
| "end": 962, |
| "text": "(Lim et al., 2018)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 1637, |
| "end": 1658, |
| "text": "(Plank and Agi\u0107, 2018", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2." |
| }, |
| { |
| "text": "The training data for South S\u00e1mi disambiguation comes from the Universal Dependencies Treebank of the related North S\u00e1mi language (Sheyanova and Tyers, 2017) . Out of all the S\u00e1mi languages, North S\u00e1mi has by and large the biggest amount of NLP resources available and therefore its use as a starting point for related languages makes perfect sense. The treebank consists of 26K tokens and comes pre-divided into a training and testing datasets.", |
| "cite_spans": [ |
| { |
| "start": 130, |
| "end": 157, |
| "text": "(Sheyanova and Tyers, 2017)", |
| "ref_id": "BIBREF23" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data and Tools", |
| "sec_num": "3." |
| }, |
| { |
| "text": "In addition to the treebank, we use FSTs for both North S\u00e1mi and South S\u00e1mi with UralicNLP (H\u00e4m\u00e4l\u00e4inen, 2019) . These transducers are integrated in the open GiellaLT infrastructure (Moshagen et al., 2014) for Uralic languages. The FSTs take in a word in an inflectional form and produce all the possible morphological readings for it. In order to evaluate our system, we use a small dataset for South S\u00e1mi that has been disambiguated automatically by a Constraint Grammar and checked manually. Currently, the dataset is not publicly available. The data consists of 1994 disambiguated sentences and we only use it for the evaluation.", |
| "cite_spans": [ |
| { |
| "start": 91, |
| "end": 109, |
| "text": "(H\u00e4m\u00e4l\u00e4inen, 2019)", |
| "ref_id": "BIBREF21" |
| }, |
| { |
| "start": 181, |
| "end": 204, |
| "text": "(Moshagen et al., 2014)", |
| "ref_id": "BIBREF22" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data and Tools", |
| "sec_num": "3." |
| }, |
| { |
| "text": "North S\u00e1mi South S\u00e1mi Average 3.1 1.8 Table 2 : Average ambiguity Table 2 shows the average morphological ambiguity in the North S\u00e1mi training set and South S\u00e1mi test set when the FSTs are used to produce all morphological readings for every word in the corpus. As we can see, North S\u00e1mi exhibits a much higher degree of morphological ambiguity than South S\u00e1mi. For generating more data, we use the South S\u00e1mi lemmas from the South S\u00e1mi-Norwegian dictionary located in the GiellaLT infrastructure (Moshagen et al., 2014) . The dictionary has 11,438 POS tagged South S\u00e1mi lemmas. We only use this dictionary for South S\u00e1mi words and omit all the Norwegian translations in our method. ", |
| "cite_spans": [ |
| { |
| "start": 497, |
| "end": 520, |
| "text": "(Moshagen et al., 2014)", |
| "ref_id": "BIBREF22" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 38, |
| "end": 45, |
| "text": "Table 2", |
| "ref_id": null |
| }, |
| { |
| "start": 66, |
| "end": 73, |
| "text": "Table 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Data and Tools", |
| "sec_num": "3." |
| }, |
| { |
| "text": "We train a sequence-to-sequence Bi-RNN model using OpenNMT (Klein et al., 2017) with the default settings except for the encoder where we use a BRNN (bi-directional recurrent neural network) instead of the default RNN (recurrent neural network) as BRNN has been shown to provide a performance gain in a variety of tasks. We use the default of two layers for both the encoder and the decoder and the default attention model, which is the general global attention presented by Luong et al. (Luong et al., 2015) .", |
| "cite_spans": [ |
| { |
| "start": 59, |
| "end": 79, |
| "text": "(Klein et al., 2017)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 475, |
| "end": 508, |
| "text": "Luong et al. (Luong et al., 2015)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Neural Disambiguation", |
| "sec_num": "4." |
| }, |
| { |
| "text": "We experiment with two models, one that is trained with the North S\u00e1mi Treebank only, and another one that is trained with South S\u00e1mi text generated by templates and the North S\u00e1mi data. Both models are trained for 60,000 training steps with the same random seed value.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Neural Disambiguation", |
| "sec_num": "4." |
| }, |
| { |
| "text": "The North S\u00e1mi data gives us the target sequence, that is the correct morphological tags and the POS tag. However, the source sequence has to be generated automatically before the training. For this, we use the North S\u00e1mi FST analyzer. We produce all the possible morphologies for each word in the Treebank. The training is done from a sorted list of homonymous readings for each word separated by a character indicating word boundary to the disambiguated set of homonymous readings from the UD (universal dependencies) TreeBank on a sentence level. In other words, the only thing the model sees are morphological tags on the source and the target side. Lemmas and words are dropped out so that the model can be used for South S\u00e1mi without the need of aligned word embeddings or dictionaries. This is illustrated in Table 1 .", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 816, |
| "end": 823, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Neural Disambiguation", |
| "sec_num": "4." |
| }, |
| { |
| "text": "For producing synthetic data, we wrote six small templates that reflect some common morpho-syntactic differences between South S\u00e1mi and North S\u00e1mi. This is, for instance, the absence of elative and inessive case in North S\u00e1mi, both of which are merged into the single locative case. For each template, we produce 20 different ambiguous sentences by selecting words fitting to the template at random from the South S\u00e1mi dictionary and inflecting them accordingly with the FST. Once the words are inflected, we can analyze them to get the ambiguous reading. The templates can be seen in Table 3 .", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 585, |
| "end": 592, |
| "text": "Table 3", |
| "ref_id": "TABREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Neural Disambiguation", |
| "sec_num": "4." |
| }, |
| { |
| "text": "We evaluate the models, the one trained only with the North S\u00e1mi Treebank and the one that had additional template generated training data, with the disambiguated gold standard that exists for South S\u00e1mi. As the South S\u00e1mi gold standard follows the GiellaLT FST tags, we converted the tags automatically into UD format, since the neural network is trained to output UD tags. The evaluation results are shown in Table 4 . The first column shows the percentage of sentences that have been fully disambiguated correctly, the second columns shows this on a word level i.e. how many words were fully correctly disambiguated and finally the last column shows the accuracy in POS tagging. The results indicate that adding the small synthetically generated data to the training boosted the results significantly. As for the incorrectly disambiguated morphological readings, there is a degree to how incorrect they are. This is shown in Table 5 , which shows the errors based on how many morphological tags were predicted wrong. In both cases, more than half of the wrongly disambiguated words only differ by one tag from the gold standard. The results for the model trained on the additional template data show that the errors the model makes are still closer to the correct reading. Below, we are having a closer look at the actual sentences and their analyses, shedding some light on the shortcomings the neural network and suggesting improvements. In ex.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 411, |
| "end": 418, |
| "text": "Table 4", |
| "ref_id": "TABREF4" |
| }, |
| { |
| "start": 928, |
| "end": 935, |
| "text": "Table 5", |
| "ref_id": "TABREF6" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results and Evaluation", |
| "sec_num": "5." |
| }, |
| { |
| "text": "(1), our system erroneously picks the nominal singular nominative instead of the adverb reading for daelie. The nominal reading, however, is very rare. There are other difficulties related to negation in the system. In the following example, the neural network predicts more tokens than the sentence contains, i.e. a negative verb (correctly) and a connegative form (erroneously) usually preceded by the negative verb.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Fully correct sentences", |
| "sec_num": null |
| }, |
| { |
| "text": "(3) --Aellieh! not.NEG.IMPRT.SG2 '-Don't!'", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Fully correct sentences", |
| "sec_num": null |
| }, |
| { |
| "text": "Uralic languages are highly ambiguous in terms of their morphology, and the linguistic resources such as annotated corpora for these languages are quite limited. This poses challenges in the use of modern NLP methods that have been successfully employed on high-resource languages. In order to overcome these limitations, we proposed a representation based on the ambiguous morphological tags of each word in a sentence. We have presented a viable way of disambiguation for South S\u00e1mi based on an FST and training data on North S\u00e1mi with minimal templates needed to cover some of the morpho-syntactic differences of the two languages. The preliminary results look promising, especially since there are nine different S\u00e1mi languages. Not to mention similar situations for other endangered languages, where data for a similar language is available. Our method is more of a hybrid pipeline of rule-based FSTs that produce the possible morphological readings and a neural network that does the disambiguation. This makes it possible to replace the FST with some other rule-based solution or a neural network based morphological analyzer, given that recent research has shown promising results for the use of neural networks in morphology of endangered languages (Schwartz et al., 2019; Silfverberg and Tyers, 2019) . Moreover, our pipeline can be further enhanced by rules. In our experiments, we had the neural network disambiguate out of all the possible morphological readings. Instead of doing that, it is possible to disambiguate first with a rulebased tool such as a Constraint Grammar, and use the neural network to disambiguate the remaining ambiguity. That way we do not need to guess what we already know. It is particularly important to make sure that if the morphology is known, the neural network would not be used to guess it again. This would allow for combining the best of the two worlds; the accuracy of the rule-based methods and the scalability of a neural network. An interesting question for the future is how far one could get in disambiguation with our proposed method if one was only to train the model by using templates. As even a small number of templates was enough to improve the results noticeably, an entirely template based approach does not seem to be entirely out of the question. Especially if the templates were constructed with more generative freedom such as by following a formalism deriving from CFG (context-free grammar). The use of synthetically generated source data is known to improve NMT (neural machine translation) models when the target data is of a high quality (see (Sennrich et al., 2016) ). Also, some promising work has been conducted in fully synthetically generated parallel data in NMT (H\u00e4m\u00e4l\u00e4inen and Alnajjar, 2019) . This year has been particularly good for Uralic languages with small UD Treebanks recently published for Skolt S\u00e1mi, Karelian, Livvi, Komi-Permyak and Moksha. This means that in the future we can try different variations of our method with these languages as well with minimal modifications to the current approach as all of these languages have rule-based FSTs available in the GiellaLT infrastructure.", |
| "cite_spans": [ |
| { |
| "start": 1258, |
| "end": 1281, |
| "text": "(Schwartz et al., 2019;", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 1282, |
| "end": 1310, |
| "text": "Silfverberg and Tyers, 2019)", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 2615, |
| "end": 2638, |
| "text": "(Sennrich et al., 2016)", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 2741, |
| "end": 2772, |
| "text": "(H\u00e4m\u00e4l\u00e4inen and Alnajjar, 2019)", |
| "ref_id": "BIBREF7" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussion and Conclusions", |
| "sec_num": "6." |
| }, |
| { |
| "text": "http://visl.sdu.dk/constraint_grammar. html (accessed 2018-10-08), also(Didriksen, 2010)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "We would like to thank Lene Antonsen and Anja Regina Fjellheim Labj for their work on the South S\u00e1mi Constraint Grammar disambiguator within the GiellaLT infrastructure and for making their automatically annotated and manually corrected South S\u00e1mi corpus available to us.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgments", |
| "sec_num": "7." |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Bayesian modeling of lexical resources for lowresource settings", |
| "authors": [ |
| { |
| "first": "N", |
| "middle": [], |
| "last": "Andrews", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Dredze", |
| "suffix": "" |
| }, |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Van Durme", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Eisner", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "1029--1039", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Andrews, N., Dredze, M., Van Durme, B., and Eisner, J. (2017). Bayesian modeling of lexical resources for low- resource settings. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguis- tics (Volume 1: Long Papers), pages 1029-1039. Asso- ciation for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Next to nothing -a cheap south saami disambiguator", |
| "authors": [ |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Antonsen", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Trosterud", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Antonsen, L. and Trosterud, T. (2011). Next to nothing -a cheap south saami disambiguator. pages 131-137, 05.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Reusing grammatical resources for new languages", |
| "authors": [ |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Antonsen", |
| "suffix": "" |
| }, |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Wiechetek", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Trosterud", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the 7th International Conference on Language Resources and Evaluation (LREC 2010)", |
| "volume": "", |
| "issue": "", |
| "pages": "2782--2789", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Antonsen, L., Wiechetek, L., and Trosterud, T. (2010). Reusing grammatical resources for new languages. In Proceedings of the 7th International Conference on Lan- guage Resources and Evaluation (LREC 2010), pages 2782-2789, Stroudsburg. The Association for Compu- tational Linguistics.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Sydsamisk grammatikk. Davvi Girji", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Bergsland", |
| "suffix": "" |
| } |
| ], |
| "year": 1994, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bergsland. (1994). Sydsamisk grammatikk. Davvi Girji.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Cross-lingual morphological tagging for low-resource languages", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Buys", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [ |
| "A" |
| ], |
| "last": "Botha", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "1954--1964", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Buys, J. and Botha, J. A. (2016). Cross-lingual morpholog- ical tagging for low-resource languages. In Proceedings of the 54th Annual Meeting of the Association for Com- putational Linguistics (Volume 1: Long Papers), pages 1954-1964. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Constraint Grammar Manual: 3rd version of the CG formalism variant", |
| "authors": [ |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Didriksen", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Didriksen, T., (2010). Constraint Grammar Manual: 3rd version of the CG formalism variant. GrammarSoft ApS, Denmark.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Morphosyntactic disambiguation in an endangered language setting", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Ens", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "H\u00e4m\u00e4l\u00e4inen", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Rueter", |
| "suffix": "" |
| }, |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Pasquier", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 22nd Nordic Conference on Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "345--349", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ens, J., H\u00e4m\u00e4l\u00e4inen, M., Rueter, J., and Pasquier, P. (2019). Morphosyntactic disambiguation in an endangered lan- guage setting. In Proceedings of the 22nd Nordic Con- ference on Computational Linguistics, pages 345-349.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "A template based approach for training nmt for low-resource uralic languages-a pilot with finnish", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "H\u00e4m\u00e4l\u00e4inen", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Alnajjar", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 2019 2nd International Conference on Algorithms, Computing and Artificial Intelligence", |
| "volume": "", |
| "issue": "", |
| "pages": "520--525", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "H\u00e4m\u00e4l\u00e4inen, M. and Alnajjar, K. (2019). A template based approach for training nmt for low-resource uralic languages-a pilot with finnish. In Proceedings of the 2019 2nd International Conference on Algorithms, Com- puting and Artificial Intelligence, pages 520-525.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Constraint Grammar: A Language-Independent System for Parsing Unrestricted Text", |
| "authors": [ |
| { |
| "first": "F", |
| "middle": [], |
| "last": "Karlsson", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Voutilainen", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Heikkil\u00e4", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Anttila", |
| "suffix": "" |
| } |
| ], |
| "year": 1995, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Karlsson, F., Voutilainen, A., Heikkil\u00e4, J., and Anttila, A. (1995). Constraint Grammar: A Language-Independent System for Parsing Unrestricted Text. Mouton de Gruyter, Berlin.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Constraint Grammar as a Framework for Parsing Running Text", |
| "authors": [ |
| { |
| "first": "F", |
| "middle": [], |
| "last": "Karlsson", |
| "suffix": "" |
| } |
| ], |
| "year": 1990, |
| "venue": "Proceedings of the 13th Conference on Computational Linguistics (COLING 1990)", |
| "volume": "3", |
| "issue": "", |
| "pages": "168--173", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Karlsson, F. (1990). Constraint Grammar as a Framework for Parsing Running Text. In Hans Karlgren, editor, Proceedings of the 13th Conference on Computational Linguistics (COLING 1990), volume 3, pages 168-173, Helsinki, Finland. Association for Computational Lin- guistics.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "OpenNMT: Open-Source Toolkit for Neural Machine Translation", |
| "authors": [ |
| { |
| "first": "G", |
| "middle": [], |
| "last": "Klein", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [], |
| "last": "Deng", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Senellart", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [ |
| "M" |
| ], |
| "last": "Rush", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proc. ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Klein, G., Kim, Y., Deng, Y., Senellart, J., and Rush, A. M. (2017). OpenNMT: Open-Source Toolkit for Neural Ma- chine Translation. In Proc. ACL.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Multilingual dependency parsing for low-resource languages: Case studies on north saami and komi-zyrian", |
| "authors": [ |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Lim", |
| "suffix": "" |
| }, |
| { |
| "first": "N", |
| "middle": [], |
| "last": "Partanen", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Poibeau", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the Eleventh International Conference on Language Resources and Evaluation", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lim, K., Partanen, N., and Poibeau, T. (2018). Multi- lingual dependency parsing for low-resource languages: Case studies on north saami and komi-zyrian. In Pro- ceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018).", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Researching literacy development in the globalised North: studying tri-lingual children's english writing in Finnish, Norwegian and Swedish S\u00e1pmi", |
| "authors": [ |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Lindgren", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Sullivan", |
| "suffix": "" |
| }, |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Outakoski", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Westum", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Super Dimensions in Globalisation and Education, Cultural Studies and Transdiciplinarity in Education", |
| "volume": "", |
| "issue": "", |
| "pages": "55--68", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lindgren, E., Sullivan, K., Outakoski, H., and Westum, A. (2016). Researching literacy development in the globalised North: studying tri-lingual children's english writing in Finnish, Norwegian and Swedish S\u00e1pmi. In David R. Cole et al., editors, Super Dimensions in Glob- alisation and Education, Cultural Studies and Transdici- plinarity in Education, pages 55-68. Springer, Singa- pore.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Effective approaches to attention-based neural machine translation", |
| "authors": [ |
| { |
| "first": "M.-T", |
| "middle": [], |
| "last": "Luong", |
| "suffix": "" |
| }, |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Pham", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [ |
| "D" |
| ], |
| "last": "Manning", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1508.04025" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Luong, M.-T., Pham, H., and Manning, C. D. (2015). Effective approaches to attention-based neural machine translation. arXiv preprint arXiv:1508.04025.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Atlas of the World s Languages in Danger", |
| "authors": [], |
| "year": 2010, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Christopher Moseley, editor. (2010). Atlas of the World s Languages in Danger. UNESCO Publishing, 3rd edi- tion. Online version: http://www.unesco.org/languages- atlas/.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Davvis\u00e1megielat\u010d\u00e1lam\u00e1htu konteaksta", |
| "authors": [ |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Outakoski", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "S\u00e1mi dieala\u0161\u00e1ige\u010d\u00e1la", |
| "volume": "1", |
| "issue": "", |
| "pages": "29--59", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Outakoski, H. (2013). Davvis\u00e1megielat\u010d\u00e1lam\u00e1htu kon- teaksta [The context of North S\u00e1mi literacy]. S\u00e1mi dieala\u0161\u00e1ige\u010d\u00e1la, 1/2015:29-59.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Distant supervision from disparate sources for low-resource part-of-speech tagging", |
| "authors": [ |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Plank", |
| "suffix": "" |
| }, |
| { |
| "first": "\u017d", |
| "middle": [], |
| "last": "Agi\u0107", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "614--620", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Plank, B. and Agi\u0107,\u017d. (2018). Distant supervision from disparate sources for low-resource part-of-speech tag- ging. In Proceedings of the 2018 Conference on Em- pirical Methods in Natural Language Processing, pages 614-620. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Bootstrapping a neural morphological analyzer for st. lawrence island yupik from a finite-state transducer", |
| "authors": [ |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Schwartz", |
| "suffix": "" |
| }, |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Hunt", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [ |
| "L" |
| ], |
| "last": "Schreiner", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 3rd Workshop on the Use of Computational Methods in the Study of Endangered Languages", |
| "volume": "1", |
| "issue": "", |
| "pages": "87--96", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Schwartz, L., Chen, E., Hunt, B., and Schreiner, S. L. (2019). Bootstrapping a neural morphological analyzer for st. lawrence island yupik from a finite-state trans- ducer. In Proceedings of the 3rd Workshop on the Use of Computational Methods in the Study of Endangered Languages Volume 1 (Papers), pages 87-96, Honolulu, February. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Improving neural machine translation models with monolingual data", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Sennrich", |
| "suffix": "" |
| }, |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Haddow", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Birch", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "86--96", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sennrich, R., Haddow, B., and Birch, A. (2016). Improv- ing neural machine translation models with monolingual data. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 86-96, Berlin, Germany, August. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Data-driven morphological analysis for uralic languages", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Silfverberg", |
| "suffix": "" |
| }, |
| { |
| "first": "F", |
| "middle": [], |
| "last": "Tyers", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the Fifth International Workshop on Computational Linguistics for Uralic Languages", |
| "volume": "", |
| "issue": "", |
| "pages": "1--14", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Silfverberg, M. and Tyers, F. (2019). Data-driven morpho- logical analysis for uralic languages. In Proceedings of the Fifth International Workshop on Computational Lin- guistics for Uralic Languages, pages 1-14, Tartu, Esto- nia, January. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "UralicNLP: An NLP library for Uralic languages", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "H\u00e4m\u00e4l\u00e4inen", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Journal of Open Source Software", |
| "volume": "4", |
| "issue": "37", |
| "pages": "", |
| "other_ids": { |
| "DOI": [ |
| "1345.10.21105/joss.01345" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "H\u00e4m\u00e4l\u00e4inen, M. (2019). UralicNLP: An NLP library for Uralic languages. Journal of Open Source Software, 4(37):1345. 10.21105/joss.01345.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "Open-Source Infrastructures for Collaborative Work on Under-Resourced Languages", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Moshagen", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Rueter", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Pirinen", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Trosterud", |
| "suffix": "" |
| }, |
| { |
| "first": "F", |
| "middle": [ |
| "M" |
| ], |
| "last": "Tyers", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "The LREC 2014 Workshop \"CCURL 2014 -Collaboration and Computing for Under-Resourced Languages in the Linked Open Data Era", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Moshagen, S., Rueter, J., Pirinen, T., Trosterud, T., and Ty- ers, F. M. (2014). Open-Source Infrastructures for Col- laborative Work on Under-Resourced Languages. The LREC 2014 Workshop \"CCURL 2014 -Collaboration and Computing for Under-Resourced Languages in the Linked Open Data Era\".", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "Annotation schemes in north s\u00e1mi dependency parsing", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Sheyanova", |
| "suffix": "" |
| }, |
| { |
| "first": "F", |
| "middle": [ |
| "M" |
| ], |
| "last": "Tyers", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 3rd International Workshop for Computational Linguistics of Uralic Languages", |
| "volume": "", |
| "issue": "", |
| "pages": "66--75", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sheyanova, M. and Tyers, F. M. (2017). Annotation schemes in north s\u00e1mi dependency parsing. In Proceed- ings of the 3rd International Workshop for Computa- tional Linguistics of Uralic Languages, pages 66-75.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "TABREF2": { |
| "num": null, |
| "text": "Templates for generating South S\u00e1mi data", |
| "html": null, |
| "type_str": "table", |
| "content": "<table/>" |
| }, |
| "TABREF4": { |
| "num": null, |
| "text": "Evaluation results of the two different models on South S\u00e1mi data", |
| "html": null, |
| "type_str": "table", |
| "content": "<table/>" |
| }, |
| "TABREF6": { |
| "num": null, |
| "text": "", |
| "html": null, |
| "type_str": "table", |
| "content": "<table><tr><td/><td colspan=\"6\">: Errors based on the number of erroneous morpho-</td></tr><tr><td colspan=\"5\">logical tags on a word level</td><td/></tr><tr><td>(1)</td><td>Daelie</td><td/><td/><td>dle</td><td colspan=\"2\">geajnam</td><td>gaavnem!</td></tr><tr><td/><td colspan=\"3\">now;thenSG.NOM</td><td>so</td><td colspan=\"2\">street.ACC</td><td>find.PRS.SG1</td></tr><tr><td/><td colspan=\"5\">'Then I find the street!'</td></tr><tr><td colspan=\"7\">Negation verbs pose a problem to the neural network. Of</td></tr><tr><td colspan=\"7\">the 238 instances only very few negation verbs -despite not</td></tr><tr><td colspan=\"7\">being homonymous with any other forms -are analyzed as</td></tr><tr><td colspan=\"7\">such. In ex. (2), im 'I don't' is analyzed a an indicative past</td></tr><tr><td colspan=\"7\">tense verb 1st person singular (the last of which is correct)</td></tr><tr><td colspan=\"7\">despite the fact that im is not ambiguous.</td></tr><tr><td>(2)</td><td>Im</td><td/><td colspan=\"2\">s\u00efjhth</td><td/><td>g\u00e5abph</td><td>gih,</td><td>men</td></tr><tr><td/><td colspan=\"2\">not.NEG.SG1</td><td colspan=\"3\">want.CONNEG</td><td>anywhere</td><td>then,</td><td>but</td></tr><tr><td/><td>tjidtjie</td><td>jeahta</td><td colspan=\"3\">m\u00e5nnoeh.</td></tr><tr><td/><td>mother</td><td>says</td><td colspan=\"3\">us.DU1.NOM</td></tr><tr><td/><td colspan=\"6\">'I don't want anywhere then, but mother says us two</td></tr><tr><td/><td colspan=\"2\">will go.'</td><td/><td/><td/></tr></table>" |
| } |
| } |
| } |
| } |