| { |
| "paper_id": "D15-1025", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T16:27:08.515907Z" |
| }, |
| "title": "Non-lexical neural architecture for fine-grained POS Tagging", |
| "authors": [ |
| { |
| "first": "Matthieu", |
| "middle": [], |
| "last": "Labeau", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Universit\u00e9 Paris-Sud and LIMSI-CNRS", |
| "location": { |
| "addrLine": "Rue John von Neumann", |
| "postCode": "91403", |
| "settlement": "Orsay cedex France" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "L\u00f6ser", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Universit\u00e9 Paris-Sud and LIMSI-CNRS", |
| "location": { |
| "addrLine": "Rue John von Neumann", |
| "postCode": "91403", |
| "settlement": "Orsay cedex France" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Alexandre", |
| "middle": [], |
| "last": "Allauzen", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Universit\u00e9 Paris-Sud and LIMSI-CNRS", |
| "location": { |
| "addrLine": "Rue John von Neumann", |
| "postCode": "91403", |
| "settlement": "Orsay cedex France" |
| } |
| }, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "In this paper we explore a POS tagging application of neural architectures that can infer word representations from the raw character stream. It relies on two modelling stages that are jointly learnt: a convolutional network that infers a word representation directly from the character stream, followed by a prediction stage. Models are evaluated on a POS and morphological tagging task for German. Experimental results show that the convolutional network can infer meaningful word representations, while for the prediction stage, a well designed and structured strategy allows the model to outperform stateof-the-art results, without any feature engineering.", |
| "pdf_parse": { |
| "paper_id": "D15-1025", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "In this paper we explore a POS tagging application of neural architectures that can infer word representations from the raw character stream. It relies on two modelling stages that are jointly learnt: a convolutional network that infers a word representation directly from the character stream, followed by a prediction stage. Models are evaluated on a POS and morphological tagging task for German. Experimental results show that the convolutional network can infer meaningful word representations, while for the prediction stage, a well designed and structured strategy allows the model to outperform stateof-the-art results, without any feature engineering.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Most modern statistical models for natural language processing (NLP) applications are strongly or fully lexicalized, for instance part-of-speech (POS) and named entity taggers, as well as language models, and parsers. In these models, the observed word form is considered as the elementary unit, while its morphological properties remain neglected. As a result, the vocabulary observed on training data heavily restricts the generalization power of lexicalized models.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Designing subword-level systems is appealing for several reasons. First, words sharing morphological properties often share grammatical function and meaning, and leveraging that information can yield improved word representations. Second, a subword-level analysis can address the outof-vocabulary issue i.e the fact that word-level models fail to meaningfully process unseen word forms. This allows a better processing of morphologically rich languages in which there is a combinatorial explosion of word forms, most of which are not observed during training. Finally, using subword units could allow processing of noisy text such as user-generated content on the Web, where abbreviations, slang usage and spelling mistakes cause the number of word types to explode.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "This work investigates models that do not rely on a fixed vocabulary to make a linguistic prediction. Our main focus in this paper is POS tagging, yet the proposed approach could be applied to a wide variety of language processing tasks. Our main contribution is to show that neural networks can successfully learn unlexicalized models that infer a useful word representation from the character stream. This approach achieves state of-the-art performance on a German POS tagging task. This task is difficult because German is a morphologically rich language 1 , as reflected by the large number of morphological tags (255) in our study, yielding a grand total of more than 600 POS+MORPH tags. An aggravating factor is that these morphological categories are overtly marked by a handful of highly ambiguous inflection marks (suffixes). We therefore believe that this case study is well suited to assess both the representation and prediction power of our models.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The architecture we explore in section 2 differs from previous work that only consider the character level. Following (Santos and Zadrozny, 2014), it consists in two stages that are jointly learnt. The lower stage is a convolutional network that infers a word embedding from a character string of arbitrary size, while the higher network infers the POS tags based on this word embedding sequence. For the latter, we investigate different architectures of increasing complexities: from a feedforward and context-free inference to a bi-recurrent network that predicts the global sequence. Experimental results (section 4) show that the proposed approach can achieve state of the art performance and that the choice of architecture for the prediction part of the model has a significant impact.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The different architectures we propose act in two stages to infer, for a sentence s = {w 1 , . . . , w |s| }, a sequence of tags {t 1 , . . . , t |s| }. Each tag belongs to the tagset T . The first stage is designed to represent each word locally, and focuses on capturing the meaningful morphological information. In the second stage, we investigate different ways to predict the tag sequence that differ in how the global information is used.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Network Architectures", |
| "sec_num": "2" |
| }, |
| { |
| "text": "To obtain word embeddings, the usual approach introduced by (Bengio et al., 2003) relies on a fixed vocabulary W and each word w \u2208 W is mapped to a vector of n f real valued features by a look-up matrix W \u2208 R |W| * n f . To avoid the use of a fixed vocabulary, we propose to derive a word representation from a sequence of character embedding: if C denotes the finite set of characters, each character is mapped on a vector of n c features gathered in the look-up matrix C.", |
| "cite_spans": [ |
| { |
| "start": 60, |
| "end": 81, |
| "text": "(Bengio et al., 2003)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "From character to word level", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "To infer a word embedding , we use a convolution layer (Waibel et al., 1990; , build as in (Santos and Zadrozny, 2014). As illustrated in figure 1, a word w is a character sequence {c 1 , .., c |w| } represented by their embeddings {C c 1 , .., C c |w| }, where C c i denotes the row in C associated to the character c i . A convolution filter W conv \u2208 R n f \u00d7 R dc * nc is applied over a sliding window of d c characters, producing local features :", |
| "cite_spans": [ |
| { |
| "start": 55, |
| "end": 76, |
| "text": "(Waibel et al., 1990;", |
| "ref_id": "BIBREF17" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "From character to word level", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "x n = W conv (C c n\u2212dc+1 : .. : C cn ) T + b conv ,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "From character to word level", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "where x n is a vector of size n f obtained for each position n in the word 2 . The i-th element of the embedding of w is the maximum over the i-th elements of the feature vectors :", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "From character to word level", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "[f ] i = tanh( max 1\u2264n\u2264|s| [x n ] i )", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "From character to word level", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Using a maximum after a sliding convolution window ensures that the embedding combines local features from the whole word, and selects the more 2 Two padding character tokens are used to deal with border effects. The first is added at the beginning and the second at the end of the word, as many times as it is necessary to obtain the same number of windows than the length of the word. Their embeddings are added to C. useful ones. The parameters of the layer are the matrices C and W conv and the bias b conv .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "From character to word level", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "n c W conv \u00d7 (.) T + b conv n f e max(.) S h e n f 1 f 2 f 3 f 4 f 5 C c1 C c2 C c3 C c4 C c5 C sow C eow", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "From character to word level", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "To predict the tag sequence associated to a sentence s, we first use a feedforward architecture, with a single hidden layer. To compute the probability of tagging the n-th word in the sentence with tag t i , we use a window of d w word embeddings 3 centered around the word w n :", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "From words to prediction", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "x n = f n\u2212 dw \u22121 2 : ... : f n+ dw \u22121 2 ,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "From words to prediction", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "followed by a hidden and output layers:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "From words to prediction", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "s n = W o tanh(W h x n + b h ) + b o .", |
| "eq_num": "(1)" |
| } |
| ], |
| "section": "From words to prediction", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "The parameters of the hidden an output layers are respectively W h , b h and W o , b o .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "From words to prediction", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "We also experiment with a a bidirectional recurrent layer, as described in . The forward and backward passes allow each prediction to be conditioned on the complete past and future contexts, instead of merely a neighboring window. As illustrated in figure 2, the forward hidden state, at position n, will be computed using the previous forward hidden state and the word embedding in position n: \u2212\u2212\u2192 W f h and \u2212\u2212\u2192 W hh are the transition matrices of the forward part of the layer, and b h is the bias. The backward hidden states are computed similarly, and the hidden states of each direction are concatenated to pass through an output layer:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "From words to prediction", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "\u2212 \u2192 h n = tanh( \u2212\u2212\u2192 W f h f n + \u2212\u2212\u2192 W hh \u2212\u2212\u2192 h n\u22121 + b h ) n f |T | n h f1 f2 f3 f4 f5", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "From words to prediction", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "s n = W o ( \u2212 \u2192 h n : \u2190 \u2212 h n ) + b o .", |
| "eq_num": "(2)" |
| } |
| ], |
| "section": "From words to prediction", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "To infer the tag sequence from the sequence of output layers defined by equations 1 or 2, we explore two strategies. The first simply applies a softmax function to the output layer of the network described in the previous section. In this case, each tag prediction is made independently of the surrounding predictions. For sequence labeling, a more appropriate solution relies on the approach of (Collobert, 2011) , also used in (Santos and Zadrozny, 2014). Let consider each possible tag sequence {t 1 , . . . , t |s| } as a possible path over a sequence of hidden states. We can add a transition matrix W trans and then compute the score of a sequence as follows:", |
| "cite_spans": [ |
| { |
| "start": 396, |
| "end": 413, |
| "text": "(Collobert, 2011)", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Inference and Training", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "s({t} |s| 1 , {w} |s| 1 ) = 1\u2264n\u2264|s| W trans t n\u22121 ,tn + [s n ] tn", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Inference and Training", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "The Viterbi algorithm (Viterbi, 1967) offers an exact solution to infer the path that gives the maximum score. It is worth noticing that both these strategies can be applied to the feedforward and bidirectional recurrent networks. For both strategies, the whole network can estimate conditional log-likelihood of a tag sequence given a sentence s and the set of parameters \u03b8. This criterion can then be optimized using a stochastic gradient ascent with the back-propagation algorithm.", |
| "cite_spans": [ |
| { |
| "start": 22, |
| "end": 37, |
| "text": "(Viterbi, 1967)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Inference and Training", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "The choice to consider words from the character level has recently been more and more explored. While its raw application to language modeling did not achieve clear improvement over the word-based models (Mikolov et al., 2012) , this approach shown impressive results for text generation (Sutskever et al., 2011; Graves, 2013) . However, for this line of work, the main issue is to learn long range dependencies at the character level since the word level is not considered by the model. More recently, the character level was considered as more interpretable and convenient way to explore and understand recurrent networks (Karpathy et al., 2015) . In (Zhang and Le-Cun, 2015) , the authors build a text understanding model that does not require any knowledge and uses hierarchical feature extraction. Here the character level allows the model to ignore the definition a priori of a vocabulary and let the model build its own representation of a sentence or a document, directly from the character level. To some extent, our work can be considered as an extension of their work, tailored for POS tagging.", |
| "cite_spans": [ |
| { |
| "start": 204, |
| "end": 226, |
| "text": "(Mikolov et al., 2012)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 288, |
| "end": 312, |
| "text": "(Sutskever et al., 2011;", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 313, |
| "end": 326, |
| "text": "Graves, 2013)", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 624, |
| "end": 647, |
| "text": "(Karpathy et al., 2015)", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 653, |
| "end": 677, |
| "text": "(Zhang and Le-Cun, 2015)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "3" |
| }, |
| { |
| "text": "(Santos and Zadrozny, 2014) applies a very similar model to the POS tagging of Portuguese and English. (Luong et al., 2013 ) also descends lower than the word level, using a dictionary of morphemes and recursive neural networks to model the structure of the words. Similarly, this allows a better representation of rare and complex words, evaluated on a word similarity task.", |
| "cite_spans": [ |
| { |
| "start": 103, |
| "end": 122, |
| "text": "(Luong et al., 2013", |
| "ref_id": "BIBREF10" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Experiments are carried out on the Part-of-Speech and Morphological tagging tasks using the German corpus TIGER Treebank (Brants et al., 2002) . To the best of our knowledge, the best results on this task were published in (Mueller et al., 2013) , who applied a high-order CRF that includes an intensive feature engineering to five different languages. German was highlighted as having 'the most ambiguous morphology'. The corpus, de- (Mueller et al., 2013) . Simple and Struct. respectively denote the position-by-position and structured prediction. * indicates our best configuration.", |
| "cite_spans": [ |
| { |
| "start": 121, |
| "end": 142, |
| "text": "(Brants et al., 2002)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 223, |
| "end": 245, |
| "text": "(Mueller et al., 2013)", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 435, |
| "end": 457, |
| "text": "(Mueller et al., 2013)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments and Results", |
| "sec_num": "4" |
| }, |
| { |
| "text": "scribed in details in (Fraser et al., 2013) , contains a training set of 40472 sentences, a development and a test set of both 5000 sentences. We consider the two tagging tasks, with first a coarse tagset (54 tags), and then a morpho-syntactical rich tagset (619 items observed on the the training set).", |
| "cite_spans": [ |
| { |
| "start": 22, |
| "end": 43, |
| "text": "(Fraser et al., 2013)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments and Results", |
| "sec_num": "4" |
| }, |
| { |
| "text": "All the models are implemented 4 with the Theano library (Bergstra et al., 2010) . For optimization, we use Adagrad (Duchi et al., 2011) , with a learning rate of 0.1. The other hyperparameters are: the window sizes, d c and d w , respectively set to 5 and 9, the dimension of character embeddings, word embeddings and of the hidden layer, n c , n f and n h , that are respectively of 100, 200 and 200 5 . The models were trained on 7 epochs. Parameter initialization and corpus ordering are random, and the results presented are the average and standard deviation of the POS Tagging error rate over 5 runs.", |
| "cite_spans": [ |
| { |
| "start": 57, |
| "end": 80, |
| "text": "(Bergstra et al., 2010)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 116, |
| "end": 136, |
| "text": "(Duchi et al., 2011)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental settings", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "4 Implementation is available at https://github. com/MatthieuLabeau/NonlexNN 5 For both the learning rate and the embedding sizes, results does not differ in a significant way in a large range of hyperparameters, and their impact resides more in convergence speed and computation time", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental settings", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "The first experiment aims to evaluate the efficiency of a convolutional encoding with the basic feedforward architecture for prediction. We compare a completely non-lexicalized model which relies only on a character-level encoding with a lexicalized model where we use conventional word embeddings stored with a fixed vocabulary 6 . Results are reported in Table 1 along with with the state-of-the-art results published in (Mueller et al., 2013) . Results show that a character-level encoding yields better results than the conventional word-level encoding. Moreover, the structured inference allows the model to achieve accuracy reasonably close to the performance of a high-order CRF that uses handcrafted features. Finally, the model that uses the concatenation of both the character and word-level embeddings outperforms the state-of-the-art system on the more difficult task, without any feature engineering.", |
| "cite_spans": [ |
| { |
| "start": 423, |
| "end": 445, |
| "text": "(Mueller et al., 2013)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 357, |
| "end": 364, |
| "text": "Table 1", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "To give an idea of how a simple model would perform on such task, the reader can refer to (Schmid and Laws, 2008) and (Mueller et al., 2013) . For instance in the former, by choosing the most probable tag position-by-position, the error rate on the development set of the TIGER dataset is 32.7 for the simple POS Tagging task.", |
| "cite_spans": [ |
| { |
| "start": 90, |
| "end": 113, |
| "text": "(Schmid and Laws, 2008)", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 118, |
| "end": 140, |
| "text": "(Mueller et al., 2013)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "We further analyze the results by looking at the error rates respectively on known and unknown words 7 . From table 2, we observe that the number of unknown words wrongly labeled is divided by 3 for POS and almost divided by 2 for POS+Morph tagging, showing the ability of character-level encoding to generalize to new words. Moreover, a strictly non-lexical encoding makes slightly more mistakes on words already seen, whereas the model that concatenates both embeddings will make less mistakes for both unknown and known words.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "This shows that information from the context and from the morphology are complementary, which is conjectured in (Mueller et al., 2013) by using a morphological analyzer in complement of higher-order CRF. Table 2 : Error counts for known/unknown words in the test set, with a structured feedforward prediction model for the tagging task.", |
| "cite_spans": [ |
| { |
| "start": 112, |
| "end": 134, |
| "text": "(Mueller et al., 2013)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 204, |
| "end": 211, |
| "text": "Table 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "In the second set of experiments, we evaluate the convolutional encoding with a bidirectional recurrent network for prediction. Results are presented in the second half of Table 1 . Surprisingly, this architecture performs poorly with simple inference, but clearly improves when predicting a structured output using the Viterbi algorithm, both for training and testing. Moreover, a non-lexical model trained to infer a tag sequence with the Viterbi algorithm achieves results that are close to the state-of-the-art, thus validating our approach. We consider that this improvement comes from the synergy between using a global training objective with a global hidden representation, complexifying the model but allowing a more efficient solution. Finally, the model that uses the combination of both the character and word-level embeddings yields the best results. It is interesting to notice that the predictive architecture has no influence on the results of the simple task when the prediction is structured, but improves them on the difficult task. This also shows that the contribution of word embeddings to our model corresponds to a difference of 1.5 to 2 points in performance.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 172, |
| "end": 179, |
| "text": "Table 1", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "In this paper, we explored new models that can infer meaningful word representations from the raw character stream, allowing the model to exploit the morphological properties of words without using any handcrafted features or external tools. These models can therefore efficiently process words that were unseen in the training data. The evaluation was carried out on a POS and morphological tagging task for German. We described different architectures that act in two stages: the first stage is a convolutional network that infers a word representation directly from the character stream, while the second stage performs the prediction. For the prediction stage, we investigated different solutions showing that a bidirectional recurrent network can outperform state-of-the-art results when using a structured inference algorithm.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Our results showed that character-level encoding can address the unknown words problem for morphologically complex languages. In the future, we plan to extend these models to other tasks such as syntactic parsing and machine translation. Moreover, we will also investigate other architectures to infer word embeddings from the character level. For instance, preliminary experiments show that bidirectional recurrent network can achieve very competitive and promising results.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Besides inflected forms, German is characterized by a possibly infinite and evolving set of compound nouns.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "Similarly, we use special word tokens for padding.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "Every word that appears in the training set.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "Unknown words refer to words present in the development or test sets, but not in the training set.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "We would like to thank the anonymous reviewers for their helpful comments and suggestions. This work has been partly funded by the European Unions Horizon 2020 research and innovation programme under grant agreement No. 645452 (QT21).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgments", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "A neural probabilistic language model", |
| "authors": [ |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| }, |
| { |
| "first": "R\u00e9jean", |
| "middle": [], |
| "last": "Ducharme", |
| "suffix": "" |
| }, |
| { |
| "first": "Pascal", |
| "middle": [], |
| "last": "Vincent", |
| "suffix": "" |
| }, |
| { |
| "first": "Christian", |
| "middle": [], |
| "last": "Jauvin", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Journal of Machine Learning Research", |
| "volume": "3", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yoshua Bengio, R\u00e9jean Ducharme, Pascal Vincent, and Christian Jauvin. 2003. A neural probabilistic lan- guage model. Journal of Machine Learning Re- search, 3:1137 1155.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Theano: a CPU and GPU math expression compiler", |
| "authors": [ |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Bergstra", |
| "suffix": "" |
| }, |
| { |
| "first": "Olivier", |
| "middle": [], |
| "last": "Breuleux", |
| "suffix": "" |
| }, |
| { |
| "first": "Fr\u00e9d\u00e9ric", |
| "middle": [], |
| "last": "Bastien", |
| "suffix": "" |
| }, |
| { |
| "first": "Pascal", |
| "middle": [], |
| "last": "Lamblin", |
| "suffix": "" |
| }, |
| { |
| "first": "Razvan", |
| "middle": [], |
| "last": "Pascanu", |
| "suffix": "" |
| }, |
| { |
| "first": "Guillaume", |
| "middle": [], |
| "last": "Desjardins", |
| "suffix": "" |
| }, |
| { |
| "first": "Joseph", |
| "middle": [], |
| "last": "Turian", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Warde-Farley", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the Python for Scientific Computing Conference (SciPy)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "James Bergstra, Olivier Breuleux, Fr\u00e9d\u00e9ric Bastien, Pascal Lamblin, Razvan Pascanu, Guillaume Des- jardins, Joseph Turian, David Warde-Farley, and Yoshua Bengio. 2010. Theano: a CPU and GPU math expression compiler. In Proceedings of the Python for Scientific Computing Conference (SciPy), June. Oral Presentation.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "The TIGER treebank", |
| "authors": [ |
| { |
| "first": "Sabine", |
| "middle": [], |
| "last": "Brants", |
| "suffix": "" |
| }, |
| { |
| "first": "Stefanie", |
| "middle": [], |
| "last": "Dipper", |
| "suffix": "" |
| }, |
| { |
| "first": "Silvia", |
| "middle": [], |
| "last": "Hansen", |
| "suffix": "" |
| }, |
| { |
| "first": "Wolfgang", |
| "middle": [], |
| "last": "Lezius", |
| "suffix": "" |
| }, |
| { |
| "first": "George", |
| "middle": [], |
| "last": "Smith", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "Proceedings of the workshop on treebanks and linguistic theories", |
| "volume": "", |
| "issue": "", |
| "pages": "24--41", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sabine Brants, Stefanie Dipper, Silvia Hansen, Wolf- gang Lezius, and George Smith. 2002. The TIGER treebank. In Proceedings of the workshop on tree- banks and linguistic theories, pages 24-41.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Natural language processing (almost) from scratch", |
| "authors": [ |
| { |
| "first": "Ronan", |
| "middle": [], |
| "last": "Collobert", |
| "suffix": "" |
| }, |
| { |
| "first": "Jason", |
| "middle": [], |
| "last": "Weston", |
| "suffix": "" |
| }, |
| { |
| "first": "L\u00e9on", |
| "middle": [], |
| "last": "Bottou", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Karlen", |
| "suffix": "" |
| }, |
| { |
| "first": "Koray", |
| "middle": [], |
| "last": "Kavukcuoglu", |
| "suffix": "" |
| }, |
| { |
| "first": "Pavel", |
| "middle": [], |
| "last": "Kuksa", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "J. Mach. Learn. Res", |
| "volume": "12", |
| "issue": "", |
| "pages": "2493--2537", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ronan Collobert, Jason Weston, L\u00e9on Bottou, Michael Karlen, Koray Kavukcuoglu, and Pavel Kuksa. 2011. Natural language processing (almost) from scratch. J. Mach. Learn. Res., 12:2493-2537, November.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Deep learning for efficient discriminative parsing", |
| "authors": [ |
| { |
| "first": "Ronan", |
| "middle": [], |
| "last": "Collobert", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics", |
| "volume": "", |
| "issue": "", |
| "pages": "224--232", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ronan Collobert. 2011. Deep learning for efficient discriminative parsing. In Proceedings of the Four- teenth International Conference on Artificial Intel- ligence and Statistics, AISTATS 2011, Fort Laud- erdale, USA, April 11-13, 2011, pages 224-232.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Adaptive subgradient methods for online learning and stochastic optimization", |
| "authors": [ |
| { |
| "first": "John", |
| "middle": [], |
| "last": "Duchi", |
| "suffix": "" |
| }, |
| { |
| "first": "Elad", |
| "middle": [], |
| "last": "Hazan", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoram", |
| "middle": [], |
| "last": "Singer", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "J. Mach. Learn. Res", |
| "volume": "12", |
| "issue": "", |
| "pages": "2121--2159", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "John Duchi, Elad Hazan, and Yoram Singer. 2011. Adaptive subgradient methods for online learning and stochastic optimization. J. Mach. Learn. Res., 12:2121-2159, July.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Knowledge sources for constituent parsing of German, a morphologically rich and less-configurational language", |
| "authors": [ |
| { |
| "first": "Alexander", |
| "middle": [], |
| "last": "Fraser", |
| "suffix": "" |
| }, |
| { |
| "first": "Helmut", |
| "middle": [], |
| "last": "Schmid", |
| "suffix": "" |
| }, |
| { |
| "first": "Rich\u00e1rd", |
| "middle": [], |
| "last": "Farkas", |
| "suffix": "" |
| }, |
| { |
| "first": "Renjing", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Hinrich", |
| "middle": [], |
| "last": "Sch\u00fctze", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Comput. Linguist", |
| "volume": "39", |
| "issue": "1", |
| "pages": "57--85", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alexander Fraser, Helmut Schmid, Rich\u00e1rd Farkas, Renjing Wang, and Hinrich Sch\u00fctze. 2013. Knowl- edge sources for constituent parsing of German, a morphologically rich and less-configurational lan- guage. Comput. Linguist., 39(1):57-85, March.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Hybrid speech recognition with deep bidirectional LSTM", |
| "authors": [ |
| { |
| "first": "Alex", |
| "middle": [], |
| "last": "Graves", |
| "suffix": "" |
| }, |
| { |
| "first": "Navdeep", |
| "middle": [], |
| "last": "Jaitly", |
| "suffix": "" |
| }, |
| { |
| "first": "Abdel-Rahman", |
| "middle": [], |
| "last": "Mohamed", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "IEEE Workshop on Automatic Speech Recognition and Understanding", |
| "volume": "", |
| "issue": "", |
| "pages": "273--278", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alex Graves, Navdeep Jaitly, and Abdel-rahman Mo- hamed. 2013. Hybrid speech recognition with deep bidirectional LSTM. In 2013 IEEE Workshop on Automatic Speech Recognition and Understanding, Olomouc, Czech Republic, December 8-12, 2013, pages 273-278.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Generating sequences with recurrent neural networks", |
| "authors": [ |
| { |
| "first": "Alex", |
| "middle": [], |
| "last": "Graves", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alex Graves. 2013. Generating sequences with recur- rent neural networks. CoRR, abs/1308.0850.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Visualizing and understanding recurrent networks", |
| "authors": [ |
| { |
| "first": "Andrej", |
| "middle": [], |
| "last": "Karpathy", |
| "suffix": "" |
| }, |
| { |
| "first": "Justin", |
| "middle": [], |
| "last": "Johnson", |
| "suffix": "" |
| }, |
| { |
| "first": "Fei-Fei", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Andrej Karpathy, Justin Johnson, and Fei-Fei Li. 2015. Visualizing and understanding recurrent networks. CoRR, abs/1506.02078.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Better word representations with recursive neural networks for morphology", |
| "authors": [ |
| { |
| "first": "Thang", |
| "middle": [], |
| "last": "Luong", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Socher", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [ |
| "D" |
| ], |
| "last": "Manning", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the Seventeenth Conference on Computational Natural Language Learning", |
| "volume": "", |
| "issue": "", |
| "pages": "104--113", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Thang Luong, Richard Socher, and Christopher D. Manning. 2013. Better word representations with recursive neural networks for morphology. In Pro- ceedings of the Seventeenth Conference on Compu- tational Natural Language Learning, CoNLL 2013, Sofia, Bulgaria, August 8-9, 2013, pages 104-113.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Subword language modeling with neural networks", |
| "authors": [ |
| { |
| "first": "Tomas", |
| "middle": [], |
| "last": "Mikolov", |
| "suffix": "" |
| }, |
| { |
| "first": "Ilya", |
| "middle": [], |
| "last": "Sutskever", |
| "suffix": "" |
| }, |
| { |
| "first": "Anoop", |
| "middle": [], |
| "last": "Deoras", |
| "suffix": "" |
| }, |
| { |
| "first": "Hai-Son", |
| "middle": [], |
| "last": "Le", |
| "suffix": "" |
| }, |
| { |
| "first": "Stefan", |
| "middle": [], |
| "last": "Kombrink", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tomas Mikolov, Ilya Sutskever, Anoop Deoras, Hai- Son Le, Stefan Kombrink, and Jan Cernocky. 2012. Subword language modeling with neural networks. Unpublished.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Efficient higher-order CRFs for morphological tagging", |
| "authors": [ |
| { |
| "first": "Thomas", |
| "middle": [], |
| "last": "Mueller", |
| "suffix": "" |
| }, |
| { |
| "first": "Helmut", |
| "middle": [], |
| "last": "Schmid", |
| "suffix": "" |
| }, |
| { |
| "first": "Hinrich", |
| "middle": [], |
| "last": "Sch\u00fctze", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "322--332", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Thomas Mueller, Helmut Schmid, and Hinrich Sch\u00fctze. 2013. Efficient higher-order CRFs for morphological tagging. In Proceedings of the 2013 Conference on Empirical Methods in Natural Lan- guage Processing, pages 322-332, Seattle, Wash- ington, USA, October. Association for Computa- tional Linguistics.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Learning character-level representations for part-ofspeech tagging", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Cicero", |
| "suffix": "" |
| }, |
| { |
| "first": "Bianca", |
| "middle": [], |
| "last": "Santos", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Zadrozny", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 31st International Conference on Machine Learning (ICML-14)", |
| "volume": "", |
| "issue": "", |
| "pages": "1818--1826", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Cicero D. Santos and Bianca Zadrozny. 2014. Learning character-level representations for part-of- speech tagging. In Tony Jebara and Eric P. Xing, ed- itors, Proceedings of the 31st International Confer- ence on Machine Learning (ICML-14), pages 1818- 1826. JMLR Workshop and Conference Proceed- ings.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Estimation of conditional probabilities with decision trees and an application to fine-grained pos tagging", |
| "authors": [ |
| { |
| "first": "Helmut", |
| "middle": [], |
| "last": "Schmid", |
| "suffix": "" |
| }, |
| { |
| "first": "Florian", |
| "middle": [], |
| "last": "Laws", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proceedings of the 22Nd International Conference on Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "777--784", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Helmut Schmid and Florian Laws. 2008. Estima- tion of conditional probabilities with decision trees and an application to fine-grained pos tagging. In Proceedings of the 22Nd International Conference on Computational Linguistics -Volume 1, COLING '08, pages 777-784, Stroudsburg, PA, USA. Associ- ation for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Generating text with recurrent neural networks", |
| "authors": [ |
| { |
| "first": "Ilya", |
| "middle": [], |
| "last": "Sutskever", |
| "suffix": "" |
| }, |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Martens", |
| "suffix": "" |
| }, |
| { |
| "first": "Geoffrey", |
| "middle": [], |
| "last": "Hinton", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the 28th International Conference on Machine Learning (ICML-11), ICML '11", |
| "volume": "", |
| "issue": "", |
| "pages": "1017--1024", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ilya Sutskever, James Martens, and Geoffrey Hinton. 2011. Generating text with recurrent neural net- works. In Lise Getoor and Tobias Scheffer, editors, Proceedings of the 28th International Conference on Machine Learning (ICML-11), ICML '11, pages 1017-1024, New York, NY, USA, June. ACM.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Error bounds for convolutional codes and an asymptotically optimum decoding algorithm", |
| "authors": [ |
| { |
| "first": "Andrew", |
| "middle": [], |
| "last": "Viterbi", |
| "suffix": "" |
| } |
| ], |
| "year": 1967, |
| "venue": "IEEE Trans. Inf. Theor", |
| "volume": "13", |
| "issue": "2", |
| "pages": "260--269", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Andrew Viterbi. 1967. Error bounds for convolutional codes and an asymptotically optimum decoding al- gorithm. IEEE Trans. Inf. Theor., 13(2):260-269, April.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Readings in Speech Recognition, chapter Phoneme Recognition Using Time-delay Neural Networks", |
| "authors": [ |
| { |
| "first": "Alexander", |
| "middle": [], |
| "last": "Waibel", |
| "suffix": "" |
| }, |
| { |
| "first": "Toshiyuki", |
| "middle": [], |
| "last": "Hanazawa", |
| "suffix": "" |
| }, |
| { |
| "first": "Geofrey", |
| "middle": [], |
| "last": "Hinton", |
| "suffix": "" |
| }, |
| { |
| "first": "Kiyohiro", |
| "middle": [], |
| "last": "Shikano", |
| "suffix": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [ |
| "J" |
| ], |
| "last": "Lang", |
| "suffix": "" |
| } |
| ], |
| "year": 1990, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "393--404", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alexander Waibel, Toshiyuki Hanazawa, Geofrey Hin- ton, Kiyohiro Shikano, and Kevin J. Lang, 1990. Readings in Speech Recognition, chapter Phoneme Recognition Using Time-delay Neural Networks, pages 393-404. Morgan Kaufmann Publishers Inc., San Francisco, CA, USA.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Text understanding from scratch", |
| "authors": [ |
| { |
| "first": "Xiang", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "Yann", |
| "middle": [], |
| "last": "Lecun", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Xiang Zhang and Yann LeCun. 2015. Text understand- ing from scratch. CoRR, abs/1502.01710.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "uris": null, |
| "num": null, |
| "type_str": "figure", |
| "text": "Architecture of the layer for characterlevel encoding of words." |
| }, |
| "FIGREF1": { |
| "uris": null, |
| "num": null, |
| "type_str": "figure", |
| "text": "Bidirectional recurrent architecture for tag prediction. The upper part is used in the case of structured inference." |
| }, |
| "TABREF0": { |
| "type_str": "table", |
| "content": "<table><tr><td/><td/><td/><td colspan=\"2\">POS</td><td>POS+Morph</td></tr><tr><td colspan=\"3\">Architecture Encoding Output</td><td>Dev</td><td>Test</td><td>Dev</td><td>Test</td></tr><tr><td>Feedforward</td><td>Lex.</td><td colspan=\"4\">Simple 4.22 \u00b1 0.05 5.89 Struct. 2.92 \u00b1 0.02 3.82 \u00b1 0.04 11.65 \u00b1 0.11 14.43 \u00b1 0.19</td></tr><tr><td/><td>Both</td><td colspan=\"4\">Simple 2.59 \u00b1 0.05 Struct. 2.22 \u00b1 0.03 * 2.86 \u00b1 0.03 * 9.11 \u00b1 0.14 3.34 \u00b1 0.09 11.89 \u00b1 0.14 14.63 \u00b1 0.22 11.29 \u00b1 0.06</td></tr><tr><td/><td>Lex</td><td colspan=\"2\">Simple 6.03 \u00b1 0.06 Struct. 3.89 \u00b1 0.06</td><td colspan=\"2\">8.05 \u00b1 0.05 17.83 \u00b1 0.11 21.33 \u00b1 0.26 5.26 \u00b1 0.05 11.88 \u00b1 0.05 17.78 \u00b1 0.12</td></tr><tr><td>biRNN</td><td>Non-Lex</td><td colspan=\"2\">Simple 4.46 \u00b1 0.08 Struct. 2.74 \u00b1 0.07</td><td colspan=\"2\">5.84 \u00b1 0.19 16.61 \u00b1 0.18 19.39 \u00b1 0.12 3.59 \u00b1 0.07 10.09 \u00b1 0.09 12.88 \u00b1 0.28</td></tr><tr><td/><td colspan=\"3\">Simple 3.63 \u00b1 0.06 Struct. 2.21 \u00b1 0.04 CRF Both 2.06</td><td colspan=\"2\">4.63 \u00b1 0.04 14.83 \u00b1 0.11 17.54 \u00b1 0.13 2.56 9.40 11.42</td></tr></table>", |
| "html": null, |
| "num": null, |
| "text": "\u00b1 0.07 13.97 \u00b1 0.14 17.46 \u00b1 0.14 Struct. 3.90 \u00b1 0.05 5.33 \u00b1 0.09 12.22 \u00b1 0.13 15.34 \u00b1 0.13 Non-lex. Simple 3.31 \u00b1 0.07 4.22 \u00b1 0.07 13.50 \u00b1 0.16 16.23 \u00b1 0.13 2.86 \u00b1 0.05 * 8.63 \u00b1 0.21 * 10.97 \u00b1 0.19 *" |
| }, |
| "TABREF1": { |
| "type_str": "table", |
| "content": "<table/>", |
| "html": null, |
| "num": null, |
| "text": "Comparison of the feedforward and bidirectional recurrent architectures for predictions, with different settings. The non-lexical encoding is convolutional. CRF refers to state-of-the-art system of" |
| } |
| } |
| } |
| } |