| { |
| "paper_id": "K18-1045", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T07:10:39.868014Z" |
| }, |
| "title": "Bringing Order to Neural Word Embeddings with Embeddings Augmented by Random Permutations (EARP)", |
| "authors": [ |
| { |
| "first": "Trevor", |
| "middle": [], |
| "last": "Cohen", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Biomedical and Health Informatics University of Washington", |
| "location": { |
| "settlement": "Seattle" |
| } |
| }, |
| "email": "cohenta@uw.edu" |
| }, |
| { |
| "first": "Dominic", |
| "middle": [], |
| "last": "Widdows", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "dominic.widdows@grab.com" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "Word order is clearly a vital part of human language, but it has been used comparatively lightly in distributional vector models. This paper presents a new method for incorporating word order information into word vector embedding models by combining the benefits of permutation-based order encoding with the more recent method of skip-gram with negative sampling. The new method introduced here is called Embeddings Augmented by Random Permutations (EARP). It operates by applying permutations to the coordinates of context vector representations during the process of training. Results show an 8% improvement in accuracy on the challenging Bigger Analogy Test Set, and smaller but consistent improvements on other analogy reference sets. These findings demonstrate the importance of order-based information in analogical retrieval tasks, and the utility of random permutations as a means to augment neural embeddings.", |
| "pdf_parse": { |
| "paper_id": "K18-1045", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "Word order is clearly a vital part of human language, but it has been used comparatively lightly in distributional vector models. This paper presents a new method for incorporating word order information into word vector embedding models by combining the benefits of permutation-based order encoding with the more recent method of skip-gram with negative sampling. The new method introduced here is called Embeddings Augmented by Random Permutations (EARP). It operates by applying permutations to the coordinates of context vector representations during the process of training. Results show an 8% improvement in accuracy on the challenging Bigger Analogy Test Set, and smaller but consistent improvements on other analogy reference sets. These findings demonstrate the importance of order-based information in analogical retrieval tasks, and the utility of random permutations as a means to augment neural embeddings.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "The recognition of the utility of corpus-derived distributed representations of words for a broad range of Natural Language Processing (NLP) tasks (Collobert et al., 2011) has led to a resurgence of interest in methods of distributional semantics. In particular, the neural-probabilistic word representations produced by the Skip-gram and Continuous Bag-of-Words (CBOW) architectures (Mikolov et al., 2013a) implemented in the word2vec and fastText software packages have been extensively evaluated in recent years.", |
| "cite_spans": [ |
| { |
| "start": 147, |
| "end": 171, |
| "text": "(Collobert et al., 2011)", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 384, |
| "end": 407, |
| "text": "(Mikolov et al., 2013a)", |
| "ref_id": "BIBREF25" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "As was the case with preceding distributional models (see for example (Sch\u00fctze, 1993; Lund and Burgess, 1996; Sch\u00fctze, 1998; Karlgren and Sahlgren, 2001 )), these architectures generate vector representations of words -or word embeddings -such that words with similar proximal neighboring terms within a corpus of text will have similar vector representations. As the relative position of these neighboring terms is generally not considered, distributional models of this nature are often (and sometimes derisively) referred to as bag-of-words models. While methods of encoding word order into neural-probabilistic representations have been evaluated, these methods generally require learning additional parameters, either for each position in the sliding window (Mnih and Kavukcuoglu, 2013) , or for each context word-byposition pair (Trask et al., 2015; Ling et al., 2015) .", |
| "cite_spans": [ |
| { |
| "start": 70, |
| "end": 85, |
| "text": "(Sch\u00fctze, 1993;", |
| "ref_id": "BIBREF35" |
| }, |
| { |
| "start": 86, |
| "end": 109, |
| "text": "Lund and Burgess, 1996;", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 110, |
| "end": 124, |
| "text": "Sch\u00fctze, 1998;", |
| "ref_id": "BIBREF36" |
| }, |
| { |
| "start": 125, |
| "end": 152, |
| "text": "Karlgren and Sahlgren, 2001", |
| "ref_id": "BIBREF20" |
| }, |
| { |
| "start": 763, |
| "end": 791, |
| "text": "(Mnih and Kavukcuoglu, 2013)", |
| "ref_id": "BIBREF29" |
| }, |
| { |
| "start": 835, |
| "end": 855, |
| "text": "(Trask et al., 2015;", |
| "ref_id": "BIBREF39" |
| }, |
| { |
| "start": 856, |
| "end": 874, |
| "text": "Ling et al., 2015)", |
| "ref_id": "BIBREF23" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this paper we evaluate an alternative method of encoding the position of words within a sliding window into neural word embeddings using Embeddings Augmented by Random Permutations (EARP). EARP leverages random permutations of context vector representations (Sahlgren et al., 2008) , a technique that has not been applied during the training of neural word embeddings previously. Unlike prior approaches to encoding position into neural word embeddings, it imposes no computational and negligible space requirements.", |
| "cite_spans": [ |
| { |
| "start": 261, |
| "end": 284, |
| "text": "(Sahlgren et al., 2008)", |
| "ref_id": "BIBREF34" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Results show that the word order information encoded through EARP leads to a nearly 8% improvement (from 29.17% to 37.07%) in the accuracy of analogy predictions in the Bigger Analogy Test Set of , with the improvement being largest (over 20%) in the category of analogies that exhibit derivational morphology (a case where the use of subword information also improves accuracy for all representations with and without order information). Smaller improvements in performance are evident on other analogy sets, and in downstream sequence labeling tasks. This makes EARP a strong contender for enriching word embeddings with order-based information, leading to greater accuracy on more challenging semantic processing tasks.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Methods of distributional semantics learn representations of words from the contexts they occur in across large text corpora, such that words that occur in similar contexts will have similar representations (Turney and Pantel, 2010) . Geometricallymotivated approaches to this problem often involve decomposition of a term-by-context matrix (Sch\u00fctze, 1993; Landauer and Dumais, 1997; Pennington et al., 2014) , resulting in word vectors of considerably lower dimensionality than the total number of contexts. Alternatively, reduceddimensional representations can be generated online while processing individual units of text, without the need to represent a large term-by-context matrix explicitly. A seminal example of the latter approach is the Random Indexing (RI) method (Kanerva et al., 2000) , which generates distributed representations of words by superposing randomly generated context vector representations.", |
| "cite_spans": [ |
| { |
| "start": 207, |
| "end": 232, |
| "text": "(Turney and Pantel, 2010)", |
| "ref_id": "BIBREF40" |
| }, |
| { |
| "start": 341, |
| "end": 356, |
| "text": "(Sch\u00fctze, 1993;", |
| "ref_id": "BIBREF35" |
| }, |
| { |
| "start": 357, |
| "end": 383, |
| "text": "Landauer and Dumais, 1997;", |
| "ref_id": "BIBREF21" |
| }, |
| { |
| "start": 384, |
| "end": 408, |
| "text": "Pennington et al., 2014)", |
| "ref_id": "BIBREF30" |
| }, |
| { |
| "start": 775, |
| "end": 797, |
| "text": "(Kanerva et al., 2000)", |
| "ref_id": "BIBREF18" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background 2.1 Distributed representations of words", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Neural-probabilistic methods have shown utility as a means to generate semantic vector representations of words (Bengio et al., 2003) . In particular, the Skip-gram and CBOW neural network architectures (Mikolov et al., 2013a) implemented within the word2vec and fastText software packages provide scalable approaches to online training of word vector representations that have been shown to perform well across a number of tasks (Mikolov et al., 2013a; Levy et al., 2015; .", |
| "cite_spans": [ |
| { |
| "start": 112, |
| "end": 133, |
| "text": "(Bengio et al., 2003)", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 203, |
| "end": 226, |
| "text": "(Mikolov et al., 2013a)", |
| "ref_id": "BIBREF25" |
| }, |
| { |
| "start": 430, |
| "end": 453, |
| "text": "(Mikolov et al., 2013a;", |
| "ref_id": "BIBREF25" |
| }, |
| { |
| "start": 454, |
| "end": 472, |
| "text": "Levy et al., 2015;", |
| "ref_id": "BIBREF22" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background 2.1 Distributed representations of words", |
| "sec_num": "2" |
| }, |
| { |
| "text": "While the definition of what constitutes a context varies across models, a popular alternative is to use words in a sliding window centered on a focus term for this purpose. Consequently where decomposition is involved the matrix in question would be a term-by-term matrix. With online methods, each term has two vector representations -a semantic vector and a context vector, corresponding to the input and output weights for each word in neural-probabilistic approaches.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background 2.1 Distributed representations of words", |
| "sec_num": "2" |
| }, |
| { |
| "text": "In most word vector embedding models based on sliding windows, the relative position of words within this sliding window is ignored, but there have been prior efforts to encode this information. Before the popularization of neural word embeddings, a number of researchers developed and evaluated methods to encode word position into distributed vector representations. The BEA-GLE model (Jones et al., 2006) uses circular convolution -as described by Plate (1995) -as a binding operator to compose representations of ngrams from randomly instantiated context vector representations of individual terms. These composite representations are then added to the semantic vector representation for the central term in a sliding window. The cosine similarity between the resulting vectors is predictive of human performance in semantic priming experiments.", |
| "cite_spans": [ |
| { |
| "start": 387, |
| "end": 407, |
| "text": "(Jones et al., 2006)", |
| "ref_id": "BIBREF17" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Order-based distributional models", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "A limitation of this approach is that it involves a large number of operations per sliding window. For example, Jones and his colleagues (2006) employ eight superposition and nine convolution operations to represent a single sliding window of three terms (excluding the focus term). Sahlgren, Holst and Kanerva (2008) report a computationally simpler method of encoding word order in the context of RI. Like BEAGLE, this approach involves adding randomly instantiated context vectors for adjacent terms to the semantic vector of the central term in a sliding window. However, RI uses sparse context vectors, consisting of mostly zeroes with a small number non-zero components initialized at random. These vectors are assigned permutations indicating their position within a sliding window. For example, with p representing the permutation assigned to a given position p and words to the right of the equation representing their context vectors, the sliding window \"[fast is fine but accuracy] is everything\" would result in the following update to S(fine), the semantic vector for the term \"fine\":", |
| "cite_spans": [ |
| { |
| "start": 283, |
| "end": 317, |
| "text": "Sahlgren, Holst and Kanerva (2008)", |
| "ref_id": "BIBREF34" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Order-based distributional models", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "S(fine) += \u22122 \u2212 \u2212 \u2192 fast + \u22121 \u2212 \u2192 is + +1 \u2212\u2192 but + +2 \u2212 \u2212\u2212\u2212\u2212\u2212 \u2192 accuracy", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Order-based distributional models", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Amongst the encoding schemes evaluated using this approach, using a pair of permutations to distinguish between words preceding the focus term and those following it resulted in the best performance on synonym test evaluations (Sahlgren et al., 2008) . Furthermore, this approach was shown to outperform BEAGLE on a number of evaluations on account of its ability to scale up to larger amounts of input data (Recchia et al., 2010) .", |
| "cite_spans": [ |
| { |
| "start": 227, |
| "end": 250, |
| "text": "(Sahlgren et al., 2008)", |
| "ref_id": "BIBREF34" |
| }, |
| { |
| "start": 408, |
| "end": 430, |
| "text": "(Recchia et al., 2010)", |
| "ref_id": "BIBREF32" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Order-based distributional models", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Some recent work has evaluated the utility of encoding word order into neural word embeddings.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Encoding order into neural embeddings", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "A straightforward way to accomplish this involves maintaining a separate set of parameters for each context word position, so that for a vocabulary of size v and p context window positions the number of output weights in the model is v \u00d7 p \u00d7 d for embeddings of dimensionality d. This approach was applied by Ling and his colleagues (2015) to distinguish between terms occurring before and after the central term in a sliding window, with improvements in performance in downstream partof-speech tagging and dependency parsing tasks.", |
| "cite_spans": [ |
| { |
| "start": 309, |
| "end": 339, |
| "text": "Ling and his colleagues (2015)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Encoding order into neural embeddings", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "Trask and his colleagues (2015) develop this idea further in a model they call a Partitioned Embedding Neural Network (PENN). In a PENN, both the input weights (word embeddings) and output weights (context embeddings) of the network have separate position-specific instantiations, so the total number of model parameters is 2v \u00d7 p \u00d7 d. In addition to evaluating binary (before/after) context positions, the utility of training separate weights for each position within a sliding window was evaluated. Incorporating order in this way resulted in improvements in accuracy over word2vec's CBOW implementation on proportional analogy problems, which were considerably enhanced by the incorporation of character-level embeddings.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Encoding order into neural embeddings", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "A more space-efficient approach to encoding word order involves learning a vector V for each position p in the sliding window. These vectors are applied to rescale context vector representations using pointwise multiplication while constructing a weighted average of the context vectors for each term in the window (Mnih and Kavukcuoglu, 2013; . Results reported using this approach are variable, with some authors reporting worse performance with positiondependent weights on a sentence completion task (Mnih and Kavukcuoglu, 2013) , and others reporting improved performance on an analogy completion task .", |
| "cite_spans": [ |
| { |
| "start": 315, |
| "end": 343, |
| "text": "(Mnih and Kavukcuoglu, 2013;", |
| "ref_id": "BIBREF29" |
| }, |
| { |
| "start": 504, |
| "end": 532, |
| "text": "(Mnih and Kavukcuoglu, 2013)", |
| "ref_id": "BIBREF29" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Encoding order into neural embeddings", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "In the current work, we extend the skipgram-withnegative-sampling (SGNS) algorithm to encode positional information without additional computation. The Skip-gram model (Mikolov et al., 2013a) predicts p(c|w): the probability of observing a context word, c, given an observed word, w. This can be accomplished by moving a slid-ing window through a large text corpus, such that the observed word is at the center of the window, and the context words surround it. The architecture itself includes two sets of parameters for each unique word: The input weights of the network ( \u2212 \u2192 w ), which represent observed words and are usually retained as semantic vectors after training, and the output weights of the network which represent context words ( \u2212 \u2192 c ) and are usually discarded.", |
| "cite_spans": [ |
| { |
| "start": 168, |
| "end": 191, |
| "text": "(Mikolov et al., 2013a)", |
| "ref_id": "BIBREF25" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Skipgram-with-negative-sampling", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "The probability of observing a word in context is calculated by appying the sigmoid function to the scalar product between the input weights for the observed word, and the output weights for the context word, such that", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Skipgram-with-negative-sampling", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "p(c|w) = \u03c3( \u2212 \u2192 w . \u2212 \u2192 c ).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Skipgram-with-negative-sampling", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "As maximizing this probability using the softmax over all possible context words would be computationally inconvenient, SGNS instead draws a small number of negative samples (\u00acc) from the vocabulary to serve as counterexamples to each observed context term. With D representing observed term/context pairs, and D representing randomly constructed counterexamples the SGNS optimization objective is as follows (Goldberg and Levy, 2014) :", |
| "cite_spans": [ |
| { |
| "start": 409, |
| "end": 434, |
| "text": "(Goldberg and Levy, 2014)", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Skipgram-with-negative-sampling", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "(w,c)\u2208D log \u03c3( \u2212 \u2192 w . \u2212 \u2192 c ) + (w,\u00acc)\u2208D log \u03c3(\u2212 \u2212 \u2192 w . \u2212 \u2192 \u00acc)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Skipgram-with-negative-sampling", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "Sliding window variants of RI are similar in some respects to the SGNS algorithm, in that training occurs through an online update step in which a vector representation for each context term is added to the semantic vector representation for a focus term (with SGNS this constitutes an update of the input weight vector for the focus term, weighted by the gradient and learning rate (for a derivation see (Rong, 2014) ). Unlike SGNS, however, context vectors in RI are immutable and sparse. With SGNS, dense context vectors are altered during the training process, providing an enhanced capacity for inferred similarity. Nonetheless, the technique of permuting context vectors to indicate the position of a context term within a sliding window is readily adaptable to SGNS. Our approach to encoding the position of context words involves assigning a randomlygenerated permutation to each position within a sliding window. For example, with a sliding window of window radius 2 (considering two positions to the left and the right of a focus term), we assign a random permutation, p , to each element ", |
| "cite_spans": [ |
| { |
| "start": 405, |
| "end": 417, |
| "text": "(Rong, 2014)", |
| "ref_id": "BIBREF33" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Embeddings Augmented by Random Permutations (EARP)", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "(w,c:p)\u2208D log \u03c3( \u2212 \u2192 w . p ( \u2212 \u2192 c )) + (w,\u00acc:p)\u2208D log \u03c3( \u2212 \u2212 \u2192 w . p ( \u2212 \u2192 \u00acc))", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Embeddings Augmented by Random Permutations (EARP)", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "For example, upon observing the term \"wyatt\" in the context of the term \"earp\", the model would attempt to maximize", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Embeddings Augmented by Random Permutations (EARP)", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "p(c|w) = \u03c3( \u2212 \u2212\u2212 \u2192 wyatt. +1 ( \u2212 \u2212 \u2192 earp))", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Embeddings Augmented by Random Permutations (EARP)", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": ", with \u2212 \u2212\u2212 \u2192 wyatt and \u2212 \u2212 \u2192 earp as semantic and context vectors respectively. The underlying architecture is illustrated in Figure 1 , using a simplified model generating 5-dimensional vector representations for a small vocabulary of 10 terms. The permutation p can be implemented by \"rewiring\" the components of the input and output weights, without explicitly generating a permuted copy of the context vector concerned. In this way, p(c|w) is estimated and maximized in place without imposing additional computational or space requirements, beyond those required to store the permutations. This is accomplished by changing the index values used to access components of \u2212 \u2212 \u2192 earp when the scalar product is calculated, and when weights are updated. The inverse permutation (or reverse rewiring -connecting components 1:3 rather than 3:1) is applied to \u2212 \u2212\u2212 \u2192 wyatt when updating \u2212 \u2212 \u2192 earp, and this procedure is used with both observed context terms and negative samples.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 127, |
| "end": 135, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Embeddings Augmented by Random Permutations (EARP)", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Within this general framework, we evaluate four word order encoding schemes, implemented by adapting the open source Semantic Vectors 1 package for distributional semantics research:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Embeddings Augmented by Random Permutations (EARP)", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "1 https://github.com/semanticvectors/semanticvectors 3.2.1 Directional (EARP dir ) Directional encoding draws a distinction between terms that occur before or after the focus term in a sliding window. As such, only two permutations (and their inverse permutations) are employed:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Embeddings Augmented by Random Permutations (EARP)", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "\u22121 for preceding terms, and +1 for all subseqent terms. Directional encoding with permutations has been shown to improve performance in synonym tests evaluations when applied in the context of RI (Sahlgren et al., 2008) .", |
| "cite_spans": [ |
| { |
| "start": 196, |
| "end": 219, |
| "text": "(Sahlgren et al., 2008)", |
| "ref_id": "BIBREF34" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Embeddings Augmented by Random Permutations (EARP)", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "With positional encoding, a permutation is used to encode each space in the sliding window. As a randomly permuted vector is highly likely to be orthogonal or close-to-orthogonal to the vector from which it originated (Kanerva, 2009) , this is likely to result in orthogonal encodings for the same context word in different positions. With RI, positional encoding degraded synonym test performance, but permitted a novel form of distributional query, in which permutation is used to retrieve words that occur in particular positions in relation to one another (Sahlgren et al., 2008) . EARP pos facilitates queries of this form also: the nearest neighboring context vector to the permuted semantic vector INV \u22121 ( \u2212 \u2212 \u2192 earp) is \u2212 \u2212\u2212 \u2192 wyatt in both static-window subword-agnostic EARP pos spaces used in the experiments that follow.", |
| "cite_spans": [ |
| { |
| "start": 218, |
| "end": 233, |
| "text": "(Kanerva, 2009)", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 560, |
| "end": 583, |
| "text": "(Sahlgren et al., 2008)", |
| "ref_id": "BIBREF34" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Positional (EARP pos )", |
| "sec_num": "3.2.2" |
| }, |
| { |
| "text": "With proximity-based encoding, the positional encoding of a particular context term occurring in the first position in a sliding window will be somewhat similar to the encoding when this term occurs in the second position, and less similar (but still not orthogonal) to the encoding when it occurs in the third. This is accomplished by randomly generating an index permutation +1 , randomly reassigning a half of its permutations to generate +2 , and repeating this process iteratively until a permutation for every position in the window is obtained (for the current experiments, we assigned two index permutations +1 and \u22121 , proceeding bidirectionally). As a low-dimensional example, if +1 were {4:1, 1:2, 2:3, 3:4}, +2 might be {4:3, 1:2, 2:1, 3:4}. The net result is that the similarity between the position-specific representations of a given context vector reflects the proximity between the positions concerned. While this method is reminiscent of interpolation between randomly generated vectors or matrices to encode character position within words (Cohen et al., 2013) and pixel position within images (Gallant and Culliton, 2016) respectively, the iterative application of permutations for this purpose is a novel approach to such positional binding.", |
| "cite_spans": [ |
| { |
| "start": 1059, |
| "end": 1079, |
| "text": "(Cohen et al., 2013)", |
| "ref_id": "BIBREF8" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Proximity-based (EARP prox )", |
| "sec_num": "3.2.3" |
| }, |
| { |
| "text": "The use of character n-grams as components of distributional semantic models was introduced by Sch\u00fctze (1993) , and has been shown to improve performance of neural-probabilistic models on analogical retrieval tasks . It is intuitive that this should be the case as standard analogy evaluation reference sets include many analogical questions that require mapping from a morphological derivative of one word (e.g. fast:faster) to the same morphological derivative of another (e.g. high:higher).", |
| "cite_spans": [ |
| { |
| "start": 95, |
| "end": 109, |
| "text": "Sch\u00fctze (1993)", |
| "ref_id": "BIBREF35" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Subword embeddings", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "Consequently, we also generated n-gram based variants of each of our models, by adapting the approach described in to our SGNS configuration. Specifically, we decomposed each word into character n-grams, after introducing characters indicating the start (<) and end (>) of a word. N-grams of size betweeen 3 and 6 characters (inclusive) were encoded, and in order to place an upper bound on memory requirements a hash function was used to map each observed n-gram to one of at the most two million vectors without constraints on collisions. The input vector V i for a word with n included n-grams (including the word itself) was then generated as 2", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Subword embeddings", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "V i = 1 n n i=1 \u2212 \u2212\u2212\u2212 \u2192 ngram i", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Subword embeddings", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "During training, we used V i as the input vector for EARP and SGNS, with updates propagating back to component word and n-gram vectors in proportion to their contribution to V i .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Subword embeddings", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "All models were trained on the first January 2018 release of the English Wikipedia 3 , to which we applied the pre-processing script distributed with the fastText package 4 , resulting in a corpus of approximately 8.8 billion words.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Training data", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "All models used 500-dimensional vectors, and were trained for a single iteration across the corpus with five negative samples per context term. For each of the four models (SGNS , EARP dir , EARP pos , EARP prox ), embeddings were generated with sliding window radii r of 2 and 5 (in each direction), with and without subword embeddings. For all experiments, we excluded numbers, and terms occurring less than 150 times in the corpus. SGNS has a number of hyperparameters that are known to influence performance (Levy et al., 2015) . We did not engage in tuning of these hyperparameters to improve performance of our models, but rather were guided by prior research in selecting hyper-parameter settings that are known to perform well with the SGNS baseline model. Specifically, we used a subsampling threshold t of 10 \u22125 . In some experiments (EARP and SGNS), we used dynamic sliding windows with uniform probability of a sliding window radius between one and r, in an effort to align our baseline model closely with other SGNS implementations. Stochastic reduction of the sliding window radius will result in distal words being ignored at times. With a dynamic sliding window, subsampled words are replaced by the next word in sequence. This increases the data available for training, but will result in relative position being distorted at times. Consequently, we also generated spaces with static fixed-width sliding windows (EARPx) for position-aware models.", |
| "cite_spans": [ |
| { |
| "start": 512, |
| "end": 531, |
| "text": "(Levy et al., 2015)", |
| "ref_id": "BIBREF22" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Training procedure", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "After finding that fastText trained on Wikipedia performed better on analogy tests than prior results obtained with word2vec (Levy et al., 2015) , we adopted a number of its hyperparameter settings. We set the probability of negative sampling for each term to f 1 2 , where f is the number of term occurrences divided by the total token count. In addition we used an initial learning rate of .05, and subsampled terms with a probability of 1 \u2212 ( t f + t f ) 5 .", |
| "cite_spans": [ |
| { |
| "start": 125, |
| "end": 144, |
| "text": "(Levy et al., 2015)", |
| "ref_id": "BIBREF22" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Training procedure", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "5 While using this formula reliably improves performance on some of the analogy sets, it differs from both the formula described in Mikolov et al. (2013b) , and the formula implemented in the canonical word2vec implementation of SGNS -see Levy et al. (2015) for details. It is also difficult to justify on theoretical grounds, as it returns values less than zero for some words that meet the subsampling threshold. Nevertheless, retaining it throughout our experiments seemed more principled than altering the fastText baseline in a manner that impaired its performance.", |
| "cite_spans": [ |
| { |
| "start": 132, |
| "end": 154, |
| "text": "Mikolov et al. (2013b)", |
| "ref_id": "BIBREF27" |
| }, |
| { |
| "start": 239, |
| "end": 257, |
| "text": "Levy et al. (2015)", |
| "ref_id": "BIBREF22" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Training procedure", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "To evaluate the nature and utility of the additional information encoded by permutation-based variants, we utilized a set of analogical retrieval reference sets, including the MSR set (Mikolov et al., 2013c) consisting of 8,000 proportional analogy questions that are morphological in nature (e.g. young:younger:quick:?) and the Google analogy set (Mikolov et al., 2013a) which includes 8,869 semantic (and predominantly geographic, e.g. brussels:belgium:dublin:?) and 10,675 morphologically-oriented \"syntactic\" questions. We also included the Bigger Analogy Test Set (BATS) set , a more challenging set of 99,200 proportional analogy questions balanced across 40 linguistic types in four categories: Inflections (e.g. plurals, infinitives), Derivation (e.g. verb+er), Lexicography (e.g. hypernyms, synonyms) and Encylopedia (e.g. country:capital, male:female). We obtained these sets from the distribution described in Finley et al (2017) 6 , in which only the first correct answer to questions with multiple correct answers in BATS is retained, and used a parallelized implementation of the widely used vector offset method, in which for a given proportional analogy a:b:c:d, all word vectors in the space are rank-ordered in accordance with their cosine similarity to the vector \u2212 \u2212\u2212\u2212\u2212\u2212 \u2192 c + b \u2212 a. We report average accuracy, where a result is considered accurate if d is the top-ranked result aside from a, b and c. 7 To evaluate the effects of encoding word order on the relative distance between terms, we used a series of widely used reference sets that mediate comparison between human and machine estimates of pairwise similarity and relatedness between term pairs. Specifically, we used Wordsim-353 (Finkelstein et al., 2001) , split into subsets emphasizing similarity and relatedness (Agirre et al., 2009) ; MEN (Bruni et al., 2014) and Simlex-999 (Hill et al., 2015) . For each of these sets, we estimated the Spearman correlation of the cosine similarity between vector representations of the words in a given pair, with the human ratings (averaged across raters) of similarity and/or relat-edness provided in the reference standards.", |
| "cite_spans": [ |
| { |
| "start": 184, |
| "end": 207, |
| "text": "(Mikolov et al., 2013c)", |
| "ref_id": "BIBREF28" |
| }, |
| { |
| "start": 348, |
| "end": 371, |
| "text": "(Mikolov et al., 2013a)", |
| "ref_id": "BIBREF25" |
| }, |
| { |
| "start": 921, |
| "end": 940, |
| "text": "Finley et al (2017)", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 941, |
| "end": 942, |
| "text": "6", |
| "ref_id": null |
| }, |
| { |
| "start": 1423, |
| "end": 1424, |
| "text": "7", |
| "ref_id": null |
| }, |
| { |
| "start": 1712, |
| "end": 1738, |
| "text": "(Finkelstein et al., 2001)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 1799, |
| "end": 1820, |
| "text": "(Agirre et al., 2009)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 1827, |
| "end": 1847, |
| "text": "(Bruni et al., 2014)", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 1863, |
| "end": 1882, |
| "text": "(Hill et al., 2015)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "3.6" |
| }, |
| { |
| "text": "Only those examples in which all relevant terms were represented in our vector spaces were considered. Consequently, our analogy test sets consisted of 6136; 19,420 and 88,108 examples for the MSR 8 , Google and BATS sets respectively. With pairwise similarity, we retained 998; 335 and all 3,000 of the Simlex, Wordsim and MEN examples respectively. These numbers were identical across models, including fastText baselines.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "3.6" |
| }, |
| { |
| "text": "In addition we evaluated the effects of incorporating word order with EARP on three standard sequence labeling tasks: part-of-speech tagging of the Wall Street Journal sections of the Penn Treebank (PTB) and the CoNLL'00 sentence chunking (Tjong Kim Sang and Buchholz, 2000) and CoNLL'03 named entity recognition (Tjong Kim Sang and De Meulder, 2003) shared tasks. As was the case with the pairwise similarity and relatedness evaluations, we conducted these evaluations using the repEval2016 package 9 after converting all vectors to the word2vec binary format. This package provides implementations of the neural NLP architecture developed by Collobert and his colleagues (2011), which uses vectors for words within a five-word window as input, a single hidden layer of 300 units and an output Softmax layer. The implementation provided in repEval2016 deviates by design from the original implementation by fixing word vectors during training in order to emphasize difference between models for the purpose of comparative evaluation, which tends to reduce performance (for further details, see (Chiu et al., 2016) ). As spaces constructed with narrower sliding windows generally perform better on these tasks (Chiu et al., 2016) , we conducted these experiments with models of window radius 2 only. To facilitate fair comparison, we added random vectors representing tokens available in the fastText-derived spaces only to all spaces, replacing the original vectors where these existed. This was important in this evaluation as only fastText retains vector representation for punctuation marks (these are eliminated by the Semantic Vectors tokenization procedure), resulting in a relatively large number of out-of-vocabulary terms and predictably reduced performance with the Semantic Vectors implementation of the same algorithm. With the random vectors added, out-of-vocabulary rates were equivalent across the two SGNS implementations, resulting in similar performance.", |
| "cite_spans": [ |
| { |
| "start": 250, |
| "end": 274, |
| "text": "Sang and Buchholz, 2000)", |
| "ref_id": "BIBREF37" |
| }, |
| { |
| "start": 324, |
| "end": 350, |
| "text": "Sang and De Meulder, 2003)", |
| "ref_id": "BIBREF38" |
| }, |
| { |
| "start": 1095, |
| "end": 1114, |
| "text": "(Chiu et al., 2016)", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 1210, |
| "end": 1229, |
| "text": "(Chiu et al., 2016)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "3.6" |
| }, |
| { |
| "text": "The results of our analogical retrieval experiments are shown in Table 1 . With the single exception of the semantic component of the Google set, the best result on every set and subset was obtained by a variant of the EARP prox model, strongly suggesting that (1) information concerning relative position is of value for solving analogical retrieval problems;", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 65, |
| "end": 72, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Analogical retrieval", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "(2) encoding this information in a flexible manner that preserves the natural relationship of proximity between sliding window positions helps more than encoding only direction, or encoding window positions as disparate \"slots\".", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Analogical retrieval", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "On the syntactically-oriented subsets (Gsyn, Binf, Bder) adding subword information improves performance of baseline SGNS and EARP models, with subword-sensitive EARPx prox models showing improvements of between \u223c6% and \u223c21% in accuracy on these subtasks, as compared with the best performing baseline 10 . The results follow the same pattern at both sliding window radii, aside from a larger decrease in performance of EARPx models on the semantic component of the Google set at radius 2, attributable to semantically useful information lost on account of subsampling without replacement. In general, better performance on syntactic subsets is obtained at radius 2 with subword-sensitive models, with semantic subsets showing the opposite trend.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Analogical retrieval", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "While better performance on the total Google set has been reported with larger training corpora (Pennington et al., 2014; , the best EARP results on the syntactic component of this set surpass those reported from orderinsensitive models trained on more comprehensive corpora for multiple iterations. With this subset, 10 To assess reproducibility, we repeated the window radius 2 experiments a further 4 times, with different stochastic initialization of network weights. Performance was remarkably consistent, with a standard error of the mean accuracy on the MSR, Google and BATS sets at or below .24% (.0024), .33% (.0033) and .12% (.0012) for all models. All differences in performance from the baseline (only SGNS semVec was repeated) were statistically significant by unpaired t-test, aside from the results of EARP dir and EARP prox on the Google set when no subwords were used.", |
| "cite_spans": [ |
| { |
| "start": 96, |
| "end": 121, |
| "text": "(Pennington et al., 2014;", |
| "ref_id": "BIBREF30" |
| }, |
| { |
| "start": 318, |
| "end": 320, |
| "text": "10", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Analogical retrieval", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "the best EARP prox model obtained an accuracy of 76.74% after a single training iteration on a \u223c9 billion word Wikipedia-derived corpus. Pennington and his colleagues (2014) report a best accuracy of 69.3% after training Glove on a corpus of 42 billion words, and Mikolov and colleagues (2017) report an accuracy of 73% when training a subword-sensitive CBOW model for five iterations across a 630 billion word corpus derived from Common Crawl. The latter performance improved to 82% with pre-processing to tag phrases and position-dependent weighting -modifications that may improve EARP performance also, as would almost certainly be the case with multiple training iterations across a much larger corpus.", |
| "cite_spans": [ |
| { |
| "start": 264, |
| "end": 293, |
| "text": "Mikolov and colleagues (2017)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Analogical retrieval", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "Regarding the performance of other ordersensitive approaches on this subset, Trask and his colleagues (2015) report a 1.41% to 3.07% increase in absolute accuracy over a standard CBOW baseline with PENN, and Mikolov and his colleagues (2017) report a 4% increase over a subword-sensitive CBOW model with incorporation of position-dependent weights 11 . By comparison, EARPx prox yields improvements of up to 4.27% over the best baseline when subwords are not considered, and 8.29% with subword-sensitive models (both at radius 5).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Analogical retrieval", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "With the more challenging BATS analogies, Drozd and his colleagues (2016) report results for models trained on a 6 billion word corpus derived from the Wikipedia and several other resources, with best results with the vector offset method for BATS components of 61%, 11.2% (both SGNS), 10.9% and 31.5% (both Glove) for the Inflectional, Derivational, Lexicography and Encyclopedia components respectively. While not strictly comparable on account of our training on Wikipedia alone and acknowledging only one of a set of possible correct answers in some cases, our best results for these sets were 71.56%, 44.30%, 10.07% and 36.52% respectively, with a fourfold increase in performance on the derivational component. These results further support the hypothesis that order-related information is of value in several classes of proportional analogy problem.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Analogical retrieval", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "These improvements in analogy retrieval were not accompanied by better correlation with human es- ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Semantic similarity/relatedness", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Results of these experiments are shown in Table 2, and suggest an advantage for models encoding position in sequence labeling tasks. In particular, for the sentence chunking shared task (CoNLL00), the best results obtained with an order-aware model (EARPx pos + subwords) exceeds the best baseline result by around 2%, with smaller improvements in performance on the other two tasks, including a .43% improvement in accuracy for part-of-speech tagging (PTB, without subwords) that is comparable to the .37% improvement over a SGNS baseline reported on this dataset by Ling and his colleagues (2015) when using separate (rather than shared) positiondependent context weights.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sequence labeling", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "All models were trained on a single multi-core machine. In general, Semantic Vectors takes slightly longer to run than fastText, which takes around an hour to generate models including building of the dictionary. With Semantic Vectors, models were generated in around 1h20 when the corpus was indexed at document level, which is desirable as this package uses Lucene 12 for tokenization and indexing. As the accuracy of fastText on analogy completion tasks dropped considerably when we attempted to train it on unsegmented documents, we adapted Semantic Vectors to treat each line of input as an individual document. As this approximately doubled the time required to generate each model, we would not recommend this other than for the purpose of comparative evaluation. Adding subword embeddings increased training time by three-to fourfold. Source code is available via GitHub 13 , and embeddings are publicly available at 14 .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Computational performance", |
| "sec_num": "4.4" |
| }, |
| { |
| "text": "While this paper focused on encoding positional information, EARP is generalizable and could be used to encode other sorts of relational information also. An interesting direction for future work may involve using EARP to encode the nature of semantic and dependency relations, as has been done with RI previously (Cohen et al., 2009; Basile et al., 2011) . As the main focus of the current paper was on comparative evaluation across models with identical hyper-parameters, we have yet to formally evaluate the extent to which hyperparameter settings (such as dimensionality) may affect performance, and it seems likely that hyperparameters that would further accentuate EARP performance remain to be identified.", |
| "cite_spans": [ |
| { |
| "start": 314, |
| "end": 334, |
| "text": "(Cohen et al., 2009;", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 335, |
| "end": 355, |
| "text": "Basile et al., 2011)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Limitations and future work", |
| "sec_num": "4.5" |
| }, |
| { |
| "text": "This paper describes EARP, a novel method through which word order can be encoded into neural word embedding representations. Of note, this additional information is encoded without the need for additional computation, and space requirements are practically identical to those of baseline models. Upon evaluation, encoding word order results in substantive improvements in performance across multiple analogical retrieval reference sets, with best performance when order information is encoded using a novel permutationbased method of positional binding.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "4.6" |
| }, |
| { |
| "text": "Words and n-grams are weighted equally, following and the fastText implementation 3 https://dumps.wikimedia.org/enwiki 4 https://github.com/facebookresearch /fastText", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "https://github.com/gpfinley/analogies 7 For a finer-grained estimate of performance -which we considered particularly important in cases in BATS in which only the first of a set of correct answers was retained -we also calculated the mean reciprocal rank (rank \u22121 ) (Voorhees et al., 1999) of d across all questions. As these results closely mirrored the accuracy results, we report accuracy only.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "The fraction of the MSR set retained is smaller, because 1000 of the examples in this set concern identifying possessives indicates by the presence of an apostrophe, and terms of this nature were eliminated by the pre-processing procedure.9 https://github.com/cambridgeltl/RepEval-2016", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "CBOW baselines were around 10% higher than our SGNS baselines, attributable to differences in corpus size, composition and preprocessing; and perhaps architecture.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "https://lucene.apache.org/ 13 https://github.com/semanticvectors/semanticvectors", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "https://doi.org/10.5281/zenodo.1345333", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "This work was supported by US National Library of Medicine grant (R01 LM011563), and conducted in part at the University of Texas School of Biomedical Informatics at Houston.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgments", |
| "sec_num": null |
| }, |
| { |
| "text": "timates of semantic similarity and relatedness. To the contrary with a window radius of 2, the best results on all sets and subsets were produced by SGNS baselines, with a best baseline Spearman Rho of .41 (SGNS fastText ), .72 (SGNS semVec ) and .77 (SGNS semVec + subwords) for SimLex-999, WS-353 and MEN respectively. Surprisingly, incorporating order-related information reduced performance on all of these sets, with best performance for SimLex-999, WS-353 and MEN of .40 (EARPx prox ), .71 (EARP pos ) and .76 (EARP dir + subwords) respectively; and worst performance of .37 (EARP pos ), .66 and .71 (both EARPx prox ) respectively. Other models fell between these ex-tremes, with a similar pattern observed with window radius of 5. The differences in performance observed with these tasks are neither as stark nor as consistent as those with analogy tasks, with EARP performance falling between that of the two baseline models in several cases. Nonetheless, EARP models tend to correlate worse with human estimates of pairwise similarity and relatedness. This drop in performance may relate to how semantic information is dispersed with positionaware models -a neighboring word is encoded differently depending on position, which may obscure the semantically useful information that two other words both occur in proximity to it.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "annex", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "A study on similarity and relatedness using distributional and wordnet-based approaches", |
| "authors": [ |
| { |
| "first": "Eneko", |
| "middle": [], |
| "last": "Agirre", |
| "suffix": "" |
| }, |
| { |
| "first": "Enrique", |
| "middle": [], |
| "last": "Alfonseca", |
| "suffix": "" |
| }, |
| { |
| "first": "Keith", |
| "middle": [], |
| "last": "Hall", |
| "suffix": "" |
| }, |
| { |
| "first": "Jana", |
| "middle": [], |
| "last": "Kravalova", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Proceedings of Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Eneko Agirre, Enrique Alfonseca, Keith Hall, Jana Kravalova, Marius Pa\u015fca, and Aitor Soroa. 2009. A study on similarity and relatedness using distribu- tional and wordnet-based approaches. In Proceed- ings of Human Language Technologies: The 2009", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Annual Conference of the North American Chapter of the Association for Computational Linguistics", |
| "authors": [], |
| "year": null, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "19--27", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Annual Conference of the North American Chap- ter of the Association for Computational Linguistics, pages 19-27. Association for Computational Lin- guistics.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Encoding syntactic dependencies by vector permutation", |
| "authors": [ |
| { |
| "first": "Pierpaolo", |
| "middle": [], |
| "last": "Basile", |
| "suffix": "" |
| }, |
| { |
| "first": "Annalina", |
| "middle": [], |
| "last": "Caputo", |
| "suffix": "" |
| }, |
| { |
| "first": "Giovanni", |
| "middle": [], |
| "last": "Semeraro", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the GEMS 2011 Workshop on GEometrical Models of Natural Language Semantics", |
| "volume": "", |
| "issue": "", |
| "pages": "43--51", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Pierpaolo Basile, Annalina Caputo, and Giovanni Se- meraro. 2011. Encoding syntactic dependencies by vector permutation. In Proceedings of the GEMS 2011 Workshop on GEometrical Models of Natural Language Semantics, pages 43-51. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "A neural probabilistic language model", |
| "authors": [ |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| }, |
| { |
| "first": "R\u00e9jean", |
| "middle": [], |
| "last": "Ducharme", |
| "suffix": "" |
| }, |
| { |
| "first": "Pascal", |
| "middle": [], |
| "last": "Vincent", |
| "suffix": "" |
| }, |
| { |
| "first": "Christian", |
| "middle": [], |
| "last": "Jauvin", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Journal of machine learning research", |
| "volume": "3", |
| "issue": "", |
| "pages": "1137--1155", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yoshua Bengio, R\u00e9jean Ducharme, Pascal Vincent, and Christian Jauvin. 2003. A neural probabilistic lan- guage model. Journal of machine learning research, 3(Feb):1137-1155.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Enriching word vectors with subword information", |
| "authors": [ |
| { |
| "first": "Piotr", |
| "middle": [], |
| "last": "Bojanowski", |
| "suffix": "" |
| }, |
| { |
| "first": "Edouard", |
| "middle": [], |
| "last": "Grave", |
| "suffix": "" |
| }, |
| { |
| "first": "Armand", |
| "middle": [], |
| "last": "Joulin", |
| "suffix": "" |
| }, |
| { |
| "first": "Tomas", |
| "middle": [], |
| "last": "Mikolov", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Transactions of the Association of Computational Linguistics", |
| "volume": "5", |
| "issue": "1", |
| "pages": "135--146", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Piotr Bojanowski, Edouard Grave, Armand Joulin, and Tomas Mikolov. 2017. Enriching word vectors with subword information. Transactions of the Associa- tion of Computational Linguistics, 5(1):135-146.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Multimodal distributional semantics", |
| "authors": [ |
| { |
| "first": "Elia", |
| "middle": [], |
| "last": "Bruni", |
| "suffix": "" |
| }, |
| { |
| "first": "Nam-Khanh", |
| "middle": [], |
| "last": "Tran", |
| "suffix": "" |
| }, |
| { |
| "first": "Marco", |
| "middle": [], |
| "last": "Baroni", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Journal of Artificial Intelligence Research", |
| "volume": "49", |
| "issue": "", |
| "pages": "1--47", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Elia Bruni, Nam-Khanh Tran, and Marco Baroni. 2014. Multimodal distributional semantics. Journal of Ar- tificial Intelligence Research, 49:1-47.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Intrinsic evaluation of word vectors fails to predict extrinsic performance", |
| "authors": [ |
| { |
| "first": "Billy", |
| "middle": [], |
| "last": "Chiu", |
| "suffix": "" |
| }, |
| { |
| "first": "Anna", |
| "middle": [], |
| "last": "Korhonen", |
| "suffix": "" |
| }, |
| { |
| "first": "Sampo", |
| "middle": [], |
| "last": "Pyysalo", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 1st Workshop on Evaluating Vector-Space Representations for NLP", |
| "volume": "", |
| "issue": "", |
| "pages": "1--6", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Billy Chiu, Anna Korhonen, and Sampo Pyysalo. 2016. Intrinsic evaluation of word vectors fails to predict extrinsic performance. In Proceedings of the 1st Workshop on Evaluating Vector-Space Representa- tions for NLP, pages 1-6.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Predication-based semantic indexing: Permutations as a means to encode predications in semantic space", |
| "authors": [ |
| { |
| "first": "Trevor", |
| "middle": [], |
| "last": "Cohen", |
| "suffix": "" |
| }, |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Roger", |
| "suffix": "" |
| }, |
| { |
| "first": "Thomas C", |
| "middle": [], |
| "last": "Schvaneveldt", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Rindflesch", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "AMIA Annual Symposium Proceedings", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Trevor Cohen, Roger W Schvaneveldt, and Thomas C Rindflesch. 2009. Predication-based semantic in- dexing: Permutations as a means to encode predi- cations in semantic space. In AMIA Annual Sympo- sium Proceedings, volume 2009, page 114. Ameri- can Medical Informatics Association.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Orthogonality and orthography: introducing measured distance into semantic space", |
| "authors": [ |
| { |
| "first": "Trevor", |
| "middle": [], |
| "last": "Cohen", |
| "suffix": "" |
| }, |
| { |
| "first": "Dominic", |
| "middle": [], |
| "last": "Widdows", |
| "suffix": "" |
| }, |
| { |
| "first": "Manuel", |
| "middle": [], |
| "last": "Wahle", |
| "suffix": "" |
| }, |
| { |
| "first": "Roger", |
| "middle": [], |
| "last": "Schvaneveldt", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "International Symposium on Quantum Interaction", |
| "volume": "", |
| "issue": "", |
| "pages": "34--46", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Trevor Cohen, Dominic Widdows, Manuel Wahle, and Roger Schvaneveldt. 2013. Orthogonality and or- thography: introducing measured distance into se- mantic space. In International Symposium on Quan- tum Interaction, pages 34-46. Springer.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Natural language processing (almost) from scratch", |
| "authors": [ |
| { |
| "first": "Ronan", |
| "middle": [], |
| "last": "Collobert", |
| "suffix": "" |
| }, |
| { |
| "first": "Jason", |
| "middle": [], |
| "last": "Weston", |
| "suffix": "" |
| }, |
| { |
| "first": "L\u00e9on", |
| "middle": [], |
| "last": "Bottou", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Karlen", |
| "suffix": "" |
| }, |
| { |
| "first": "Koray", |
| "middle": [], |
| "last": "Kavukcuoglu", |
| "suffix": "" |
| }, |
| { |
| "first": "Pavel", |
| "middle": [], |
| "last": "Kuksa", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Journal of Machine Learning Research", |
| "volume": "12", |
| "issue": "", |
| "pages": "2493--2537", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ronan Collobert, Jason Weston, L\u00e9on Bottou, Michael Karlen, Koray Kavukcuoglu, and Pavel Kuksa. 2011. Natural language processing (almost) from scratch. Journal of Machine Learning Research, 12(Aug):2493-2537.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Word embeddings, analogies, and machine learning: Beyond king-man+ woman= queen", |
| "authors": [ |
| { |
| "first": "Aleksandr", |
| "middle": [], |
| "last": "Drozd", |
| "suffix": "" |
| }, |
| { |
| "first": "Anna", |
| "middle": [], |
| "last": "Gladkova", |
| "suffix": "" |
| }, |
| { |
| "first": "Satoshi", |
| "middle": [], |
| "last": "Matsuoka", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers", |
| "volume": "", |
| "issue": "", |
| "pages": "3519--3530", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Aleksandr Drozd, Anna Gladkova, and Satoshi Mat- suoka. 2016. Word embeddings, analogies, and ma- chine learning: Beyond king-man+ woman= queen. In Proceedings of COLING 2016, the 26th Inter- national Conference on Computational Linguistics: Technical Papers, pages 3519-3530.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Placing search in context: The concept revisited", |
| "authors": [ |
| { |
| "first": "Lev", |
| "middle": [], |
| "last": "Finkelstein", |
| "suffix": "" |
| }, |
| { |
| "first": "Evgeniy", |
| "middle": [], |
| "last": "Gabrilovich", |
| "suffix": "" |
| }, |
| { |
| "first": "Yossi", |
| "middle": [], |
| "last": "Matias", |
| "suffix": "" |
| }, |
| { |
| "first": "Ehud", |
| "middle": [], |
| "last": "Rivlin", |
| "suffix": "" |
| }, |
| { |
| "first": "Zach", |
| "middle": [], |
| "last": "Solan", |
| "suffix": "" |
| }, |
| { |
| "first": "Gadi", |
| "middle": [], |
| "last": "Wolfman", |
| "suffix": "" |
| }, |
| { |
| "first": "Eytan", |
| "middle": [], |
| "last": "Ruppin", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the 10th international conference on World Wide Web", |
| "volume": "", |
| "issue": "", |
| "pages": "406--414", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lev Finkelstein, Evgeniy Gabrilovich, Yossi Matias, Ehud Rivlin, Zach Solan, Gadi Wolfman, and Ey- tan Ruppin. 2001. Placing search in context: The concept revisited. In Proceedings of the 10th inter- national conference on World Wide Web, pages 406- 414. ACM.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "What analogies reveal about word vectors and their compositionality", |
| "authors": [ |
| { |
| "first": "Gregory", |
| "middle": [], |
| "last": "Finley", |
| "suffix": "" |
| }, |
| { |
| "first": "Stephanie", |
| "middle": [], |
| "last": "Farmer", |
| "suffix": "" |
| }, |
| { |
| "first": "Serguei", |
| "middle": [], |
| "last": "Pakhomov", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 6th Joint Conference on Lexical and Computational Semantics (* SEM 2017)", |
| "volume": "", |
| "issue": "", |
| "pages": "1--11", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Gregory Finley, Stephanie Farmer, and Serguei Pakho- mov. 2017. What analogies reveal about word vec- tors and their compositionality. In Proceedings of the 6th Joint Conference on Lexical and Computa- tional Semantics (* SEM 2017), pages 1-11.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Positional binding with distributed representations", |
| "authors": [ |
| { |
| "first": "I", |
| "middle": [], |
| "last": "Stephen", |
| "suffix": "" |
| }, |
| { |
| "first": "Phil", |
| "middle": [], |
| "last": "Gallant", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Culliton", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Image, Vision and Computing (ICIVC), International Conference on", |
| "volume": "", |
| "issue": "", |
| "pages": "108--113", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Stephen I Gallant and Phil Culliton. 2016. Positional binding with distributed representations. In Image, Vision and Computing (ICIVC), International Con- ference on, pages 108-113. IEEE.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Analogy-based detection of morphological and semantic relations with word embeddings: what works and what doesn't", |
| "authors": [ |
| { |
| "first": "Anna", |
| "middle": [], |
| "last": "Gladkova", |
| "suffix": "" |
| }, |
| { |
| "first": "Aleksandr", |
| "middle": [], |
| "last": "Drozd", |
| "suffix": "" |
| }, |
| { |
| "first": "Satoshi", |
| "middle": [], |
| "last": "Matsuoka", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the NAACL Student Research Workshop", |
| "volume": "", |
| "issue": "", |
| "pages": "8--15", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Anna Gladkova, Aleksandr Drozd, and Satoshi Mat- suoka. 2016. Analogy-based detection of morpho- logical and semantic relations with word embed- dings: what works and what doesn't. In Proceedings of the NAACL Student Research Workshop, pages 8- 15.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "word2vec explained: Deriving mikolov et al.'s negativesampling word-embedding method", |
| "authors": [ |
| { |
| "first": "Yoav", |
| "middle": [], |
| "last": "Goldberg", |
| "suffix": "" |
| }, |
| { |
| "first": "Omer", |
| "middle": [], |
| "last": "Levy", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1402.3722" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yoav Goldberg and Omer Levy. 2014. word2vec explained: Deriving mikolov et al.'s negative- sampling word-embedding method. arXiv preprint arXiv:1402.3722.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Simlex-999: Evaluating semantic models with (genuine) similarity estimation", |
| "authors": [ |
| { |
| "first": "Felix", |
| "middle": [], |
| "last": "Hill", |
| "suffix": "" |
| }, |
| { |
| "first": "Roi", |
| "middle": [], |
| "last": "Reichart", |
| "suffix": "" |
| }, |
| { |
| "first": "Anna", |
| "middle": [], |
| "last": "Korhonen", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Computational Linguistics", |
| "volume": "41", |
| "issue": "4", |
| "pages": "665--695", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Felix Hill, Roi Reichart, and Anna Korhonen. 2015. Simlex-999: Evaluating semantic models with (gen- uine) similarity estimation. Computational Linguis- tics, 41(4):665-695.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "High-dimensional semantic space accounts of priming", |
| "authors": [ |
| { |
| "first": "N", |
| "middle": [], |
| "last": "Michael", |
| "suffix": "" |
| }, |
| { |
| "first": "Walter", |
| "middle": [], |
| "last": "Jones", |
| "suffix": "" |
| }, |
| { |
| "first": "Douglas Jk", |
| "middle": [], |
| "last": "Kintsch", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Mewhort", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Journal of memory and language", |
| "volume": "55", |
| "issue": "4", |
| "pages": "534--552", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Michael N Jones, Walter Kintsch, and Douglas JK Me- whort. 2006. High-dimensional semantic space ac- counts of priming. Journal of memory and lan- guage, 55(4):534-552.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Random indexing of text samples for latent semantic analysis", |
| "authors": [ |
| { |
| "first": "Pentii", |
| "middle": [], |
| "last": "Kanerva", |
| "suffix": "" |
| }, |
| { |
| "first": "Jan", |
| "middle": [], |
| "last": "Kristoferson", |
| "suffix": "" |
| }, |
| { |
| "first": "Anders", |
| "middle": [], |
| "last": "Holst", |
| "suffix": "" |
| } |
| ], |
| "year": 2000, |
| "venue": "Proceedings of the Annual Meeting of the Cognitive Science Society", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Pentii Kanerva, Jan Kristoferson, and Anders Holst. 2000. Random indexing of text samples for la- tent semantic analysis. In Proceedings of the An- nual Meeting of the Cognitive Science Society, vol- ume 22.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Hyperdimensional computing: An introduction to computing in distributed representation with high-dimensional random vectors", |
| "authors": [ |
| { |
| "first": "Pentti", |
| "middle": [], |
| "last": "Kanerva", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Cognitive Computation", |
| "volume": "1", |
| "issue": "2", |
| "pages": "139--159", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Pentti Kanerva. 2009. Hyperdimensional computing: An introduction to computing in distributed rep- resentation with high-dimensional random vectors. Cognitive Computation, 1(2):139-159.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "From words to understanding", |
| "authors": [ |
| { |
| "first": "Jussi", |
| "middle": [], |
| "last": "Karlgren", |
| "suffix": "" |
| }, |
| { |
| "first": "Magnus", |
| "middle": [], |
| "last": "Sahlgren", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Foundations of Real-World Intelligence", |
| "volume": "", |
| "issue": "", |
| "pages": "294--308", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jussi Karlgren and Magnus Sahlgren. 2001. From words to understanding. Foundations of Real-World Intelligence, pages 294-308.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "A solution to Plato's problem: The latent semantic analysis theory of acquisition", |
| "authors": [ |
| { |
| "first": "Thomas", |
| "middle": [], |
| "last": "Landauer", |
| "suffix": "" |
| }, |
| { |
| "first": "Susan", |
| "middle": [], |
| "last": "Dumais", |
| "suffix": "" |
| } |
| ], |
| "year": 1997, |
| "venue": "Psychological Review", |
| "volume": "104", |
| "issue": "2", |
| "pages": "211--240", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Thomas Landauer and Susan Dumais. 1997. A solu- tion to Plato's problem: The latent semantic anal- ysis theory of acquisition. Psychological Review, 104(2):211-240.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "Improving distributional similarity with lessons learned from word embeddings", |
| "authors": [ |
| { |
| "first": "Omer", |
| "middle": [], |
| "last": "Levy", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoav", |
| "middle": [], |
| "last": "Goldberg", |
| "suffix": "" |
| }, |
| { |
| "first": "Ido", |
| "middle": [], |
| "last": "Dagan", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Transactions of the Association for Computational Linguistics", |
| "volume": "3", |
| "issue": "", |
| "pages": "211--225", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Omer Levy, Yoav Goldberg, and Ido Dagan. 2015. Im- proving distributional similarity with lessons learned from word embeddings. Transactions of the Associ- ation for Computational Linguistics, 3:211-225.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "Two/too simple adaptations of word2vec for syntax problems", |
| "authors": [ |
| { |
| "first": "Wang", |
| "middle": [], |
| "last": "Ling", |
| "suffix": "" |
| }, |
| { |
| "first": "Chris", |
| "middle": [], |
| "last": "Dyer", |
| "suffix": "" |
| }, |
| { |
| "first": "Alan", |
| "middle": [ |
| "W" |
| ], |
| "last": "Black", |
| "suffix": "" |
| }, |
| { |
| "first": "Isabel", |
| "middle": [], |
| "last": "Trancoso", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "1299--1304", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wang Ling, Chris Dyer, Alan W Black, and Isabel Trancoso. 2015. Two/too simple adaptations of word2vec for syntax problems. In Proceedings of the 2015 Conference of the North American Chap- ter of the Association for Computational Linguistics: Human Language Technologies, pages 1299-1304.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "Producing high-dimensional semantic spaces from lexical cooccurrence", |
| "authors": [ |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Lund", |
| "suffix": "" |
| }, |
| { |
| "first": "Curt", |
| "middle": [], |
| "last": "Burgess", |
| "suffix": "" |
| } |
| ], |
| "year": 1996, |
| "venue": "Behavior research methods, instruments, & computers", |
| "volume": "28", |
| "issue": "", |
| "pages": "203--208", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kevin Lund and Curt Burgess. 1996. Producing high-dimensional semantic spaces from lexical co- occurrence. Behavior research methods, instru- ments, & computers, 28(2):203-208.", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "Efficient estimation of word representations in vector space", |
| "authors": [ |
| { |
| "first": "Tomas", |
| "middle": [], |
| "last": "Mikolov", |
| "suffix": "" |
| }, |
| { |
| "first": "Kai", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Greg", |
| "middle": [], |
| "last": "Corrado", |
| "suffix": "" |
| }, |
| { |
| "first": "Jeffrey", |
| "middle": [], |
| "last": "Dean", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1301.3781" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tomas Mikolov, Kai Chen, Greg Corrado, and Jef- frey Dean. 2013a. Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781.", |
| "links": null |
| }, |
| "BIBREF26": { |
| "ref_id": "b26", |
| "title": "Advances in pre-training distributed word representations", |
| "authors": [ |
| { |
| "first": "Tomas", |
| "middle": [], |
| "last": "Mikolov", |
| "suffix": "" |
| }, |
| { |
| "first": "Edouard", |
| "middle": [], |
| "last": "Grave", |
| "suffix": "" |
| }, |
| { |
| "first": "Piotr", |
| "middle": [], |
| "last": "Bojanowski", |
| "suffix": "" |
| }, |
| { |
| "first": "Christian", |
| "middle": [], |
| "last": "Puhrsch", |
| "suffix": "" |
| }, |
| { |
| "first": "Armand", |
| "middle": [], |
| "last": "Joulin", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1712.09405" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tomas Mikolov, Edouard Grave, Piotr Bojanowski, Christian Puhrsch, and Armand Joulin. 2017. Ad- vances in pre-training distributed word representa- tions. arXiv preprint arXiv:1712.09405.", |
| "links": null |
| }, |
| "BIBREF27": { |
| "ref_id": "b27", |
| "title": "Distributed representations of words and phrases and their compositionality", |
| "authors": [ |
| { |
| "first": "Tomas", |
| "middle": [], |
| "last": "Mikolov", |
| "suffix": "" |
| }, |
| { |
| "first": "Ilya", |
| "middle": [], |
| "last": "Sutskever", |
| "suffix": "" |
| }, |
| { |
| "first": "Kai", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Greg", |
| "middle": [ |
| "S" |
| ], |
| "last": "Corrado", |
| "suffix": "" |
| }, |
| { |
| "first": "Jeff", |
| "middle": [], |
| "last": "Dean", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Advances in neural information processing systems", |
| "volume": "", |
| "issue": "", |
| "pages": "3111--3119", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Cor- rado, and Jeff Dean. 2013b. Distributed representa- tions of words and phrases and their compositional- ity. In Advances in neural information processing systems, pages 3111-3119.", |
| "links": null |
| }, |
| "BIBREF28": { |
| "ref_id": "b28", |
| "title": "Linguistic regularities in continuous space word representations", |
| "authors": [ |
| { |
| "first": "Tomas", |
| "middle": [], |
| "last": "Mikolov", |
| "suffix": "" |
| }, |
| { |
| "first": "Yih", |
| "middle": [], |
| "last": "Wen-Tau", |
| "suffix": "" |
| }, |
| { |
| "first": "Geoffrey", |
| "middle": [], |
| "last": "Zweig", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "746--751", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tomas Mikolov, Wen-tau Yih, and Geoffrey Zweig. 2013c. Linguistic regularities in continuous space word representations. In Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 746-751.", |
| "links": null |
| }, |
| "BIBREF29": { |
| "ref_id": "b29", |
| "title": "Learning word embeddings efficiently with noise-contrastive estimation", |
| "authors": [ |
| { |
| "first": "Andriy", |
| "middle": [], |
| "last": "Mnih", |
| "suffix": "" |
| }, |
| { |
| "first": "Koray", |
| "middle": [], |
| "last": "Kavukcuoglu", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Advances in neural information processing systems", |
| "volume": "", |
| "issue": "", |
| "pages": "2265--2273", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Andriy Mnih and Koray Kavukcuoglu. 2013. Learning word embeddings efficiently with noise-contrastive estimation. In Advances in neural information pro- cessing systems, pages 2265-2273.", |
| "links": null |
| }, |
| "BIBREF30": { |
| "ref_id": "b30", |
| "title": "Glove: Global vectors for word representation", |
| "authors": [ |
| { |
| "first": "Jeffrey", |
| "middle": [], |
| "last": "Pennington", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Socher", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [], |
| "last": "Manning", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP)", |
| "volume": "", |
| "issue": "", |
| "pages": "1532--1543", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jeffrey Pennington, Richard Socher, and Christopher Manning. 2014. Glove: Global vectors for word representation. In Proceedings of the 2014 confer- ence on empirical methods in natural language pro- cessing (EMNLP), pages 1532-1543.", |
| "links": null |
| }, |
| "BIBREF31": { |
| "ref_id": "b31", |
| "title": "Holographic reduced representations", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Tony", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Plate", |
| "suffix": "" |
| } |
| ], |
| "year": 1995, |
| "venue": "IEEE Transactions on Neural networks", |
| "volume": "6", |
| "issue": "3", |
| "pages": "623--641", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tony A Plate. 1995. Holographic reduced represen- tations. IEEE Transactions on Neural networks, 6(3):623-641.", |
| "links": null |
| }, |
| "BIBREF32": { |
| "ref_id": "b32", |
| "title": "Encoding sequential information in vector space models of semantics: Comparing holographic reduced representation and random permutation", |
| "authors": [ |
| { |
| "first": "Gabriel", |
| "middle": [], |
| "last": "Recchia", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Jones", |
| "suffix": "" |
| }, |
| { |
| "first": "Magnus", |
| "middle": [], |
| "last": "Sahlgren", |
| "suffix": "" |
| }, |
| { |
| "first": "Pentti", |
| "middle": [], |
| "last": "Kanerva", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the Annual Meeting of the Cognitive Science Society", |
| "volume": "32", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Gabriel Recchia, Michael Jones, Magnus Sahlgren, and Pentti Kanerva. 2010. Encoding sequential informa- tion in vector space models of semantics: Compar- ing holographic reduced representation and random permutation. In Proceedings of the Annual Meeting of the Cognitive Science Society, volume 32.", |
| "links": null |
| }, |
| "BIBREF33": { |
| "ref_id": "b33", |
| "title": "word2vec parameter learning explained", |
| "authors": [ |
| { |
| "first": "Xin", |
| "middle": [], |
| "last": "Rong", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1411.2738" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Xin Rong. 2014. word2vec parameter learning ex- plained. arXiv preprint arXiv:1411.2738.", |
| "links": null |
| }, |
| "BIBREF34": { |
| "ref_id": "b34", |
| "title": "Permutations as a means to encode order in word space", |
| "authors": [ |
| { |
| "first": "Magnus", |
| "middle": [], |
| "last": "Sahlgren", |
| "suffix": "" |
| }, |
| { |
| "first": "Anders", |
| "middle": [], |
| "last": "Holst", |
| "suffix": "" |
| }, |
| { |
| "first": "Pentti", |
| "middle": [], |
| "last": "Kanerva", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Magnus Sahlgren, Anders Holst, and Pentti Kanerva. 2008. Permutations as a means to encode order in word space.", |
| "links": null |
| }, |
| "BIBREF35": { |
| "ref_id": "b35", |
| "title": "Word space", |
| "authors": [ |
| { |
| "first": "Hinrich", |
| "middle": [], |
| "last": "Sch\u00fctze", |
| "suffix": "" |
| } |
| ], |
| "year": 1993, |
| "venue": "Advances in neural information processing systems", |
| "volume": "", |
| "issue": "", |
| "pages": "895--902", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hinrich Sch\u00fctze. 1993. Word space. In Advances in neural information processing systems, pages 895- 902.", |
| "links": null |
| }, |
| "BIBREF36": { |
| "ref_id": "b36", |
| "title": "Automatic word sense discrimination", |
| "authors": [ |
| { |
| "first": "Hinrich", |
| "middle": [], |
| "last": "Sch\u00fctze", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "Computational Linguistics", |
| "volume": "24", |
| "issue": "1", |
| "pages": "97--124", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hinrich Sch\u00fctze. 1998. Automatic word sense discrim- ination. Computational Linguistics, 24(1):97-124.", |
| "links": null |
| }, |
| "BIBREF37": { |
| "ref_id": "b37", |
| "title": "Introduction to the conll-2000 shared task: Chunking", |
| "authors": [ |
| { |
| "first": "Erik F Tjong Kim", |
| "middle": [], |
| "last": "Sang", |
| "suffix": "" |
| }, |
| { |
| "first": "Sabine", |
| "middle": [], |
| "last": "Buchholz", |
| "suffix": "" |
| } |
| ], |
| "year": 2000, |
| "venue": "Proceedings of the 2nd workshop on Learning language in logic and the 4th conference on Computational natural language learning", |
| "volume": "7", |
| "issue": "", |
| "pages": "127--132", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Erik F Tjong Kim Sang and Sabine Buchholz. 2000. Introduction to the conll-2000 shared task: Chunk- ing. In Proceedings of the 2nd workshop on Learn- ing language in logic and the 4th conference on Computational natural language learning-Volume 7, pages 127-132. Association for Computational Lin- guistics.", |
| "links": null |
| }, |
| "BIBREF38": { |
| "ref_id": "b38", |
| "title": "Introduction to the conll-2003 shared task: Language-independent named entity recognition", |
| "authors": [ |
| { |
| "first": "Erik F Tjong Kim", |
| "middle": [], |
| "last": "Sang", |
| "suffix": "" |
| }, |
| { |
| "first": "Fien", |
| "middle": [], |
| "last": "De Meulder", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Proceedings of the seventh conference on Natural language learning at HLT-NAACL 2003", |
| "volume": "4", |
| "issue": "", |
| "pages": "142--147", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Erik F Tjong Kim Sang and Fien De Meulder. 2003. Introduction to the conll-2003 shared task: Language-independent named entity recognition. In Proceedings of the seventh conference on Natural language learning at HLT-NAACL 2003-Volume 4, pages 142-147. Association for Computational Lin- guistics.", |
| "links": null |
| }, |
| "BIBREF39": { |
| "ref_id": "b39", |
| "title": "Modeling order in neural word embeddings at scale", |
| "authors": [ |
| { |
| "first": "Andrew", |
| "middle": [], |
| "last": "Trask", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Gilmore", |
| "suffix": "" |
| }, |
| { |
| "first": "Matthew", |
| "middle": [], |
| "last": "Russell", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 32nd International Conference on International Conference on Machine Learning", |
| "volume": "37", |
| "issue": "", |
| "pages": "2266--2275", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Andrew Trask, David Gilmore, and Matthew Rus- sell. 2015. Modeling order in neural word em- beddings at scale. In Proceedings of the 32nd In- ternational Conference on International Conference on Machine Learning-Volume 37, pages 2266-2275. JMLR. org.", |
| "links": null |
| }, |
| "BIBREF40": { |
| "ref_id": "b40", |
| "title": "From frequency to meaning: Vector space models of semantics", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Peter", |
| "suffix": "" |
| }, |
| { |
| "first": "Patrick", |
| "middle": [], |
| "last": "Turney", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Pantel", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Journal of artificial intelligence research", |
| "volume": "37", |
| "issue": "", |
| "pages": "141--188", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Peter D Turney and Patrick Pantel. 2010. From fre- quency to meaning: Vector space models of se- mantics. Journal of artificial intelligence research, 37:141-188.", |
| "links": null |
| }, |
| "BIBREF41": { |
| "ref_id": "b41", |
| "title": "The trec-8 question answering track report", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Ellen", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Voorhees", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "Trec", |
| "volume": "99", |
| "issue": "", |
| "pages": "77--82", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ellen M Voorhees et al. 1999. The trec-8 question an- swering track report. In Trec, volume 99, pages 77- 82.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "num": null, |
| "type_str": "figure", |
| "text": "The Architecture of EARP of the set of positions {\u22122; \u22121; +1; +2}. With p indicating the application of a context-specific random permutation for position p, the optimization objective becomes:", |
| "uris": null |
| }, |
| "TABREF1": { |
| "type_str": "table", |
| "html": null, |
| "text": "Performance on Sequence Labeling Tasks. % accuracy shown for PTB, and % F-measure otherwise", |
| "content": "<table/>", |
| "num": null |
| } |
| } |
| } |
| } |