| { |
| "paper_id": "P15-1049", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T09:09:36.341102Z" |
| }, |
| "title": "S-MART: Novel Tree-based Structured Learning Algorithms Applied to Tweet Entity Linking", |
| "authors": [ |
| { |
| "first": "Yi", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Georgia Institute of Technology", |
| "location": {} |
| }, |
| "email": "yiyang@gatech.edu" |
| }, |
| { |
| "first": "Ming-Wei", |
| "middle": [], |
| "last": "Chang", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Microsoft Research", |
| "location": {} |
| }, |
| "email": "minchang@microsoft.com" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "Non-linear models recently receive a lot of attention as people are starting to discover the power of statistical and embedding features. However, tree-based models are seldom studied in the context of structured learning despite their recent success on various classification and ranking tasks. In this paper, we propose SMART , a tree-based structured learning framework based on multiple additive regression trees. SMART is especially suitable for handling tasks with dense features, and can be used to learn many different structures under various loss functions. We apply SMART to the task of tweet entity linking-a core component of tweet information extraction, which aims to identify and link name mentions to entities in a knowledge base. A novel inference algorithm is proposed to handle the special structure of the task. The experimental results show that SMART significantly outperforms state-of-the-art tweet entity linking systems.", |
| "pdf_parse": { |
| "paper_id": "P15-1049", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "Non-linear models recently receive a lot of attention as people are starting to discover the power of statistical and embedding features. However, tree-based models are seldom studied in the context of structured learning despite their recent success on various classification and ranking tasks. In this paper, we propose SMART , a tree-based structured learning framework based on multiple additive regression trees. SMART is especially suitable for handling tasks with dense features, and can be used to learn many different structures under various loss functions. We apply SMART to the task of tweet entity linking-a core component of tweet information extraction, which aims to identify and link name mentions to entities in a knowledge base. A novel inference algorithm is proposed to handle the special structure of the task. The experimental results show that SMART significantly outperforms state-of-the-art tweet entity linking systems.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Many natural language processing (NLP) problems can be formalized as structured prediction tasks. Standard algorithms for structured learning include Conditional Random Field (CRF) (Lafferty et al., 2001) and Structured Supported Vector Machine (SSVM) (Tsochantaridis et al., 2004) . These algorithms, usually equipped with a linear model and sparse lexical features, achieve stateof-the-art performances in many NLP applications such as part-of-speech tagging, named entity recognition and dependency parsing.", |
| "cite_spans": [ |
| { |
| "start": 181, |
| "end": 204, |
| "text": "(Lafferty et al., 2001)", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 252, |
| "end": 281, |
| "text": "(Tsochantaridis et al., 2004)", |
| "ref_id": "BIBREF35" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "This classical combination of linear models and sparse features is challenged by the recent emerg-ing usage of dense features such as statistical and embedding features. Tasks with these low dimensional dense features require models to be more sophisticated to capture the relationships between features. Therefore, non-linear models start to receive more attention as they are often more expressive than linear models.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Tree-based models such as boosted trees (Friedman, 2001 ) are flexible non-linear models. They can handle categorical features and count data better than other non-linear models like Neural Networks. Unfortunately, to the best of our knowledge, little work has utilized tree-based methods for structured prediction, with the exception of TreeCRF (Dietterich et al., 2004) .", |
| "cite_spans": [ |
| { |
| "start": 40, |
| "end": 55, |
| "text": "(Friedman, 2001", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 346, |
| "end": 371, |
| "text": "(Dietterich et al., 2004)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this paper, we propose a novel structured learning framework called S-MART (Structured Multiple Additive Regression Trees). Unlike TreeCRF, S-MART is very versatile, as it can be applied to tasks beyond sequence tagging and can be trained under various objective functions. S-MART is also powerful, as the high order relationships between features can be captured by nonlinear regression trees.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "We further demonstrate how S-MART can be applied to tweet entity linking, an important and challenging task underlying many applications including product feedback (Asur and Huberman, 2010) and topic detection and tracking (Mathioudakis and Koudas, 2010). We apply S-MART to entity linking using a simple logistic function as the loss function and propose a novel inference algorithm to prevent overlaps between entities.", |
| "cite_spans": [ |
| { |
| "start": 164, |
| "end": 189, |
| "text": "(Asur and Huberman, 2010)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Our contributions are summarized as follows:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 We propose a novel structured learning framework called S-MART. S-MART combines non-linearity and efficiency of treebased models with structured prediction, leading to a family of new algorithms. (Section 2)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 We apply S-MART to tweet entity linking. Building on top of S-MART, we propose a novel inference algorithm for nonoverlapping structure with the goal of preventing conflicting entity assignments. (Section 3)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 We provide a systematic study of evaluation criteria in tweet entity linking by conducting extensive experiments over major data sets. The results show that S-MART significantly outperforms state-of-the-art entity linking systems, including the system that is used to win the NEEL 2014 challenge (Cano and others, 2014) . (Section 4)", |
| "cite_spans": [ |
| { |
| "start": 298, |
| "end": 321, |
| "text": "(Cano and others, 2014)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The goal of a structured learning algorithm is to learn a joint scoring function S between an input x and an output structure y, S : (x, y) \u2192 R. The structured output y often contains many interdependent variables, and the number of the possible structures can be exponentially large with respect to the size of x. At test time, the prediction y for x is obtained by arg max", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "y\u2208Gen(x) S(x, y),", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "where Gen(x) represents the set of all valid output structures for x. Standard learning algorithms often directly optimize the model parameters. For example, assume that the joint scoring function S is parameterized by \u03b8. Then, gradient descent algorithms can be used to optimize the model parameters \u03b8 iteratively. More specifically,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "\u03b8 m = \u03b8 m\u22121 \u2212 \u03b7 m \u2202L(y * , S(x, y; \u03b8)) \u2202\u03b8 m\u22121 ,", |
| "eq_num": "(1)" |
| } |
| ], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "where y * is the gold structure, L(y * , S(x, y; \u03b8)) is a loss function and \u03b7 m is the learning rate of the m-th iteration.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "In this paper, we propose a framework called Structured Multiple Additive Regression Trees (S-MART), which generalizes Multiple Additive Regression Trees (MART) for structured learning problems. Different from Equation (1), S-MART does not directly optimize the model parameters; instead, it approximates the optimal scoring function that minimize the loss by adding (weighted) regression tree models iteratively.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Due to the fact that there are exponentially many input-output pairs in the training data, S-MART assumes that the joint scoring function can be decomposed as", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "S(x, y) = k\u2208\u2126(x) F (x, y k ),", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "where \u2126(x) contains the set of the all factors for input x and y k is the sub-structure of y that corresponds to the k-th factor in \u2126(x). For instance, in the task of word alignment, each factor can be defined as a pair of words from source and target languages respectively. Note that we can recover y from the union of {y k } K", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "1 . The factor scoring function F (x, y k ) can be optimized by performing gradient descent in the function space in the following manner:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "F m (x, y k ) = F m\u22121 (x, y k ) \u2212 \u03b7 m g m (x, y k ) (2)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "where function g m (x, y k ) is the functional gradient.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Note that g m is a function rather than a vector. Therefore, modeling g m theoretically requires an infinite number of data points. We can address this difficulty by approximating g m with a finite number of point-wise functional gradients", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "g m (x, y k = u k ) =", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "(3)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "\u2202L(y * , S(x, y k = u k )) \u2202F (x, y k = u k ) F (x,y k )=F m\u22121 (x,y k )", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "where u k index a valid sub-structure for the k-th factor of x. The key point of S-MART is that it approximates \u2212g m by modeling the point-wise negative functional gradients using a regression tree h m . Then the factor scoring function can be obtained by", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "F (x, y k ) = M m=1 \u03b7 m h m (x, y k ),", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "where h m (x, y k ) is also called a basis function and \u03b7 m can be simply set to 1 (Murphy, 2012).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "The detailed S-MART algorithm is presented in Algorithm 1. The factor scoring function F (x, y k ) is simply initialized to zero at first (line 1). After this, we iteratively update the function by adding regression trees. Note that the scoring function is shared by all the factors. Specifically, given the current decision function F m\u22121 , we can consider line 3 to line 9 a process of generating the pseudo Algorithm 1 S-MART: A family of structured learning algorithms with multiple additive regression trees 1: F0(x, y k ) = 0 2: for m = 1 to M do:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "going over all trees 3:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "D \u2190 \u2205 4:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "for all examples do: going over all examples 5:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "for y k \u2208 \u2126(x) do: going over all factors 6:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "For all u k , obtain g ku by Equation (3) 7:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "D \u2190 D \u222a {(\u03a6(x, y k = u k ), \u2212g ku )} 8:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "end for 9: end for 10:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "hm(x, y k ) \u2190 TrainRegressionTree(D) 11:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Fm(x, y k ) = Fm\u22121(x, y k ) + hm(x, y k ) 12: end for training data D for modeling the regression tree. For each training example, S-MART first computes the point-wise functional gradients according to Equation (3) (line 6). Here we use g ku as the abbreviation for g m (x, y k = u k ). In line 7, for each sub-structure u k , we create a new training example for the regression problem by the feature vector \u03a6(x, y k = u k ) and the negative gradient \u2212g ku . In line 10, a regression tree is constructed by minimizing differences between the prediction values and the point-wise negative gradients. Then a new basis function (modeled by a regression tree) will be added into the overall F (line 11).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "It is crucial to note that S-MART is a family of algorithms rather than a single algorithm. S-MART is flexible in the choice of the loss functions. For example, we can use either logistic loss or hinge loss, which means that S-MART can train probabilistic models as well as non-probabilistic ones. Depending on the choice of factors, S-MART can handle various structures such as linear chains, trees, and even the semi-Markov chain (Sarawagi and Cohen, 2004) .", |
| "cite_spans": [ |
| { |
| "start": 432, |
| "end": 458, |
| "text": "(Sarawagi and Cohen, 2004)", |
| "ref_id": "BIBREF30" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "S-MART versus MART There are two key differences between S-MART and MART. First, S-MART decomposes the joint scoring function S(x, y) into factors to address the problem of the exploding number of input-output pairs for structured learning problems. Second, S-MART models a single scoring function F (x, y k ) over inputs and output variables directly rather than O different functions F o (x), each of which corresponds to a label class.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Structured Multiple Additive Regression Trees", |
| "sec_num": "2" |
| }, |
| { |
| "text": "TreeCRF can be viewed as a special case of S-MART, and there are two points where S-MART improves upon TreeCRF. First, the model designed in (Dietterich et al., 2004) is tailored for sequence tagging problems. Similar to MART, for a tagging task with O tags, they choose to model O functions F o (x, o ) instead of directly modeling the joint score of the factor. This imposes limitations on the feature functions, and TreeCRF is consequently unsuitable for many tasks such as entity linking. 1 Second, S-MART is more general in terms of the objective functions and applicable structures. In the next section, we will see how S-MART can be applied to a non-linear-chain structure and various loss functions.", |
| "cite_spans": [ |
| { |
| "start": 141, |
| "end": 166, |
| "text": "(Dietterich et al., 2004)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "S-MART versus TreeCRF", |
| "sec_num": null |
| }, |
| { |
| "text": "We first formally define the task of tweet entity linking. As input, we are given a tweet, an entity database (e.g., Wikipedia where each article is an entity), and a lexicon 2 which maps a surface form into a set of entity candidates. For each incoming tweet, all n-grams of this tweet will be used to find matches in the lexicon, and each match will form a mention candidate. As output, we map every mention candidate (e.g., \"new york giants\") in the message to an entity (e.g., NEW YORK GIANTS) or to Nil (i.e., a non-entity). A mention candidate can often potentially link to multiple entities, which we call possible entity assignments.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "S-MART for Tweet Entity Linking", |
| "sec_num": "3" |
| }, |
| { |
| "text": "This task is a structured learning problem, as the final entity assignments of a tweet should not overlap with each other. 3 We decompose this learning problem as follows: we make each mention candidate a factor, and the score of the entity assignments of a tweet is the sum of the score of each entity and mention candidate pair. Although all mention candidates are decomposed, the nonoverlapping constraint requires the system to perform global inference.", |
| "cite_spans": [ |
| { |
| "start": 123, |
| "end": 124, |
| "text": "3", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "S-MART for Tweet Entity Linking", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Consider the example tweet in Figure 1 , where we show the tweet with the mention candidates in brackets. To link the mention candidate \"new york giants\" to a non-Nil entity, the system has to link previous overlapping mention candidates to Nil. It is important to note that this is not a linear chain problem because of the non-overlapping constraint, and the inference algorithm needs to be Figure 1 : Example tweet and its mention candidates. Each mention candidate is marked as a pair of brackets in the original tweet and forms a column in the graph. The graph demonstrates the non-overlapping constraint. To link the mention candidate \"new york giants\" to a non-Nil entity, the system has to link previous four overlapping mention candidates to Nil. The mention candidate \"eli manning\" is not affected by \"new york giants\". Note that this is not a standard linear chain problem.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 30, |
| "end": 38, |
| "text": "Figure 1", |
| "ref_id": null |
| }, |
| { |
| "start": 393, |
| "end": 401, |
| "text": "Figure 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "S-MART for Tweet Entity Linking", |
| "sec_num": "3" |
| }, |
| { |
| "text": "carefully designed.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "S-MART for Tweet Entity Linking", |
| "sec_num": "3" |
| }, |
| { |
| "text": "We derive specific model for tweet entity linking task with S-MART and use logistic loss as our running example. The hinge loss version of the model can be derived in a similar way.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Applying S-MART", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "Note that the tweet and the mention candidates are given. Let x be the tweet, u k be the entity assignment of the k-th mention candidate. We use function F (x, y k = u k ) to model the score of the k-th mention candidate choosing entity u k . 4 The overall scoring function can be decomposed as follows:", |
| "cite_spans": [ |
| { |
| "start": 243, |
| "end": 244, |
| "text": "4", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Applying S-MART", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "S(x, y = {u k } K k=1 ) = K k=1 F (x, y k = u k )", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Applying S-MART", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "S-MART utilizes regression trees to model the scoring function F (x, y k = u k ), which requires point-wise functional gradient for each entity of every mention candidate. Let's first write down the logistic loss function as", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Applying S-MART", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "L(y * , S(x, y)) = \u2212 log P (y * |x) = log Z(x) \u2212 S(x, y * )", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Applying S-MART", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "where Z(x) = y exp(S(x, y)) is the potential function. Then the point-wise gradients can be computed as", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Applying S-MART", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "g ku = \u2202L \u2202F (x, y k = u k ) = P (y k = u k |x) \u2212 1[y * k = u k ],", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Applying S-MART", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "where", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Applying S-MART", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "1[\u2022]", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Applying S-MART", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "represents an indicator function. The conditional probability P (y k = u k |x) can be computed by a variant of the forward-backward algorithm, which we will detail in the next subsection. 4 Note that each mention candidate has different own entity sets.", |
| "cite_spans": [ |
| { |
| "start": 188, |
| "end": 189, |
| "text": "4", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Applying S-MART", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "The non-overlapping structure is distinct from linear chain and semi-Markov chain (Sarawagi and Cohen, 2004) structures. Hence, we propose a carefully designed forward-backward algorithm to calculate P (y k = u k |x) based on current scoring function F (x, y k = u k ) given by the regression trees. The non-overlapping constraint distinguishes our inference algorithm from other forward-backward variants.", |
| "cite_spans": [ |
| { |
| "start": 82, |
| "end": 108, |
| "text": "(Sarawagi and Cohen, 2004)", |
| "ref_id": "BIBREF30" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Inference", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "To compute the forward probability, we sort 5 the mention candidates by their end indices and define forward recursion by", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Inference", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "\u03b1(u 1 , 1) = exp(F (x, y 1 = u 1 )) \u03b1(u k , k) = exp(F (x, y k = u k )) \u2022 P \u22121 p=1 exp(F (x, y k\u2212p = Nil)) \u2022 u k\u2212P \u03b1(u k\u2212P , k \u2212 P )", |
| "eq_num": "(4)" |
| } |
| ], |
| "section": "Inference", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "where k \u2212 P is the index of the previous nonoverlapping mention candidate. Intuitively, for the k-th mention candidate, we need to identify its nearest non-overlapping fellow and recursively compute the probability. The overlapping mention candidates can only take the Nil entity.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Inference", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Similarly, we can sort the mention candidates by their start indices and define backward recur-sion by", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Inference", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "\u03b2(u K , K) =1 \u03b2(u k , k) = u k+Q exp(F (x, y k+Q = u k+Q )) \u2022 Q\u22121 q=1 exp(F (x, y k+q = Nil)) \u2022 \u03b2(u k+Q , k + Q)", |
| "eq_num": "(5)" |
| } |
| ], |
| "section": "Inference", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "where k + Q is the index of the next nonoverlapping mention candidate. Note that the third terms of equation 4or (5) will vanish if there are no corresponding non-overlapping mention candidates.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Inference", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Given the potential function can be computed by", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Inference", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Z(x) = u k \u03b1(u k , k)\u03b2(u k , k)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Inference", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": ", for entities that are not Nil,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Inference", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "P (y k = u k |x) = exp(F (x, y k = u k )) \u2022 \u03b2(u k , k) Z(x) \u2022 P \u22121 p=1 exp(F (x, y k\u2212p = Nil)) \u2022 u k\u2212P \u03b1(u k\u2212P , k \u2212 P )", |
| "eq_num": "(6)" |
| } |
| ], |
| "section": "Inference", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "The probability for the special token Nil can be obtained by", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Inference", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "P (y k = Nil|x) = 1 \u2212 u k =Nil P (y k = u k |x) (7)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Inference", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "In the worst case, the total cost of the forwardbackward algorithm is O(max{T K, K 2 }), where T is the number of entities of a mention candidate. 6 Finally, at test time, the decoding problem arg max y S(x, y) can be solved by a variant of the Viterbi algorithm.", |
| "cite_spans": [ |
| { |
| "start": 147, |
| "end": 148, |
| "text": "6", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Inference", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "It is important for entity linking systems to take advantage of the entity-to-entity information while making local decisions. For instance, the identification of entity \"eli manning\" leads to a strong clue for linking \"new york giants\" to the NFL team.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Beyond S-MART: Modeling entity-entity relationships", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "Instead of defining a more complicated structure and learning everything jointly, we employ a two-stage approach as the solution for modeling entity-entity relationships after we found that S-MART achieves high precision and reasonable recall. Specifically, in the first stage, the system identifies all possible entities with basic features, which enables the extraction of entity-entity features. In the second stage, we re-train S-MART on a union of basic features and entity-entity features. We define entity-entity features based on the Jaccard distance introduced by Guo et al. (2013) .", |
| "cite_spans": [ |
| { |
| "start": 573, |
| "end": 590, |
| "text": "Guo et al. (2013)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Beyond S-MART: Modeling entity-entity relationships", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "Let \u0393(e i ) denotes the set of Wikipedia pages that contain a hyperlink to an entity e i and \u0393(t \u2212i ) denotes the set of pages that contain a hyperlink to any identified entity e j of the tweet t in the first stage excluding e i . The Jaccard distance between e i and t is", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Beyond S-MART: Modeling entity-entity relationships", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "Jac(e i , t) = |\u0393(e i ) \u2229 \u0393(t \u2212i )| |\u0393(e i ) \u222a \u0393(t \u2212i )| .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Beyond S-MART: Modeling entity-entity relationships", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "In addition to the Jaccard distance, we add one additional binary feature to indicate if the current entity has the highest Jaccard distance among all entities for this mention candidate.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Beyond S-MART: Modeling entity-entity relationships", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "Our experiments are designed to answer the following three research questions in the context of tweet entity linking:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments", |
| "sec_num": "4" |
| }, |
| { |
| "text": "\u2022 Do non-linear learning algorithms perform better than linear learning algorithms?", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments", |
| "sec_num": "4" |
| }, |
| { |
| "text": "\u2022 Do structured entity linking models perform better than non-structured ones?", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments", |
| "sec_num": "4" |
| }, |
| { |
| "text": "\u2022 How can we best capture the relationships between entities?", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments", |
| "sec_num": "4" |
| }, |
| { |
| "text": "We evaluate each entity linking system using two evaluation policies: Information Extraction (IE) driven evaluation and Information Retrieval (IR) driven evaluation. For both evaluation settings, precision, recall and F1 scores are reported. Our data is constructed from two publicly available sources: Named Entity Extraction & Linking (NEEL) Challenge datasets, and the datasets released by Fang and Chang (2014) . Note that we gather two datasets from Fang and Chang (2014) and they are used in two different evaluation settings. We refer to these two datasets as TACL-IE and TACL-IR, respectively. We perform some data cleaning and unification on these sets. 7 The statistics of the datasets are presented in Table 1 .", |
| "cite_spans": [ |
| { |
| "start": 393, |
| "end": 414, |
| "text": "Fang and Chang (2014)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 455, |
| "end": 476, |
| "text": "Fang and Chang (2014)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 663, |
| "end": 664, |
| "text": "7", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 713, |
| "end": 720, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Evaluation Methodology and Data", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "The IE-driven evaluation is the standard evaluation for an end-to-end entity linking system. We follow Carmel et al. (2014) and relax the definition of the correct mention boundaries, as they are often ambiguous. A mention boundary is considered to be correct if it overlaps (instead of being the same) with the gold mention boundary. Please see (Carmel et al., 2014) for more details on the procedure of calculating the precision, recall and F1 score. The NEEL and TACL-IE datasets have different annotation guidelines and different choices of knowledge bases, so we perform the following procedure to clean the data and unify the annotations. We first filter out the annotations that link to entities excluded by our knowledge base. We use the same knowledge base as the ERD 2014 competition (Carmel et al., 2014) , which includes the union of entities in Wikipedia and Freebase. Second, we follow NEEL annotation guideline and re-annotate TACL-IE dataset. For instance, in order to be consistent with NEEL, all the user tags (e.g. @BarackObama) are re-labeled as entities in TACL-IE.", |
| "cite_spans": [ |
| { |
| "start": 103, |
| "end": 123, |
| "text": "Carmel et al. (2014)", |
| "ref_id": null |
| }, |
| { |
| "start": 346, |
| "end": 367, |
| "text": "(Carmel et al., 2014)", |
| "ref_id": null |
| }, |
| { |
| "start": 794, |
| "end": 815, |
| "text": "(Carmel et al., 2014)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "IE-driven evaluation", |
| "sec_num": null |
| }, |
| { |
| "text": "We train all the models with NEEL Train dataset and evaluate different systems on NEEL Test and TACL-IE datasets. In addition, we sample 800 tweets from NEEL Train dataset as our development set to perform parameter tuning.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "IE-driven evaluation", |
| "sec_num": null |
| }, |
| { |
| "text": "The IR-driven evaluation is proposed by Fang and Chang (2014) . It is motivated by a key application of entity linking -retrieval of relevant tweets for target entities, which is crucial for downstream applications such as product research and sentiment analysis. In particular, given a query entity we can search for tweets based on the match with some potential surface forms of the query entity. Then, an entity linking system is evaluated by its ability to correctly identify the presence or absence of the query entity in every tweet. Our IR-driven evaluation is based on the TACL-IR set, which includes 980 tweets sampled for ten query entities of five entity types (roughly 100 tweets per entity). About 37% of the sampled tweets did not mention the query entity due to the anchor ambiguity. ", |
| "cite_spans": [ |
| { |
| "start": 40, |
| "end": 61, |
| "text": "Fang and Chang (2014)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "IR-driven evaluation", |
| "sec_num": null |
| }, |
| { |
| "text": "Features We employ a total number of 37 dense features as our basic feature set. Most of the features are adopted from (Guo et al., 2013) 8 , including various statistical features such as the probability of the surface to be used as anchor text in Wikipedia. We also add additional Entity Type features correspond to the following entity types: Character, Event, Product and Brand. Finally, we include several NER features to indicate each mention candidate belongs to one the following NER types: Twitter user, Twitter hashtag, Person, Location, Organization, Product, Event and Date.", |
| "cite_spans": [ |
| { |
| "start": 119, |
| "end": 137, |
| "text": "(Guo et al., 2013)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental Settings", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Algorithms Table 2 summarizes all the algorithms that are compared in our experiments. First, we consider two linear structured learning algorithms: Structured Perceptron (Collins, 2002) and Linear Structured SVM (SSVM) (Tsochantaridis et al., 2004) . For non-linear models, we consider polynomial SSVM, which employs polynomial kernel inside the structured SVM algorithm. We also include LambdaRank (Quoc and Le, 2007) , a neuralbased learning to rank algorithm, which is widely used in the information retrieval literature. We further compare with MART, which is designed for performing multiclass classification using log loss without considering the structured information. Finally, we have our proposed log-loss S-MART algorithm, as described in Section 3. 9 Note that our baseline systems are quite strong. Linear SSVM has been used in one of the stateof-the-art tweet entity linking systems (Guo et al., 2013) , and the system based on MART is the winning system of the 2014 NEEL Challenge (Cano and others, 2014) 10 . tured (e.g. they perform dynamic programming at test time) except for MART and LambdaRank, which treat mention candidates independently.", |
| "cite_spans": [ |
| { |
| "start": 171, |
| "end": 186, |
| "text": "(Collins, 2002)", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 220, |
| "end": 249, |
| "text": "(Tsochantaridis et al., 2004)", |
| "ref_id": "BIBREF35" |
| }, |
| { |
| "start": 400, |
| "end": 419, |
| "text": "(Quoc and Le, 2007)", |
| "ref_id": "BIBREF27" |
| }, |
| { |
| "start": 762, |
| "end": 763, |
| "text": "9", |
| "ref_id": null |
| }, |
| { |
| "start": 898, |
| "end": 916, |
| "text": "(Guo et al., 2013)", |
| "ref_id": "BIBREF16" |
| }, |
| { |
| "start": 997, |
| "end": 1020, |
| "text": "(Cano and others, 2014)", |
| "ref_id": null |
| }, |
| { |
| "start": 1021, |
| "end": 1023, |
| "text": "10", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 11, |
| "end": 18, |
| "text": "Table 2", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Experimental Settings", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Parameter tuning All the hyper-parameters are tuned on the development set. Then, we re-train our models on full training data (including the dev set) with the best parameters. We choose the soft margin parameter C from {0.5, 1, 5, 10} for two structured SVM methods. After a preliminary parameter search, we fixed the number of trees to 300 and the minimum number of documents in a leaf to 30 for all tree-based models. For LambdaRank, we use a two layer feed forward network. We select the number of hidden units from {10, 20, 30, 40} and learning rate from {0.1, 0.01, 0.001}.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental Settings", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "It is widely known that F1 score can be affected by the trade-off between precision and recall. In order to make the comparisons between all algorithms fairer in terms of F1 score, we include a post-processing step to balance precision and recall for all the systems. Note the tuning is only conducted for the purpose of robust evaluation. In particular, we adopt a simple tuning strategy that works well for all the algorithms, in which we add a bias term b to the scoring function value of Nil:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental Settings", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "F (x, y k = Nil) \u2190 F (x, y k = Nil) + b.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental Settings", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "We choose the bias term b from values between \u22123.0 to 3.0 on the dev set and apply the same bias term at test time. Table 3 presents the empirical findings for S-MART and competitive methods on tweet entity linking task in both IE and IR settings. In the following, we analyze the empirical results in details.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 116, |
| "end": 123, |
| "text": "Table 3", |
| "ref_id": "TABREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Experimental Settings", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Linear models vs. non-linear models Table 3 clearly shows that linear models perform worse than non-linear models when they are restricted to the IE setting of the tweet entity linking task. The story is similar in IR-driven evaluation, with the exception of LambdaRank. Among the linear models, linear SSVM demonstrates its superiority over Structured Perceptron on all datasets, which aligns with the results of (Tsochantaridis et al., 2005) on the named entity recognition task.", |
| "cite_spans": [ |
| { |
| "start": 414, |
| "end": 443, |
| "text": "(Tsochantaridis et al., 2005)", |
| "ref_id": "BIBREF36" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 36, |
| "end": 43, |
| "text": "Table 3", |
| "ref_id": "TABREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "We have many interesting observations on the non-linear models side. First, by adopting a polynomial kernel, the non-linear SSVM further improves the entity linking performances on the NEEL datasets and TACL-IR dataset. Second, LambdaRank, a neural network based model, achieves better results than linear models in IEdriven evaluation, but the results in IR-driven evaluation are worse than all the other methods. We believe the reason for this dismal performance is that the neural-based method tends to overfit the IR setting given the small number of training examples. Third, both MART and S-MART significantly outperform alternative linear and non-linear methods in IE-driven evaluation and performs better or similar to other methods in IR-driven evaluation. This suggests that tree-based non-linear models are suitable for tweet entity linking task. Finally, S-MART outperforms previous state-ofthe-art method Structured SVM by a surprisingly large margin. In the NEEL Test dataset, the difference is more than 10% F1. Overall, the results show that the shallow linear models are not expressive enough to capture the complex patterns in the data, which are represented by a few dense features.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Structured learning models To showcase structured learning technique is crucial for entity linking with non-linear models, we compare S-MART against MART directly. As shown in Table 3 , S-MART can achieve higher precision and recall points compared to MART on all datasets in terms of IE-driven evaluation, and can improve F1 by 4 points on NEEL Test and TACL-IR datasets. The task of entity linking is to produce non-overlapping entity assignments that match the gold mentions. By adopting structured learning technique, S-MART is able to automatically take into account the non-overlapping constraint during learning and inference, and produce global optimal entity assignments for mention candidates of a tweet. One effect is that S-MART can easily eliminate some common errors caused by popular entities (e.g. new york in Figure 1 ).", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 176, |
| "end": 183, |
| "text": "Table 3", |
| "ref_id": "TABREF4" |
| }, |
| { |
| "start": 826, |
| "end": 834, |
| "text": "Figure 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Modeling entity-entity relationships Entityentity relationships provide strong clues for entity disambiguation. In this paper, we use the simple two-stage approach described in Section 3.3 to capture the relationships between entities. As shown in Table 3 , the significant improvement in IR-driven evaluation indicates the importance of incorporating entity-entity information. Interestingly, while IR-driven results are significantly improved, IE-driven results are similar or even worse given entity-entity features. We believe the reason is that IE-driven and IR-driven evaluations focus on different aspects of tweet entity linking task. As Guo et al. (2013) shows that most mentions in tweets should be linked to the most popular entities, IE setting actually pays more attention on mention detection sub-problem. In contrast to IE setting, IR setting focuses on entity disambiguation, since we only need to decide whether the tweet is relevant to the query entity. Therefore, we believe that both evaluation policies are needed for tweet entity linking. Figure 2 shows the results of tuning the bias term for balancing precision and recall on the dev set. The results show that S-MART outperforms competitive approaches without any tuning, with similar margins to the results after tuning. Balancing precision and recall improves F1 scores for all the systems, which suggests that the simple tuning method performs quite well. Finally, we have an interesting observation that different methods have various scales of model scores.", |
| "cite_spans": [ |
| { |
| "start": 646, |
| "end": 663, |
| "text": "Guo et al. (2013)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 248, |
| "end": 255, |
| "text": "Table 3", |
| "ref_id": "TABREF4" |
| }, |
| { |
| "start": 1061, |
| "end": 1069, |
| "text": "Figure 2", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Linear structured learning methods have been proposed and widely used in the literature. Popular models include Structured Perceptron (Collins, 2002) , Conditional Random Field (Lafferty et al., 2001) and Structured SVM (Taskar et al., 2004; Tsochantaridis et al., 2005) . Recently, many structured learning models based on neural networks have been proposed and are widely used in language modeling (Bengio et al., 2006; Mikolov et al., 2010) , sentiment classification (Socher et al., 2013) , as well as parsing (Socher et al., 2011) . Cortes et al. (2014) recently proposed a boosting framework which treats different structured learning algorithms as base learners to ensemble structured prediction results.", |
| "cite_spans": [ |
| { |
| "start": 134, |
| "end": 149, |
| "text": "(Collins, 2002)", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 177, |
| "end": 200, |
| "text": "(Lafferty et al., 2001)", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 220, |
| "end": 241, |
| "text": "(Taskar et al., 2004;", |
| "ref_id": "BIBREF34" |
| }, |
| { |
| "start": 242, |
| "end": 270, |
| "text": "Tsochantaridis et al., 2005)", |
| "ref_id": "BIBREF36" |
| }, |
| { |
| "start": 400, |
| "end": 421, |
| "text": "(Bengio et al., 2006;", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 422, |
| "end": 443, |
| "text": "Mikolov et al., 2010)", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 471, |
| "end": 492, |
| "text": "(Socher et al., 2013)", |
| "ref_id": "BIBREF33" |
| }, |
| { |
| "start": 514, |
| "end": 535, |
| "text": "(Socher et al., 2011)", |
| "ref_id": "BIBREF32" |
| }, |
| { |
| "start": 538, |
| "end": 558, |
| "text": "Cortes et al. (2014)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Tree-based models have been shown to provide more robust and accurate performances than neural networks in some tasks of computer vision (Roe et al., 2005; Babenko et al., 2011) and information retrieval (Li et al., 2007; Wu et al., 2010) , suggesting that it is worth to investigate tree-based non-linear models for structured learning problems. To the best of our knowledge, TreeCRF (Dietterich et al., 2004) is the only work that explores tree-based methods for structured learning problems. The relationships between TreeCRF and our work have been discussed in Section 2.", |
| "cite_spans": [ |
| { |
| "start": 137, |
| "end": 155, |
| "text": "(Roe et al., 2005;", |
| "ref_id": "BIBREF29" |
| }, |
| { |
| "start": 156, |
| "end": 177, |
| "text": "Babenko et al., 2011)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 204, |
| "end": 221, |
| "text": "(Li et al., 2007;", |
| "ref_id": "BIBREF20" |
| }, |
| { |
| "start": 222, |
| "end": 238, |
| "text": "Wu et al., 2010)", |
| "ref_id": "BIBREF37" |
| }, |
| { |
| "start": 377, |
| "end": 410, |
| "text": "TreeCRF (Dietterich et al., 2004)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Early research on entity linking has focused on well written documents (Bunescu and Pasca, 2006; Cucerzan, 2007; Milne and Witten, 2008) . Due to the raise of social media, many techniques have been proposed or tailored to short texts including tweets, for the problem of entity linking (Ferragina and Scaiella, 2010; Meij et al., 2012; Guo et al., 2013) as well as the related problem of named entity recognition (NER) (Ritter et al., 2011) . Recently, non-textual information such as spatial and temporal signals have also been used to improve entity linking systems (Fang and Chang, 2014) . The task of entity linking has attracted a lot of attention, and many shared tasks have been hosted to promote entity linking research (Ji et al., 2010; Ji and Grishman, 2011; Cano and others, 2014; Carmel et al., 2014) .", |
| "cite_spans": [ |
| { |
| "start": 71, |
| "end": 96, |
| "text": "(Bunescu and Pasca, 2006;", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 97, |
| "end": 112, |
| "text": "Cucerzan, 2007;", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 113, |
| "end": 136, |
| "text": "Milne and Witten, 2008)", |
| "ref_id": "BIBREF25" |
| }, |
| { |
| "start": 287, |
| "end": 317, |
| "text": "(Ferragina and Scaiella, 2010;", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 318, |
| "end": 336, |
| "text": "Meij et al., 2012;", |
| "ref_id": "BIBREF23" |
| }, |
| { |
| "start": 337, |
| "end": 354, |
| "text": "Guo et al., 2013)", |
| "ref_id": "BIBREF16" |
| }, |
| { |
| "start": 420, |
| "end": 441, |
| "text": "(Ritter et al., 2011)", |
| "ref_id": "BIBREF28" |
| }, |
| { |
| "start": 569, |
| "end": 591, |
| "text": "(Fang and Chang, 2014)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 729, |
| "end": 746, |
| "text": "(Ji et al., 2010;", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 747, |
| "end": 769, |
| "text": "Ji and Grishman, 2011;", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 770, |
| "end": 792, |
| "text": "Cano and others, 2014;", |
| "ref_id": null |
| }, |
| { |
| "start": 793, |
| "end": 813, |
| "text": "Carmel et al., 2014)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Building an end-to-end entity linking system involves in solving two interrelated sub-problems: mention detection and entity disambiguation. Earlier research on entity linking has been largely focused on the entity disambiguation problem, including most work on entity linking for wellwritten documents such as news and encyclopedia articles (Cucerzan, 2007) and also few for tweets (Liu et al., 2013) . Recently, people have focused on building systems that consider mention detection and entity disambiguation jointly. For example, Cucerzan (2012) delays the mention detection decision and consider the mention detection and entity linking problem jointly. Similarly, Sil and Yates (2013) proposed to use a reranking approach to obtain overall better results on mention detection and entity disambiguation.", |
| "cite_spans": [ |
| { |
| "start": 342, |
| "end": 358, |
| "text": "(Cucerzan, 2007)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 383, |
| "end": 401, |
| "text": "(Liu et al., 2013)", |
| "ref_id": "BIBREF21" |
| }, |
| { |
| "start": 670, |
| "end": 690, |
| "text": "Sil and Yates (2013)", |
| "ref_id": "BIBREF31" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "In this paper, we propose S-MART, a family of structured learning algorithms which is flexible on the choices of the loss functions and structures. We demonstrate the power of S-MART by applying it to tweet entity linking, and it significantly outperforms the current state-of-the-art entity linking systems. In the future, we would like to investigate the advantages and disadvantages between treebased models and other non-linear models such as deep neural networks or recurrent neural networks.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion and Future Work", |
| "sec_num": "6" |
| }, |
| { |
| "text": "For example, entity linking systems need to model the similarity between an entity and the document. The TreeCRF formulation does not support such features.2 We use the standard techniques to construct the lexicon from anchor texts, redirect pages and other information resources.3 We follow the common practice and do not allow embedded entities.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "Sorting helps the algorithms find non-overlapping candidates.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "The cost is O(K 2 ) only if every mention candidate of the tweet overlaps other mention candidates. In practice, the algorithm is nearly linear w.r.t K.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "We plan to release the cleaned data and evaluation code if license permitted.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "We consider features of Base, Capitalization Rate, Popularity, Context Capitalization and Entity Type categories.9 Our pilot experiments show that the log-loss S-MART consistently outperforms the hinge-loss S-MART.10 Note that the numbers we reported here are different from the results in NEEL challenge due to the fact that we have cleaned the datasets and the evaluation metrics are slightly different in this paper.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "Acknowledgments We thank the reviewers for their insightful feedback. We also thank Yin Li and Ana Smith for their valuable comments on earlier version of this paper.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "acknowledgement", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Predicting the future with social media", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Asur", |
| "suffix": "" |
| }, |
| { |
| "first": "B", |
| "middle": [ |
| "A" |
| ], |
| "last": "Huberman", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1003.5699" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "S. Asur and B.A. Huberman. 2010. Predict- ing the future with social media. arXiv preprint arXiv:1003.5699.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Robust object tracking with online multiple instance learning. Pattern Analysis and Machine Intelligence", |
| "authors": [ |
| { |
| "first": "Boris", |
| "middle": [], |
| "last": "Babenko", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming-Hsuan", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "" |
| }, |
| { |
| "first": "Serge", |
| "middle": [], |
| "last": "Belongie", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "IEEE Transactions on", |
| "volume": "", |
| "issue": "", |
| "pages": "1619--1632", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Boris Babenko, Ming-Hsuan Yang, and Serge Be- longie. 2011. Robust object tracking with online multiple instance learning. Pattern Analysis and Machine Intelligence, IEEE Transactions on, pages 1619-1632.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Neural probabilistic language models", |
| "authors": [ |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| }, |
| { |
| "first": "Holger", |
| "middle": [], |
| "last": "Schwenk", |
| "suffix": "" |
| }, |
| { |
| "first": "Jean-S\u00e9bastien", |
| "middle": [], |
| "last": "Sen\u00e9cal", |
| "suffix": "" |
| }, |
| { |
| "first": "Fr\u00e9deric", |
| "middle": [], |
| "last": "Morin", |
| "suffix": "" |
| }, |
| { |
| "first": "Jean-Luc", |
| "middle": [], |
| "last": "Gauvain", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Innovations in Machine Learning", |
| "volume": "", |
| "issue": "", |
| "pages": "137--186", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yoshua Bengio, Holger Schwenk, Jean-S\u00e9bastien Sen\u00e9cal, Fr\u00e9deric Morin, and Jean-Luc Gauvain. 2006. Neural probabilistic language models. In In- novations in Machine Learning, pages 137-186.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Using encyclopedic knowledge for named entity disambiguation", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Bunescu", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Pasca", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of the European Chapter of the ACL (EACL)", |
| "volume": "", |
| "issue": "", |
| "pages": "9--16", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "R. C Bunescu and M. Pasca. 2006. Using encyclo- pedic knowledge for named entity disambiguation. In Proceedings of the European Chapter of the ACL (EACL), pages 9-16.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Microposts2014 neel challenge", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [ |
| "E" |
| ], |
| "last": "Cano", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Microposts2014 NEEL Challenge", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "AE Cano et al. 2014. Microposts2014 neel challenge. In Microposts2014 NEEL Challenge.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Making sense of microposts (# microposts2014) named entity extraction & linking challenge", |
| "authors": [ |
| { |
| "first": "Amparo", |
| "middle": [ |
| "E" |
| ], |
| "last": "Cano", |
| "suffix": "" |
| }, |
| { |
| "first": "Giuseppe", |
| "middle": [], |
| "last": "Rizzo", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrea", |
| "middle": [], |
| "last": "Varga", |
| "suffix": "" |
| }, |
| { |
| "first": "Matthew", |
| "middle": [], |
| "last": "Rowe", |
| "suffix": "" |
| }, |
| { |
| "first": "Milan", |
| "middle": [], |
| "last": "Stankovic", |
| "suffix": "" |
| }, |
| { |
| "first": "Aba-Sah", |
| "middle": [], |
| "last": "Dadzie", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Making Sense of Microposts (# Microp-osts2014)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Amparo E Cano, Giuseppe Rizzo, Andrea Varga, Matthew Rowe, Milan Stankovic, and Aba-Sah Dadzie. 2014. Making sense of microposts (# microposts2014) named entity extraction & linking challenge. Making Sense of Microposts (# Microp- osts2014).", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Erd'14: entity recognition and disambiguation challenge", |
| "authors": [], |
| "year": null, |
| "venue": "ACM SIGIR Forum", |
| "volume": "", |
| "issue": "", |
| "pages": "63--77", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Erd'14: entity recognition and disambiguation chal- lenge. In ACM SIGIR Forum, pages 63-77.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Discriminative training methods for hidden markov models: Theory and experiments with perceptron algorithms", |
| "authors": [ |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Collins", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "Proceedings of the conference on Empirical methods in natural language processing (EMNLP)", |
| "volume": "", |
| "issue": "", |
| "pages": "1--8", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Michael Collins. 2002. Discriminative training meth- ods for hidden markov models: Theory and exper- iments with perceptron algorithms. In Proceedings of the conference on Empirical methods in natural language processing (EMNLP), pages 1-8.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Learning ensembles of structured prediction rules", |
| "authors": [ |
| { |
| "first": "Corinna", |
| "middle": [], |
| "last": "Cortes", |
| "suffix": "" |
| }, |
| { |
| "first": "Vitaly", |
| "middle": [], |
| "last": "Kuznetsov", |
| "suffix": "" |
| }, |
| { |
| "first": "Mehryar", |
| "middle": [], |
| "last": "Mohri", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Corinna Cortes, Vitaly Kuznetsov, and Mehryar Mohri. 2014. Learning ensembles of structured prediction rules. In Proceedings of ACL.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Large-scale named entity disambiguation based on wikipedia data", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Silviu Cucerzan", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP)", |
| "volume": "", |
| "issue": "", |
| "pages": "708--716", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Silviu Cucerzan. 2007. Large-scale named entity dis- ambiguation based on wikipedia data. In Proceed- ings of the Conference on Empirical Methods in Nat- ural Language Processing (EMNLP), pages 708- 716.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "The msr system for entity linking at tac 2012", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Silviu Cucerzan", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Text Analysis Conference", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Silviu Cucerzan. 2012. The msr system for entity link- ing at tac 2012. In Text Analysis Conference.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Training conditional random fields via gradient tree boosting", |
| "authors": [ |
| { |
| "first": "G", |
| "middle": [], |
| "last": "Thomas", |
| "suffix": "" |
| }, |
| { |
| "first": "Adam", |
| "middle": [], |
| "last": "Dietterich", |
| "suffix": "" |
| }, |
| { |
| "first": "Yaroslav", |
| "middle": [], |
| "last": "Ashenfelter", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Bulatov", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proceedings of the twenty-first international conference on Machine learning (ICML)", |
| "volume": "", |
| "issue": "", |
| "pages": "28--35", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Thomas G Dietterich, Adam Ashenfelter, and Yaroslav Bulatov. 2004. Training conditional random fields via gradient tree boosting. In Proceedings of the twenty-first international conference on Machine learning (ICML), pages 28-35.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Entity linking on microblogs with spatial and temporal signals", |
| "authors": [ |
| { |
| "first": "Yuan", |
| "middle": [], |
| "last": "Fang", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming-Wei", |
| "middle": [], |
| "last": "Chang", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Transactions of the Association for Computational Linguistics (ACL)", |
| "volume": "", |
| "issue": "", |
| "pages": "259--272", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yuan Fang and Ming-Wei Chang. 2014. Entity link- ing on microblogs with spatial and temporal signals. Transactions of the Association for Computational Linguistics (ACL), pages 259-272.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "TAGME: on-thefly annotation of short text fragments (by Wikipedia entities)", |
| "authors": [ |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Ferragina", |
| "suffix": "" |
| }, |
| { |
| "first": "U", |
| "middle": [], |
| "last": "Scaiella", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of ACM Conference on Information and Knowledge Management (CIKM)", |
| "volume": "", |
| "issue": "", |
| "pages": "1625--1628", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "P. Ferragina and U. Scaiella. 2010. TAGME: on-the- fly annotation of short text fragments (by Wikipedia entities). In Proceedings of ACM Conference on Information and Knowledge Management (CIKM), pages 1625-1628.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Greedy function approximation: a gradient boosting machine", |
| "authors": [ |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Jerome", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Friedman", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Annals of Statistics", |
| "volume": "", |
| "issue": "", |
| "pages": "1189--1232", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jerome H Friedman. 2001. Greedy function approx- imation: a gradient boosting machine. Annals of Statistics, pages 1189-1232.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "To link or not to link? a study on end-to-end tweet entity linking", |
| "authors": [ |
| { |
| "first": "Stephen", |
| "middle": [], |
| "last": "Guo", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming-Wei", |
| "middle": [], |
| "last": "Chang", |
| "suffix": "" |
| }, |
| { |
| "first": "Emre", |
| "middle": [], |
| "last": "Kiciman", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the North American Chapter of the Association for Computational Linguistics (NAACL)", |
| "volume": "", |
| "issue": "", |
| "pages": "1020--1030", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Stephen Guo, Ming-Wei Chang, and Emre Kiciman. 2013. To link or not to link? a study on end-to-end tweet entity linking. In Proceedings of the North American Chapter of the Association for Computa- tional Linguistics (NAACL), pages 1020-1030.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Knowledge base population: Successful approaches and challenges", |
| "authors": [ |
| { |
| "first": "Heng", |
| "middle": [], |
| "last": "Ji", |
| "suffix": "" |
| }, |
| { |
| "first": "Ralph", |
| "middle": [], |
| "last": "Grishman", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "1148--1158", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Heng Ji and Ralph Grishman. 2011. Knowledge base population: Successful approaches and challenges. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics, pages 1148-1158.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Overview of the tac 2010 knowledge base population track", |
| "authors": [ |
| { |
| "first": "Heng", |
| "middle": [], |
| "last": "Ji", |
| "suffix": "" |
| }, |
| { |
| "first": "Ralph", |
| "middle": [], |
| "last": "Grishman", |
| "suffix": "" |
| }, |
| { |
| "first": "Hoa", |
| "middle": [ |
| "Trang" |
| ], |
| "last": "Dang", |
| "suffix": "" |
| }, |
| { |
| "first": "Kira", |
| "middle": [], |
| "last": "Griffitt", |
| "suffix": "" |
| }, |
| { |
| "first": "Joe", |
| "middle": [ |
| "Ellis" |
| ], |
| "last": "", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Third Text Analysis Conference", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Heng Ji, Ralph Grishman, Hoa Trang Dang, Kira Grif- fitt, and Joe Ellis. 2010. Overview of the tac 2010 knowledge base population track. In Third Text Analysis Conference (TAC).", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Conditional random fields: Probabilistic models for segmenting and labeling sequence data", |
| "authors": [ |
| { |
| "first": "John", |
| "middle": [], |
| "last": "Lafferty", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew", |
| "middle": [], |
| "last": "Mccallum", |
| "suffix": "" |
| }, |
| { |
| "first": "Fernando Cn", |
| "middle": [], |
| "last": "Pereira", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the 18th international conference on Machine learning (ICML)", |
| "volume": "", |
| "issue": "", |
| "pages": "282--289", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "John Lafferty, Andrew McCallum, and Fernando CN Pereira. 2001. Conditional random fields: Prob- abilistic models for segmenting and labeling se- quence data. In Proceedings of the 18th inter- national conference on Machine learning (ICML), pages 282-289.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "Mcrank: Learning to rank using multiple classification and gradient boosting", |
| "authors": [ |
| { |
| "first": "Ping", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Qiang", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [ |
| "J" |
| ], |
| "last": "Burges", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Advances in neural information processing systems (NIPS)", |
| "volume": "", |
| "issue": "", |
| "pages": "897--904", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ping Li, Qiang Wu, and Christopher J Burges. 2007. Mcrank: Learning to rank using multiple classifica- tion and gradient boosting. In Advances in neural information processing systems (NIPS), pages 897- 904.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "Entity linking for tweets", |
| "authors": [ |
| { |
| "first": "Xiaohua", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Yitong", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Haocheng", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| }, |
| { |
| "first": "Furu", |
| "middle": [], |
| "last": "Wei", |
| "suffix": "" |
| }, |
| { |
| "first": "Yi", |
| "middle": [], |
| "last": "Lu", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the Association for Computational Linguistics (ACL)", |
| "volume": "", |
| "issue": "", |
| "pages": "1304--1311", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Xiaohua Liu, Yitong Li, Haocheng Wu, Ming Zhou, Furu Wei, and Yi Lu. 2013. Entity linking for tweets. In Proceedings of the Association for Com- putational Linguistics (ACL), pages 1304-1311.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "Twittermonitor: trend detection over the twitter stream", |
| "authors": [ |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Mathioudakis", |
| "suffix": "" |
| }, |
| { |
| "first": "Nick", |
| "middle": [], |
| "last": "Koudas", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the 2010 ACM SIGMOD International Conference on Management of data (SIG-MOD)", |
| "volume": "", |
| "issue": "", |
| "pages": "1155--1158", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Michael Mathioudakis and Nick Koudas. 2010. Twit- termonitor: trend detection over the twitter stream. In Proceedings of the 2010 ACM SIGMOD Inter- national Conference on Management of data (SIG- MOD), pages 1155-1158.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "Adding semantics to microblog posts", |
| "authors": [ |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Meij", |
| "suffix": "" |
| }, |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Weerkamp", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "De Rijke", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Proceedings of International Conference on Web Search and Web Data Mining (WSDM)", |
| "volume": "", |
| "issue": "", |
| "pages": "563--572", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "E. Meij, W. Weerkamp, and M. de Rijke. 2012. Adding semantics to microblog posts. In Proceed- ings of International Conference on Web Search and Web Data Mining (WSDM), pages 563-572.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "Recurrent neural network based language model", |
| "authors": [ |
| { |
| "first": "Tomas", |
| "middle": [], |
| "last": "Mikolov", |
| "suffix": "" |
| }, |
| { |
| "first": "Martin", |
| "middle": [], |
| "last": "Karafi\u00e1t", |
| "suffix": "" |
| }, |
| { |
| "first": "Lukas", |
| "middle": [], |
| "last": "Burget", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "IN-TERSPEECH", |
| "volume": "", |
| "issue": "", |
| "pages": "1045--1048", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tomas Mikolov, Martin Karafi\u00e1t, Lukas Burget, Jan Cernock\u1ef3, and Sanjeev Khudanpur. 2010. Recur- rent neural network based language model. In IN- TERSPEECH, pages 1045-1048.", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "Learning to link with Wikipedia", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Milne", |
| "suffix": "" |
| }, |
| { |
| "first": "I", |
| "middle": [ |
| "H" |
| ], |
| "last": "Witten", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proceedings of ACM Conference on Information and Knowledge Management (CIKM)", |
| "volume": "", |
| "issue": "", |
| "pages": "509--518", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "D. Milne and I. H. Witten. 2008. Learning to link with Wikipedia. In Proceedings of ACM Conference on Information and Knowledge Management (CIKM), pages 509-518.", |
| "links": null |
| }, |
| "BIBREF26": { |
| "ref_id": "b26", |
| "title": "Machine learning: a probabilistic perspective", |
| "authors": [ |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Kevin", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Murphy", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kevin P Murphy. 2012. Machine learning: a proba- bilistic perspective. MIT press.", |
| "links": null |
| }, |
| "BIBREF27": { |
| "ref_id": "b27", |
| "title": "Learning to rank with nonsmooth cost functions", |
| "authors": [ |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Quoc", |
| "suffix": "" |
| }, |
| { |
| "first": "Viet", |
| "middle": [], |
| "last": "Le", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "193--200", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "C Quoc and Viet Le. 2007. Learning to rank with nonsmooth cost functions. pages 193-200.", |
| "links": null |
| }, |
| "BIBREF28": { |
| "ref_id": "b28", |
| "title": "Named entity recognition in tweets: an experimental study", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Ritter", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Clark", |
| "suffix": "" |
| }, |
| { |
| "first": "O", |
| "middle": [], |
| "last": "Mausam", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Etzioni", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the Conference on Empirical Methods for Natural Language Processing (EMNLP)", |
| "volume": "", |
| "issue": "", |
| "pages": "1524--1534", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "A. Ritter, S. Clark, Mausam, and O. Etzioni. 2011. Named entity recognition in tweets: an experimen- tal study. In Proceedings of the Conference on Em- pirical Methods for Natural Language Processing (EMNLP), pages 1524-1534.", |
| "links": null |
| }, |
| "BIBREF29": { |
| "ref_id": "b29", |
| "title": "Boosted decision trees as an alternative to artificial neural networks for particle identification", |
| "authors": [ |
| { |
| "first": "Hai-Jun", |
| "middle": [], |
| "last": "Byron P Roe", |
| "suffix": "" |
| }, |
| { |
| "first": "Ji", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "" |
| }, |
| { |
| "first": "Yong", |
| "middle": [], |
| "last": "Zhu", |
| "suffix": "" |
| }, |
| { |
| "first": "Ion", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Gordon", |
| "middle": [], |
| "last": "Stancu", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Mcgregor", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Nuclear Instruments and Methods in Physics Research Section A: Accelerators, Spectrometers, Detectors and Associated Equipment", |
| "volume": "", |
| "issue": "", |
| "pages": "577--584", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Byron P Roe, Hai-Jun Yang, Ji Zhu, Yong Liu, Ion Stancu, and Gordon McGregor. 2005. Boosted de- cision trees as an alternative to artificial neural net- works for particle identification. Nuclear Instru- ments and Methods in Physics Research Section A: Accelerators, Spectrometers, Detectors and Associ- ated Equipment, pages 577-584.", |
| "links": null |
| }, |
| "BIBREF30": { |
| "ref_id": "b30", |
| "title": "Semimarkov conditional random fields for information extraction", |
| "authors": [ |
| { |
| "first": "Sunita", |
| "middle": [], |
| "last": "Sarawagi", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "William W Cohen", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Advances in Neural Information Processing Systems (NIPS)", |
| "volume": "", |
| "issue": "", |
| "pages": "1185--1192", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sunita Sarawagi and William W Cohen. 2004. Semi- markov conditional random fields for information extraction. In Advances in Neural Information Pro- cessing Systems (NIPS), pages 1185-1192.", |
| "links": null |
| }, |
| "BIBREF31": { |
| "ref_id": "b31", |
| "title": "Re-ranking for joint named-entity recognition and linking", |
| "authors": [ |
| { |
| "first": "Avirup", |
| "middle": [], |
| "last": "Sil", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexander", |
| "middle": [], |
| "last": "Yates", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of ACM Conference on Information and Knowledge Management (CIKM)", |
| "volume": "", |
| "issue": "", |
| "pages": "2369--2374", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Avirup Sil and Alexander Yates. 2013. Re-ranking for joint named-entity recognition and linking. In Proceedings of ACM Conference on Information and Knowledge Management (CIKM), pages 2369- 2374.", |
| "links": null |
| }, |
| "BIBREF32": { |
| "ref_id": "b32", |
| "title": "Parsing natural scenes and natural language with recursive neural networks", |
| "authors": [ |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Socher", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Cliff", |
| "suffix": "" |
| }, |
| { |
| "first": "Chris", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew Y", |
| "middle": [], |
| "last": "Manning", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Ng", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the 28th International Conference on Machine Learning (ICML)", |
| "volume": "", |
| "issue": "", |
| "pages": "129--136", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Richard Socher, Cliff C Lin, Chris Manning, and An- drew Y Ng. 2011. Parsing natural scenes and nat- ural language with recursive neural networks. In Proceedings of the 28th International Conference on Machine Learning (ICML), pages 129-136.", |
| "links": null |
| }, |
| "BIBREF33": { |
| "ref_id": "b33", |
| "title": "Recursive deep models for semantic compositionality over a sentiment treebank", |
| "authors": [ |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Socher", |
| "suffix": "" |
| }, |
| { |
| "first": "Alex", |
| "middle": [], |
| "last": "Perelygin", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [], |
| "last": "Jean", |
| "suffix": "" |
| }, |
| { |
| "first": "Jason", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Chuang", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Christopher", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Manning", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [], |
| "last": "Andrew", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [], |
| "last": "Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Potts", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP)", |
| "volume": "", |
| "issue": "", |
| "pages": "1631--1642", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Richard Socher, Alex Perelygin, Jean Y Wu, Jason Chuang, Christopher D Manning, Andrew Y Ng, and Christopher Potts. 2013. Recursive deep mod- els for semantic compositionality over a sentiment treebank. In Proceedings of the Conference on Em- pirical Methods in Natural Language Processing (EMNLP), pages 1631-1642.", |
| "links": null |
| }, |
| "BIBREF34": { |
| "ref_id": "b34", |
| "title": "Max-margin markov networks", |
| "authors": [ |
| { |
| "first": "Ben", |
| "middle": [], |
| "last": "Taskar", |
| "suffix": "" |
| }, |
| { |
| "first": "Carlos", |
| "middle": [], |
| "last": "Guestrin", |
| "suffix": "" |
| }, |
| { |
| "first": "Daphne", |
| "middle": [], |
| "last": "Roller", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Advances in neural information processing systems", |
| "volume": "16", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ben Taskar, Carlos Guestrin, and Daphne Roller. 2004. Max-margin markov networks. Advances in neural information processing systems, 16:25.", |
| "links": null |
| }, |
| "BIBREF35": { |
| "ref_id": "b35", |
| "title": "Support vector machine learning for interdependent and structured output spaces", |
| "authors": [ |
| { |
| "first": "Ioannis", |
| "middle": [], |
| "last": "Tsochantaridis", |
| "suffix": "" |
| }, |
| { |
| "first": "Thomas", |
| "middle": [], |
| "last": "Hofmann", |
| "suffix": "" |
| }, |
| { |
| "first": "Thorsten", |
| "middle": [], |
| "last": "Joachims", |
| "suffix": "" |
| }, |
| { |
| "first": "Yasemin", |
| "middle": [], |
| "last": "Altun", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proceedings of the twentyfirst international conference on Machine learning (ICML)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ioannis Tsochantaridis, Thomas Hofmann, Thorsten Joachims, and Yasemin Altun. 2004. Support vec- tor machine learning for interdependent and struc- tured output spaces. In Proceedings of the twenty- first international conference on Machine learning (ICML), page 104.", |
| "links": null |
| }, |
| "BIBREF36": { |
| "ref_id": "b36", |
| "title": "Large margin methods for structured and interdependent output variables", |
| "authors": [ |
| { |
| "first": "Ioannis", |
| "middle": [], |
| "last": "Tsochantaridis", |
| "suffix": "" |
| }, |
| { |
| "first": "Thorsten", |
| "middle": [], |
| "last": "Joachims", |
| "suffix": "" |
| }, |
| { |
| "first": "Thomas", |
| "middle": [], |
| "last": "Hofmann", |
| "suffix": "" |
| }, |
| { |
| "first": "Yasemin", |
| "middle": [], |
| "last": "Altun", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "In Journal of Machine Learning Research", |
| "volume": "", |
| "issue": "", |
| "pages": "1453--1484", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ioannis Tsochantaridis, Thorsten Joachims, Thomas Hofmann, and Yasemin Altun. 2005. Large mar- gin methods for structured and interdependent out- put variables. In Journal of Machine Learning Re- search, pages 1453-1484.", |
| "links": null |
| }, |
| "BIBREF37": { |
| "ref_id": "b37", |
| "title": "Adapting boosting for information retrieval measures. Information Retrieval", |
| "authors": [ |
| { |
| "first": "Qiang", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [ |
| "C" |
| ], |
| "last": "Christopher", |
| "suffix": "" |
| }, |
| { |
| "first": "Krysta", |
| "middle": [ |
| "M" |
| ], |
| "last": "Burges", |
| "suffix": "" |
| }, |
| { |
| "first": "Jianfeng", |
| "middle": [], |
| "last": "Svore", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Gao", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "254--270", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Qiang Wu, Christopher JC Burges, Krysta M Svore, and Jianfeng Gao. 2010. Adapting boosting for information retrieval measures. Information Re- trieval, pages 254-270.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "type_str": "figure", |
| "num": null, |
| "uris": null, |
| "text": "Balance precisions and recalls. X-axis corresponds to values of the bias terms for the special token Nil. Note that S-MART is still the overall winning system without tuning the threshold." |
| }, |
| "TABREF1": { |
| "type_str": "table", |
| "text": "", |
| "html": null, |
| "num": null, |
| "content": "<table><tr><td>summarizes several properties of the al-</td></tr><tr><td>gorithms. For example, most algorithms are struc-</td></tr></table>" |
| }, |
| "TABREF2": { |
| "type_str": "table", |
| "text": "Included algorithms and their properties.", |
| "html": null, |
| "num": null, |
| "content": "<table/>" |
| }, |
| "TABREF4": { |
| "type_str": "table", |
| "text": "IE-driven and IR-driven evaluation results for different models. The best results with basic features are in bold. The results are underlined if adding entity-entity features gives the overall best results.", |
| "html": null, |
| "num": null, |
| "content": "<table/>" |
| } |
| } |
| } |
| } |