ACL-OCL / Base_JSON /prefixP /json /P13 /P13-1009.json
Benjamin Aw
Add updated pkl file v3
6fa4bc9
{
"paper_id": "P13-1009",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T09:34:00.144884Z"
},
"title": "Language-Independent Discriminative Parsing of Temporal Expressions",
"authors": [
{
"first": "Gabor",
"middle": [],
"last": "Angeli",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "Stanford University Stanford",
"location": {
"postCode": "94305",
"region": "CA"
}
},
"email": "angeli@stanford.edu"
},
{
"first": "Jakob",
"middle": [],
"last": "Uszkoreit",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "Google, Inc",
"location": {
"addrLine": "1600 Amphitheatre Parkway Mountain View",
"postCode": "94303",
"region": "CA"
}
},
"email": "uszkoreit@google.com"
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "Temporal resolution systems are traditionally tuned to a particular language, requiring significant human effort to translate them to new languages. We present a language independent semantic parser for learning the interpretation of temporal phrases given only a corpus of utterances and the times they reference. We make use of a latent parse that encodes a language-flexible representation of time, and extract rich features over both the parse and associated temporal semantics. The parameters of the model are learned using a weakly supervised bootstrapping approach, without the need for manually tuned parameters or any other language expertise. We achieve state-of-the-art accuracy on all languages in the TempEval-2 temporal normalization task, reporting a 4% improvement in both English and Spanish accuracy, and to our knowledge the first results for four other languages.",
"pdf_parse": {
"paper_id": "P13-1009",
"_pdf_hash": "",
"abstract": [
{
"text": "Temporal resolution systems are traditionally tuned to a particular language, requiring significant human effort to translate them to new languages. We present a language independent semantic parser for learning the interpretation of temporal phrases given only a corpus of utterances and the times they reference. We make use of a latent parse that encodes a language-flexible representation of time, and extract rich features over both the parse and associated temporal semantics. The parameters of the model are learned using a weakly supervised bootstrapping approach, without the need for manually tuned parameters or any other language expertise. We achieve state-of-the-art accuracy on all languages in the TempEval-2 temporal normalization task, reporting a 4% improvement in both English and Spanish accuracy, and to our knowledge the first results for four other languages.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "Temporal resolution is the task of mapping from a textual phrase describing a potentially complex time, date, or duration to a normalized (grounded) temporal representation. For example, possibly complex phrases such as the week before last 1 are often more useful in their grounded form -e.g., August 4 -August 11.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Many approaches to this problem make use of rule-based methods, combining regularexpression matching and hand-written interpretation functions. In contrast, we would like to learn the interpretation of a temporal expression probabilistically. This allows propagation of uncertainty to higher-level components, and the potential to dynamically back off to a rule-based system in the case of low confidence parses. In addition, we would like to use a representation of time which is broadly applicable to multiple languages, without the need for language-specific rules or manually tuned parameters.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Our system requires annotated data consisting only of an input phrase and an associated grounded time, relative to some reference time; the language-flexible parse is entirely latent. Training data of this weakly-supervised form is generally easier to collect than the alternative of manually creating and tuning potentially complex interpretation rules.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "A large number of languages conceptualize time as lying on a one dimensional line. Although the surface forms of temporal expressions differ, the basic operations many languages use can be mapped to operations on this time line (see Section 3). Furthermore, many common languages share temporal units (hours, weekdays, etc.) . By structuring a latent parse to reflect these semantics, we can define a single model which performs well on multiple languages. A discriminative parsing model allows us to define sparse features over not only lexical cues but also the temporal value of our prediction. For example, it allows us to learn that we are much more likely to express March 14 th than 2pm in Marchdespite the fact that both interpretations are composed of similar types of components. Furthermore, it allows us to define both sparse n-gram and denser but less informative bag-of-words features over multi-word phrases, and allows us to handle numbers in a flexible way.",
"cite_spans": [
{
"start": 301,
"end": 324,
"text": "(hours, weekdays, etc.)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "We briefly describe our temporal representation and grammar, followed by a description of the learning algorithm; we conclude with experimental results on the six languages of the TempEval-2 A task.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Our approach follows the work of Angeli et al. (2012) , both in the bootstrapping training methodology and the temporal grammar. Our foremost contributions over this prior work are: (i) the utilization of a discriminative parser trained with rich features; (ii) simplifications to the temporal grammar which nonetheless maintain high accuracy; and (iii) experimental results on 6 different languages, with state-of-the-art performance on both datasets on which we know of prior work.",
"cite_spans": [
{
"start": 33,
"end": 53,
"text": "Angeli et al. (2012)",
"ref_id": "BIBREF2"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "As in this previous work, our approach draws inspiration from work on semantic parsing. The latent parse parallels the formal semantics in previous work. Supervised approaches to semantic parsing prominently include Zelle and Mooney (1996) , Zettlemoyer and Collins (2005) , Kate et al. (2005) , Zettlemoyer and Collins (2007) , inter alia. For example, Zettlemoyer and Collins (2007) learn a mapping from textual queries to a logical form. Importantly, the logical form of these parses contain all of the predicates and entities used in the parse -unlike the label provided in our case, where a grounded time can correspond to any of a number of latent parses. Along this line, recent work by Clarke et al. (2010) and Liang et al. (2011) relax supervision to require only annotated answers rather than full logical forms.",
"cite_spans": [
{
"start": 216,
"end": 239,
"text": "Zelle and Mooney (1996)",
"ref_id": "BIBREF22"
},
{
"start": 242,
"end": 272,
"text": "Zettlemoyer and Collins (2005)",
"ref_id": "BIBREF23"
},
{
"start": 275,
"end": 293,
"text": "Kate et al. (2005)",
"ref_id": "BIBREF9"
},
{
"start": 296,
"end": 326,
"text": "Zettlemoyer and Collins (2007)",
"ref_id": "BIBREF24"
},
{
"start": 354,
"end": 384,
"text": "Zettlemoyer and Collins (2007)",
"ref_id": "BIBREF24"
},
{
"start": 694,
"end": 714,
"text": "Clarke et al. (2010)",
"ref_id": "BIBREF6"
},
{
"start": 719,
"end": 738,
"text": "Liang et al. (2011)",
"ref_id": "BIBREF11"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "Related work on interpreting temporal expressions has focused on constructing hand-crafted interpretation rules (Mani and Wilson, 2000; Saquete et al., 2003; Puscasu, 2004; Grover et al., 2010) . Of these, HeidelTime (Str\u00f6tgen and Gertz, 2010) and SUTime (Chang and Manning, 2012) provide a strong comparison in English.",
"cite_spans": [
{
"start": 112,
"end": 135,
"text": "(Mani and Wilson, 2000;",
"ref_id": "BIBREF13"
},
{
"start": 136,
"end": 157,
"text": "Saquete et al., 2003;",
"ref_id": "BIBREF16"
},
{
"start": 158,
"end": 172,
"text": "Puscasu, 2004;",
"ref_id": "BIBREF14"
},
{
"start": 173,
"end": 193,
"text": "Grover et al., 2010)",
"ref_id": "BIBREF8"
},
{
"start": 217,
"end": 243,
"text": "(Str\u00f6tgen and Gertz, 2010)",
"ref_id": "BIBREF17"
},
{
"start": 255,
"end": 280,
"text": "(Chang and Manning, 2012)",
"ref_id": "BIBREF4"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "Recent probabilistic approaches to temporal resolution include UzZaman and Allen (2010), who employ a parser to produce deep logical forms, in conjunction with a CRF classifier. In a similar vein, Kolomiyets and Moens (2010) employ a maximum entropy classifier to detect the location and temporal type of expressions; the grounding is then done via deterministic rules.",
"cite_spans": [
{
"start": 197,
"end": 224,
"text": "Kolomiyets and Moens (2010)",
"ref_id": "BIBREF10"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "In addition, there has been work on parsing Spanish expressions; UC3M (Vicente-D\u00edez et al., 2010) produce the strongest results on the TempEval-2 corpus. Of the systems entered in the original task, TIPSem (Llorens et al., 2010 ) was the only system to perform bilingual interpretation for English and Spanish. Both of the above systems rely primarily on hand-built rules.",
"cite_spans": [
{
"start": 206,
"end": 227,
"text": "(Llorens et al., 2010",
"ref_id": "BIBREF12"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "We define a compositional representation of time, similar to Angeli et al. (2012) , but with a greater focus on efficiency and simplicity. The representation makes use of a notion of temporal types and their associated semantic values; a grammar is constructed over these types, and is grounded by appealing to the associated values.",
"cite_spans": [
{
"start": 61,
"end": 81,
"text": "Angeli et al. (2012)",
"ref_id": "BIBREF2"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Temporal Representation",
"sec_num": "3"
},
{
"text": "A summary of the temporal type system is provided in Section 3.1; the grammar is described in Section 3.2; key modifications from previous work are highlighted in Section 3.3.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Temporal Representation",
"sec_num": "3"
},
{
"text": "Temporal expressions are represented either as a Range, Sequence, or Duration. The root of a parse tree should be one of these types. In addition, phrases can be tagged as a Function; or, as a special Nil type corresponding to segments without a direct temporal interpretation. Lastly, a type is allocated for numbers. We describe each of these briefly below.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Temporal Types",
"sec_num": "3.1"
},
{
"text": "A period between two dates (or times), as per an interval-based theory of time (Allen, 1981) . This includes entities such as Today, 1987, or Now.",
"cite_spans": [
{
"start": 79,
"end": 92,
"text": "(Allen, 1981)",
"ref_id": "BIBREF0"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Range [and Instant]",
"sec_num": null
},
{
"text": "Sequence A sequence of Ranges, occurring at regular but not necessarily constant intervals. This includes entities such as Friday, November 27 th , or last Friday. A Sequence is defined in terms of a partial completion of calendar fields. For example, November 27 th would define a Sequence whose year is unspecified, month is November, and day is the 27 th ; spanning the entire range of the lower order fields (in this case, a day). This example is illustrated in Figure 1 . Note that a Sequence implicitly selects a possibly infinite number of possible Ranges.",
"cite_spans": [],
"ref_spans": [
{
"start": 466,
"end": 474,
"text": "Figure 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "Range [and Instant]",
"sec_num": null
},
{
"text": "To select a particular grounded time for a Sequence, we appeal to a notion of a reference time (Reichenbach, 1947) . For the TempEval-2 corpus, we approximate this as the publication time of the article. While this is conflating Reichenbach's reference time with speech time, and comes at the expense of certain mistakes (see Section 5.3), it is nonetheless useful in practice.",
"cite_spans": [
{
"start": 95,
"end": 114,
"text": "(Reichenbach, 1947)",
"ref_id": "BIBREF15"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Range [and Instant]",
"sec_num": null
},
{
"text": "To a first approximation, grounding a sequence given a reference time corresponds to filling in the unspecified fields of the sequence with the fullyspecified fields of the reference time. This pro- Figure 1: An illustration of grounding a Sequence. When grounding the Sequence November 27 th with a reference time 2013-08-06 03:25:00, we complete the missing fields in the Sequence (the year) with the corresponding field in the reference time (2013).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Range [and Instant]",
"sec_num": null
},
{
"text": "cess has a number of special cases not enumerated here, 2 but the complexity remains constant time.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Range [and Instant]",
"sec_num": null
},
{
"text": "Duration A period of time. This includes entities like Week, Month, and 7 days. A special case of the Duration type is defined to represent approximate durations, such as a few years or some days.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Range [and Instant]",
"sec_num": null
},
{
"text": "Function A function of arity less than or equal to two representing some general modification to one of the above types. This captures semantic entities such as those implied in last x, the third x [of y], or x days ago. The particular functions are enumerated in Table 2 .",
"cite_spans": [],
"ref_spans": [
{
"start": 264,
"end": 271,
"text": "Table 2",
"ref_id": "TABREF3"
}
],
"eq_spans": [],
"section": "Range [and Instant]",
"sec_num": null
},
{
"text": "Nil A special Nil type denotes terms which are not directly contributing to the semantic meaning of the expression. This is intended for words such as a or the, which serve as cues without bearing temporal content themselves.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Range [and Instant]",
"sec_num": null
},
{
"text": "Number Lastly, a special Number type is defined for tagging numeric expressions.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Range [and Instant]",
"sec_num": null
},
{
"text": "Our approach assumes that natural language descriptions of time are compositional in nature; that is, each word attached to a temporal phrase is compositionally modifying the meaning of the phrase. We define a grammar jointly over temporal types and values. The types serve to constrain the parse and allow for coarse features; the values encode specific semantics, and allow for finer features. At the root of a parse tree, we recursively apply the functions in the tree to obtain a final temporal value. This approach can be presented as a rule-to-rule translation (Bach, 1976; Allen, 1995, p. 263) , or a constrained Synchronous PCFG (Yamada and Knight, 2001) .",
"cite_spans": [
{
"start": 567,
"end": 579,
"text": "(Bach, 1976;",
"ref_id": "BIBREF3"
},
{
"start": 580,
"end": 600,
"text": "Allen, 1995, p. 263)",
"ref_id": null
},
{
"start": 637,
"end": 662,
"text": "(Yamada and Knight, 2001)",
"ref_id": "BIBREF21"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Temporal Grammar",
"sec_num": "3.2"
},
{
"text": "Formally, we define our grammar as G = (\u03a3, S, V, T, R). The alphabet \u03a3 and start symbol S retain their usual interpretations. We define a set V to be the set of types, as described in Section 3.1. For each v \u2208 V we define an (infinite) set T v corresponding to the possible instances of type v. Each node in the tree defines a pair",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Temporal Grammar",
"sec_num": "3.2"
},
{
"text": "(v, \u03c4 ) such that \u03c4 \u2208 T v . A rule R \u2208 R is defined as a pair R = v i \u2192 v j v k , f : (T v j , T v k ) \u2192 T v i .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Temporal Grammar",
"sec_num": "3.2"
},
{
"text": "This definition is trivially adapted for the case of unary rules.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Temporal Grammar",
"sec_num": "3.2"
},
{
"text": "The form of our rules reveals the synchronous aspect of our grammar. The structure of the tree is bound by the first part over types v -these types are used to populate the chart, and allow for efficient inference. The second part is used to evaluate the semantics of the parse, \u03c4 \u2208 T v i , and allows partial derivations to be discriminated based on richer information than the coarse types.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Temporal Grammar",
"sec_num": "3.2"
},
{
"text": "We adopt the preterminals of Angeli et al. (2012) . Each preterminal consists of a type and a value; neither which are lexically informed. That is, the word week and preterminal (Week, Duration) are not tied in any way. A total of 62 preterminals are defined corresponding to instances of Ranges, Sequences, and Durations; these are summarized in Table 1 .",
"cite_spans": [
{
"start": 29,
"end": 49,
"text": "Angeli et al. (2012)",
"ref_id": "BIBREF2"
}
],
"ref_spans": [
{
"start": 347,
"end": 354,
"text": "Table 1",
"ref_id": "TABREF2"
}
],
"eq_spans": [],
"section": "Temporal Grammar",
"sec_num": "3.2"
},
{
"text": "In addition, 10 functions are defined for manipulating temporal expressions (see Table 2 ). The majority of these mirror generic operations on intervals on a timeline, or manipulations of a sequence. Notably, like intervals, times can be moved (3 weeks ago) or their size changed (the first two days of the month), or a new interval can be started from one of the endpoints (the last 2 days). Additionally, a sequence can be modified by shifting its origin (last Friday), or taking the n th element of the sequence within some bound (fourth Sunday in November).",
"cite_spans": [],
"ref_spans": [
{
"start": 81,
"end": 88,
"text": "Table 2",
"ref_id": "TABREF3"
}
],
"eq_spans": [],
"section": "Temporal Grammar",
"sec_num": "3.2"
},
{
"text": "Combination rules in the grammar mirror typechecked curried function application. For instance, the function moveLeft1 applied to week (as in last week) yields a grammar rule:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Temporal Grammar",
"sec_num": "3.2"
},
{
"text": "( EveryWeek -1 , Seq. ) ( moveLeft1 , Seq.\u2192Seq. ) ( EveryWeek , Seq. )",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Temporal Grammar",
"sec_num": "3.2"
},
{
"text": "In more generality, we create grammar rules for applying a function on either the left or the right, for all possible type signatures of f :",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Temporal Grammar",
"sec_num": "3.2"
},
{
"text": "f (x, y) x or x f (x, y).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Temporal Grammar",
"sec_num": "3.2"
},
{
"text": "Additionally, a grammar rule is created for intersecting two Ranges or Sequences, for multiplying a duration by a number, and for absorbing a Nil span. Each of these can be though of as an implicit function application (in the last case, the identity function).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Temporal Grammar",
"sec_num": "3.2"
},
{
"text": "While the grammar formalism is strongly inspired by Angeli et al. (2012) , a number of key differences are implemented to both simplify the framework, and make inference more efficient.",
"cite_spans": [
{
"start": 52,
"end": 72,
"text": "Angeli et al. (2012)",
"ref_id": "BIBREF2"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Differences From Previous Work",
"sec_num": "3.3"
},
{
"text": "Grounding The most timeconsuming and conceptually nuanced aspect of temporal inference in Angeli et al. (2012) is intersecting Sequences. In particular, there are two modes of expressing dates which resist intersection: a day-of-month-based mode and a week-based mode. Properly grounding a sequence which defines both a day of the month and a day of the week (or week of the year) requires backing off to an expensive search problem.",
"cite_spans": [
{
"start": 90,
"end": 110,
"text": "Angeli et al. (2012)",
"ref_id": "BIBREF2"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Sequence",
"sec_num": null
},
{
"text": "To illustrate, consider the example: Friday the 13 th . Although both a Friday and a 13 th of the month are easily found, the intersection of the two requires iterating through elements of one until it overlaps with an element of the other. At training time, a number of candidate parses are generated for each phrase. When considering that these parses can become both complex and pragmatically unreasonable, this can result in a noticeable efficiency hit; e.g., during training a sentence could have a [likely incorrect] candidate interpretation of: nineteen ninety-six Friday the 13 th s from now.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Sequence",
"sec_num": null
},
{
"text": "In our Sequence representation, such intersections are disallowed, in the same fashion as February 30 th would be.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Sequence",
"sec_num": null
},
{
"text": "Sequence Pragmatics For the sake of simplicity the pragmatic distribution over possible groundings of a sequence is replaced with the single most likely offset, as learned empirically from the English TempEval-2 corpus by Angeli et al. (2012) .",
"cite_spans": [
{
"start": 222,
"end": 242,
"text": "Angeli et al. (2012)",
"ref_id": "BIBREF2"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Sequence",
"sec_num": null
},
{
"text": "No Tag Splitting The Number and Nil types are no longer split according to their ordinality/magnitude and subsumed phrase, respectively. More precisely, there is a single nonterminal (Nil), rather than a nonterminal symbol characterizing the phrase it is subsuming (Nil-the, Nil-a, etc.). This information is encoded more elegantly as features.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Sequence",
"sec_num": null
},
{
"text": "The system is trained using a discriminative kbest parser, which is able to incorporate arbitrary features over partial derivations. We describe the parser below, followed by the features implemented.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Learning",
"sec_num": "4"
},
{
"text": "Inference A discriminative k-best parser was used to allow for arbitrary features in the parse tree. In the first stage, spans of the input sentence are tagged as either text or numbers. A rule-based number recognizer was used for each language to recognize and ground numeric expressions, including information on whether the number was an ordinal (e.g., two versus second). Note that unlike conventional parsing, a tag can span multiple words. Numeric expressions are treated as if the numeric value replaced the expression.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Parser",
"sec_num": "4.1"
},
{
"text": "Each rule of the parse derivation was assigned a score according to a log-linear factor. Specifically, each rule R = (",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Parser",
"sec_num": "4.1"
},
{
"text": "v i \u2192 v j v k , f )",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Parser",
"sec_num": "4.1"
},
{
"text": "with features over the rule and derivation so far \u03c6(R), subject to parameters \u03b8, is given a probability:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Parser",
"sec_num": "4.1"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "P (v i | v j , v k , f ; \u03b8) \u221d e \u03b8 T \u03c6(R)",
"eq_num": "(1)"
}
],
"section": "Parser",
"sec_num": "4.1"
},
{
"text": "Na\u00efvely, this parsing algorithm gives us a complexity of O(n 3 k 2 ), where n is the length of the sentence, and k is the size of the beam. However, we can approximate the algorithm in O(n 3 k log k) time with cube pruning (Chiang, 2007) . With features which are not context-free, we are not guaranteed an optimal beam with this approach; however, empirically the approximation yields a significant efficiency improvement without noticeable loss in performance.",
"cite_spans": [
{
"start": 223,
"end": 237,
"text": "(Chiang, 2007)",
"ref_id": "BIBREF5"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Parser",
"sec_num": "4.1"
},
{
"text": "Training We adopt an EM-style bootstrapping approach similar to Angeli et al. (2012) , in order to handle the task of parsing the temporal expression without annotations for the latent parses. Each training instance is a tuple consisting of the words in the temporal phrase, the annotated grounded time \u03c4 * , and the reference time.",
"cite_spans": [
{
"start": 64,
"end": 84,
"text": "Angeli et al. (2012)",
"ref_id": "BIBREF2"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Parser",
"sec_num": "4.1"
},
{
"text": "Given an input sentence, our parser will output k possible parses; when grounded to the reference time these correspond to k candidate times: \u03c4 1 . . . \u03c4 k , each with a normalized probability P (\u03c4 i ). This corresponds to an approximate E step in the EM algorithm, where the distribution over latent parses is approximated by a beam of size k. Although for long sentences the number of parses is far greater than the beam size, as the parameters improve, increasingly longer sentences will have correct derivations in the beam. In this way, a progressively larger percentage of the data is available to be learned from at each iteration.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Parser",
"sec_num": "4.1"
},
{
"text": "To approximate the M step, we define a multiclass hinge loss l(\u03b8) over the beam, and optimize using Stochastic Gradient Descent with AdaGrad (Duchi et al., 2010) :",
"cite_spans": [
{
"start": 141,
"end": 161,
"text": "(Duchi et al., 2010)",
"ref_id": "BIBREF7"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Parser",
"sec_num": "4.1"
},
{
"text": "l(\u03b8) = max 0\u2264i<k 1[\u03c4 i = \u03c4 * ] + P \u03b8 (\u03c4 i ) \u2212 P \u03b8 (\u03c4 * ) (2)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Parser",
"sec_num": "4.1"
},
{
"text": "We proceed to describe our features.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Parser",
"sec_num": "4.1"
},
{
"text": "Our framework allows us to define arbitrary features over partial derivations. Importantly, this allows us to condition not only on the PCFG probabilities over types but also the partial semantics of the derivation. We describe the features used below; a summary of these features for a short phrase is illustrated in Figure 2 .",
"cite_spans": [],
"ref_spans": [
{
"start": 318,
"end": 326,
"text": "Figure 2",
"ref_id": "FIGREF1"
}
],
"eq_spans": [],
"section": "Features",
"sec_num": "4.2"
},
{
"text": "Bracketing Features A feature is defined over every nonterminal combination, consisting of the pair of children being combined in that rule.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Features",
"sec_num": "4.2"
},
{
"text": "In particular, let us consider a rule Unlike syntactic parsing, child types of a parse tree uniquely define the parent type of the rule; this is a direct consequence of our combination rules being functions with domains defined in terms of the temporal types, and therefore necessarily projecting their inputs into a single output type. Therefore, the first class of bracketing featuresover types -reduce to have the exact same expressive power as the nonterminal CFG rules of Angeli et al. (2012) . Examples of features in this class are features 13 and 15 in Figure 2 ( A summary of the input, latent parse, and output for a particular example is given in (a). The features extracted for each fragment of the parse are given in (b), and described in detail in Section 4.2.",
"cite_spans": [
{
"start": 477,
"end": 497,
"text": "Angeli et al. (2012)",
"ref_id": "BIBREF2"
}
],
"ref_spans": [
{
"start": 561,
"end": 571,
"text": "Figure 2 (",
"ref_id": "FIGREF1"
}
],
"eq_spans": [],
"section": "Features",
"sec_num": "4.2"
},
{
"text": "R = (v i \u2192 v j v k , f ) corresponding to a CFG rule v i \u2192 v j v",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Features",
"sec_num": "4.2"
},
{
"text": "We now also have the flexibility to extract a second class of features from the semantics of the derivation. We define a feature bracketing the most recent semantic function applied to each of the two child derivations; along with the function being applied in the rule application. If the child is a preterminal, the semantics of the preterminal are used; otherwise, the outermost (most recent) function to be applied to the derivation is used. To illustrate, a tree fragment combining August and 2013 into August 2013 would yield the feature <INTERSECT, AUGUST, 2013>. This can be read as a feature for the rule applying the intersect function to August and 2013. Furthermore, intersecting August 2013 with the 12 th of the month would yield a feature <INTERSECT, INTERSECT, 12 th >. This can be read as applying the intersect function to a subtree which is the intersection of two terms, and to the 12 th of the month. Features 14 and 16 in Figure 2 In the first case, a feature is extracted over the value of the preterminal being extracted, and the phrase it is subsuming (e.g., features 1-4 and 6 in Figure 2 (b) ). As the type of the preterminal is deterministic from the value, encoding a feature on the type of the preterminal would be a coarser encoding of the same information, and is empirically not useful in this case. Since a multi-word expression can parse to a single nonterminal, a feature is extracted for the entire n-gram in addition to features for each of the individual words. For example, the phrase of this -of type Nil -would have features extracted: <NIL, of >, <NIL, this>, and <NIL, of this>.",
"cite_spans": [],
"ref_spans": [
{
"start": 944,
"end": 952,
"text": "Figure 2",
"ref_id": "FIGREF1"
},
{
"start": 1106,
"end": 1118,
"text": "Figure 2 (b)",
"ref_id": "FIGREF1"
}
],
"eq_spans": [],
"section": "Features",
"sec_num": "4.2"
},
{
"text": "In the second case -absorbing Nil-tagged spans -we extract features over the words under the Nil span joined with the type and value of the other derivation (e.g., features 7-12 in Figure 2 (b) Table 4 : Spanish results for TempEval-2 attribute scores for our system and the best known previous system. The scores are calculated using gold extents, forcing an interpretation for each parse.",
"cite_spans": [],
"ref_spans": [
{
"start": 181,
"end": 193,
"text": "Figure 2 (b)",
"ref_id": "FIGREF1"
},
{
"start": 194,
"end": 201,
"text": "Table 4",
"ref_id": null
}
],
"eq_spans": [],
"section": "Features",
"sec_num": "4.2"
},
{
"text": "<of, EVERYWEEK>, <this, EVERYWEEK>, and <of this, EVERYWEEK>.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Features",
"sec_num": "4.2"
},
{
"text": "In both cases, numbers are featurized according to their order of magnitude, and whether they are ordinal. Thus, the number tagged from thirty-first would be featurized as an ordinal number of magnitude 2.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Features",
"sec_num": "4.2"
},
{
"text": "Although some constraints can be imposed to help ensure that a top-level parse will be valid, absolute guarantees are difficult. For instance, February 30 is never a valid date; but, it would be difficult to disallow any local rule in its derivation. To mediate this, an indicator feature is extracted denoting whether the grounded semantics of the derivation is valid. This is illustrated in Figure 2 (b) by feature 17.",
"cite_spans": [],
"ref_spans": [
{
"start": 393,
"end": 401,
"text": "Figure 2",
"ref_id": "FIGREF1"
}
],
"eq_spans": [],
"section": "Semantic Validity",
"sec_num": null
},
{
"text": "Nil Bias Lastly, an indicator feature is extracted for each Nil span tagged (feature 5 in Figure 2 (b)). In part, this discourages over-generation of the type; in another part, it encourages Nil spans to absorb as many adjacent words as possible.",
"cite_spans": [],
"ref_spans": [
{
"start": 90,
"end": 98,
"text": "Figure 2",
"ref_id": "FIGREF1"
}
],
"eq_spans": [],
"section": "Semantic Validity",
"sec_num": null
},
{
"text": "We proceed to describe our experimental setup and results.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Semantic Validity",
"sec_num": null
},
{
"text": "We evaluate our model on all six languages in the TempEval-2 Task A dataset (Verhagen et al., 2010) , comparing against state-of-the-art systems for English and Spanish. New results are reported on smaller datasets from the four other languages. To our knowledge, there has not been any prior work on these corpora.",
"cite_spans": [
{
"start": 76,
"end": 99,
"text": "(Verhagen et al., 2010)",
"ref_id": "BIBREF19"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Evaluation",
"sec_num": "5"
},
{
"text": "We describe the TempEval-2 datasets in Section 5.1, present experimental results in Section 5.2, and discuss system errors in Section 5.3.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Evaluation",
"sec_num": "5"
},
{
"text": "TempEval-2, from SemEval 2010, focused on retrieving and reasoning about temporal information from newswire. Our system evaluates against Task A -detecting and resolving temporal expressions. Since we perform only the second of these, we evaluate our system assuming gold detection.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "TempEval-2 Datasets",
"sec_num": "5.1"
},
{
"text": "The dataset annotates six languages: English, Spanish, Italian, French, Chinese, and Korean; of these, English and Spanish are the most mature. We describe each of these languages, along with relevant quirks, below:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "TempEval-2 Datasets",
"sec_num": "5.1"
},
{
"text": "English The English dataset consists of 1052 training examples, and 156 test examples. Evaluation was done using the official evaluation script, which checks for exact match between TIMEX3 tags. Note that this is stricter than our training objective; for instance, 24 hours and a day have the same interpretation, but have different TIMEX3 strings. System output was heuristically converted to the TIMEX3 format; where ambiguities arose, the convention which maximized training accuracy was chosen. Table 5 : Our system's accuracy on all 6 languages of the TempEval-2 corpus. Chinese is divided into two results: one for the entire corpus, and one which considers only examples for which a temporal value is annotated. Languages with a dagger ( \u2020 ) were evaluated based on semantic rather than string-match correctness.",
"cite_spans": [],
"ref_spans": [
{
"start": 499,
"end": 506,
"text": "Table 5",
"ref_id": null
}
],
"eq_spans": [],
"section": "TempEval-2 Datasets",
"sec_num": "5.1"
},
{
"text": "to the training objective: if two TIMEX3 values ground to the same grounded time, they are considered equal. For example, in the example above, 24 hours and a day would be marked identical despite having different TIMEX3 strings.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "TempEval-2 Datasets",
"sec_num": "5.1"
},
{
"text": "Most TIMEX3 values convert naturally to a grounded representation; values with wildcards representing Sequences (e.g., 1998-QX or 1998-XX-12) ground to the same value as the Sequence encoding that value would. For instance, 1998-QX is parsed as every quarter in 1998.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "TempEval-2 Datasets",
"sec_num": "5.1"
},
{
"text": "Korean The Korean dataset consists of 287 training examples, and 91 test examples. 40 of the training examples encoded dates as a long integer For example: 003000000200001131951006 grounds to January 13, 2000 at the time 19:51. These were removed from the training set, yielding 247 examples; however, all three such examples were left in the test set. Evaluation was done identically to the Chinese data.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "TempEval-2 Datasets",
"sec_num": "5.1"
},
{
"text": "French Lastly, a dataset for French temporal expressions was compiled from the TempEval-2 data. Unlike the other 5 languages, the French data included only the raw TIMEX3 annotated newswire documents, encoded as XML. These documents were scraped to recover 206 training examples and 83 test examples. Evaluation was done identically to the Chinese and Korean data.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "TempEval-2 Datasets",
"sec_num": "5.1"
},
{
"text": "We proceed to describe our experimental results on these datasets.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "TempEval-2 Datasets",
"sec_num": "5.1"
},
{
"text": "We compare our system with state-of-the-art systems for both English and Spanish. To the best of our knowledge, no prior work exists for the other four languages.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Results",
"sec_num": "5.2"
},
{
"text": "We evaluate in the same framework as Angeli et al. (2012) . We compare to previous system scores when constrained to make a prediction on every example; if no guess is made, the output is considered incorrect. This in general yields lower results for those systems, as the system is not allowed to abstain on expressions it does not recognize.",
"cite_spans": [
{
"start": 37,
"end": 57,
"text": "Angeli et al. (2012)",
"ref_id": "BIBREF2"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Results",
"sec_num": "5.2"
},
{
"text": "The systems compared against are:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Results",
"sec_num": "5.2"
},
{
"text": "\u2022 GUTime (Mani and Wilson, 2000) , a widely used, older rule-based system.",
"cite_spans": [
{
"start": 9,
"end": 32,
"text": "(Mani and Wilson, 2000)",
"ref_id": "BIBREF13"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Results",
"sec_num": "5.2"
},
{
"text": "\u2022 HeidelTime (Str\u00f6tgen and Gertz, 2010 ), the top system at the TempEval-2 task for English.",
"cite_spans": [
{
"start": 13,
"end": 38,
"text": "(Str\u00f6tgen and Gertz, 2010",
"ref_id": "BIBREF17"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Results",
"sec_num": "5.2"
},
{
"text": "\u2022 SUTime (Chang and Manning, 2012) , a more recent rule-based system for English.",
"cite_spans": [
{
"start": 9,
"end": 34,
"text": "(Chang and Manning, 2012)",
"ref_id": "BIBREF4"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Results",
"sec_num": "5.2"
},
{
"text": "\u2022 ParsingTime (Angeli et al., 2012) , a semantic parser for temporal expressions, similar to this system (see Section 2).",
"cite_spans": [
{
"start": 14,
"end": 35,
"text": "(Angeli et al., 2012)",
"ref_id": "BIBREF2"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Results",
"sec_num": "5.2"
},
{
"text": "\u2022 UC3M (Vicente-D\u00edez et al., 2010), a rulebased system for Spanish.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Results",
"sec_num": "5.2"
},
{
"text": "Results for the English corpus are shown in Table 3. Results for Spanish are shown in Table 4 . Lastly, a summary of results in all six languages is shown in Table 5 .",
"cite_spans": [],
"ref_spans": [
{
"start": 86,
"end": 93,
"text": "Table 4",
"ref_id": null
},
{
"start": 158,
"end": 165,
"text": "Table 5",
"ref_id": null
}
],
"eq_spans": [],
"section": "Results",
"sec_num": "5.2"
},
{
"text": "A salient trend emerges from the results -while training accuracy is consistently high, test accuracy drops sharply for the data-impoverished languages. This is consistent with what would be expected from a discriminatively trained model in data-impoverished settings; however, the consistent training accuracy suggests that the model nonetheless captures the phenomena it sees in Table 6 : A summary of errors of our system, by percentage of incorrect examples for the English and Spanish datasets. The top section describes errors which could be handled in our framework, while the bottom section describes examples which are either ambiguous (missing context), or are annotated inconsistently relative the reference time.",
"cite_spans": [],
"ref_spans": [
{
"start": 381,
"end": 388,
"text": "Table 6",
"ref_id": null
}
],
"eq_spans": [],
"section": "Results",
"sec_num": "5.2"
},
{
"text": "training. This suggests the possibility for improving accuracy further by making use of more data during training.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Results",
"sec_num": "5.2"
},
{
"text": "We characterize the examples our system parses incorrectly on the English and Spanish datasets in Table 6 , expanding on each class of error below.",
"cite_spans": [],
"ref_spans": [
{
"start": 98,
"end": 105,
"text": "Table 6",
"ref_id": null
}
],
"eq_spans": [],
"section": "Discussion",
"sec_num": "5.3"
},
{
"text": "Pragmatics This class of errors is a result of pragmatic ambiguity over possible groundings of a sequence -for instance, it is often ambiguous whether next weekend refers to the coming or subsequent weekend. These errors manifest in either dropping a function (next, last), or imagining one that is not supported by the text (e.g., this week parsed as next week).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion",
"sec_num": "5.3"
},
{
"text": "Type error Another large class of errors -particularly in the English dataset -arise from not matching the annotation's type, but otherwise producing a reasonable response. For instance, the system may mistake a day on the calendar (a Range), with a day, the period of time.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion",
"sec_num": "5.3"
},
{
"text": "Incorrect number A class of mistakes arises from either omitting numbers from the parse, or incorrectly parsing numbers -the second case is particularly prevalent for written years, such as seventeen seventy-six.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion",
"sec_num": "5.3"
},
{
"text": "Relative Range These errors arise from attempting to parse a grounded Range by applying functions to the reference time. For example, from a reference time of August 8 th , it is possible to \"correctly\" parse the phrase August 1 as a week ago; but, naturally, this parse does not generalize well. This class of errors, although relatively small, merits special designation as it suggests a class of correct responses which are correct for the wrong reasons. Future work could explore mitigating these errors for domains where the text is further removed from the events it is describing than most news stories are.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion",
"sec_num": "5.3"
},
{
"text": "Incorrect parse Errors in this class are a result of failing to find the correct parse, for a number of reasons not individually identified. A small subset of these errors (notably, 6% on the Spanish dataset) are a result of the grammar being insufficiently expressive with the preterminals we defined. For instance, we cannot capture fractional units, such as in half an hour.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion",
"sec_num": "5.3"
},
{
"text": "Missing context A fairly large percentage of our errors arise from failing to classify inputs which express ambiguous or poorly defined times. For example, from time to time (annotated as the future), or that time (annotated as 5 years). Many of these require either some sort of inference or a broader understanding of the context in which the temporal phrase is uttered, which our system does not attempt to capture.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion",
"sec_num": "5.3"
},
{
"text": "Bad reference time The last class of errors cover cases where the temporal phrase is clear, but annotation differs from our judgment of what would be reasonable. These are a result of assuming that the reference time of an utterance is the publication time of the article.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion",
"sec_num": "5.3"
},
{
"text": "We have presented a discriminative, multilingual approach to resolving temporal expressions, using a language-flexible latent parse and rich features on both the types and values of partial derivations in the parse. We showed state-of-the-art results on both languages in TempEval-2 with prior work, and presented results on four additional languages.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion",
"sec_num": "6"
},
{
"text": "Spoken on, for instance, August 20.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "Some of these special cases are caused by variable days of the month, daylight savings time, etc. Another class arises from pragmatically peculiar utterances; e.g., the next Monday in August uttered in the last week of August should ground to August of next year (rather than the reference time's year).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
}
],
"back_matter": [
{
"text": "Acknowledgments Work was done in the summer of 2012 while the first author was an intern at Google. We would like to thank Chris Manning, and our co-workers at Google for their insight and help.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "acknowledgement",
"sec_num": null
}
],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "An interval-based representation of temporal knowledge",
"authors": [
{
"first": "James",
"middle": [
"F"
],
"last": "Allen",
"suffix": ""
}
],
"year": 1981,
"venue": "Proceedings of the 7th international joint conference on Artificial intelligence",
"volume": "",
"issue": "",
"pages": "221--226",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "James F. Allen. 1981. An interval-based representa- tion of temporal knowledge. In Proceedings of the 7th international joint conference on Artificial intel- ligence, pages 221-226, San Francisco, CA, USA. Morgan Kaufmann Publishers Inc.",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "Natural Language Understanding",
"authors": [
{
"first": "James",
"middle": [],
"last": "Allen",
"suffix": ""
}
],
"year": 1995,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "James Allen. 1995. Natural Language Understanding. Benjamin/Cummings, Redwood City, CA.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "Parsing time: Learning to interpret time expressions",
"authors": [
{
"first": "Gabor",
"middle": [],
"last": "Angeli",
"suffix": ""
},
{
"first": "Christopher",
"middle": [
"D"
],
"last": "Manning",
"suffix": ""
},
{
"first": "Daniel",
"middle": [],
"last": "Jurafsky",
"suffix": ""
}
],
"year": 2012,
"venue": "NAACL-HLT",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Gabor Angeli, Christopher D. Manning, and Daniel Ju- rafsky. 2012. Parsing time: Learning to interpret time expressions. In NAACL-HLT.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "An extension of classical transformational grammar",
"authors": [
{
"first": "E",
"middle": [],
"last": "Bach",
"suffix": ""
}
],
"year": 1976,
"venue": "Problems of Linguistic Metatheory (Proceedings of the 1976 Conference)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "E. Bach. 1976. An extension of classical transforma- tional grammar. In Problems of Linguistic Metathe- ory (Proceedings of the 1976 Conference), Michigan State University.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "SUTIME: a library for recognizing and normalizing time expressions",
"authors": [
{
"first": "Angel",
"middle": [],
"last": "Chang",
"suffix": ""
},
{
"first": "Chris",
"middle": [],
"last": "Manning",
"suffix": ""
}
],
"year": 2012,
"venue": "Language Resources and Evaluation",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Angel Chang and Chris Manning. 2012. SUTIME: a library for recognizing and normalizing time expres- sions. In Language Resources and Evaluation.",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "Hierarchical phrase-based translation. computational linguistics",
"authors": [
{
"first": "David",
"middle": [],
"last": "Chiang",
"suffix": ""
}
],
"year": 2007,
"venue": "",
"volume": "33",
"issue": "",
"pages": "201--228",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "David Chiang. 2007. Hierarchical phrase-based trans- lation. computational linguistics, 33(2):201-228.",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "Driving semantic parsing from the world's response",
"authors": [
{
"first": "James",
"middle": [],
"last": "Clarke",
"suffix": ""
},
{
"first": "Dan",
"middle": [],
"last": "Goldwasser",
"suffix": ""
},
{
"first": "Ming-Wei",
"middle": [],
"last": "Chang",
"suffix": ""
},
{
"first": "Dan",
"middle": [
"Roth"
],
"last": "",
"suffix": ""
}
],
"year": 2010,
"venue": "CoNLL",
"volume": "",
"issue": "",
"pages": "18--27",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "James Clarke, Dan Goldwasser, Ming-Wei Chang, and Dan Roth. 2010. Driving semantic parsing from the world's response. In CoNLL, pages 18-27, Uppsala, Sweden.",
"links": null
},
"BIBREF7": {
"ref_id": "b7",
"title": "Adaptive subgradient methods for online learning and stochastic optimization",
"authors": [
{
"first": "John",
"middle": [],
"last": "Duchi",
"suffix": ""
},
{
"first": "Elad",
"middle": [],
"last": "Hazan",
"suffix": ""
},
{
"first": "Yoram",
"middle": [],
"last": "Singer",
"suffix": ""
}
],
"year": 2010,
"venue": "Journal of Machine Learning Research",
"volume": "12",
"issue": "",
"pages": "2121--2159",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "John Duchi, Elad Hazan, and Yoram Singer. 2010. Adaptive subgradient methods for online learning and stochastic optimization. Journal of Machine Learning Research, 12:2121-2159.",
"links": null
},
"BIBREF8": {
"ref_id": "b8",
"title": "Edinburgh-LTG: TempEval-2 system description",
"authors": [
{
"first": "Claire",
"middle": [],
"last": "Grover",
"suffix": ""
},
{
"first": "Richard",
"middle": [],
"last": "Tobin",
"suffix": ""
},
{
"first": "Beatrice",
"middle": [
"Alex"
],
"last": "",
"suffix": ""
},
{
"first": "Kate",
"middle": [],
"last": "Byrne",
"suffix": ""
}
],
"year": 2010,
"venue": "Proceedings of the 5th International Workshop on Semantic Evaluation",
"volume": "",
"issue": "",
"pages": "333--336",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Claire Grover, Richard Tobin, Beatrice Alex, and Kate Byrne. 2010. Edinburgh-LTG: TempEval-2 system description. In Proceedings of the 5th International Workshop on Semantic Evaluation, Sem-Eval, pages 333-336.",
"links": null
},
"BIBREF9": {
"ref_id": "b9",
"title": "Learning to transform natural to formal languages",
"authors": [
{
"first": "J",
"middle": [],
"last": "Rohit",
"suffix": ""
},
{
"first": "Yuk",
"middle": [
"Wah"
],
"last": "Kate",
"suffix": ""
},
{
"first": "Raymond",
"middle": [
"J"
],
"last": "Wong",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Mooney",
"suffix": ""
}
],
"year": 2005,
"venue": "AAAI",
"volume": "",
"issue": "",
"pages": "1062--1068",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Rohit J. Kate, Yuk Wah Wong, and Raymond J. Mooney. 2005. Learning to transform natural to formal languages. In AAAI, pages 1062-1068, Pitts- burgh, PA.",
"links": null
},
"BIBREF10": {
"ref_id": "b10",
"title": "KUL: recognition and normalization of temporal expressions",
"authors": [
{
"first": "Oleksandr",
"middle": [],
"last": "Kolomiyets",
"suffix": ""
},
{
"first": "Marie-Francine",
"middle": [],
"last": "Moens",
"suffix": ""
}
],
"year": 2010,
"venue": "Proceedings of the 5th International Workshop on Semantic Evaluation, Sem-Eval '10",
"volume": "",
"issue": "",
"pages": "325--328",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Oleksandr Kolomiyets and Marie-Francine Moens. 2010. KUL: recognition and normalization of tem- poral expressions. In Proceedings of the 5th Interna- tional Workshop on Semantic Evaluation, Sem-Eval '10, pages 325-328.",
"links": null
},
"BIBREF11": {
"ref_id": "b11",
"title": "Learning dependency-based compositional semantics",
"authors": [
{
"first": "P",
"middle": [],
"last": "Liang",
"suffix": ""
},
{
"first": "M",
"middle": [
"I"
],
"last": "Jordan",
"suffix": ""
},
{
"first": "D",
"middle": [],
"last": "Klein",
"suffix": ""
}
],
"year": 2011,
"venue": "ACL",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "P. Liang, M. I. Jordan, and D. Klein. 2011. Learn- ing dependency-based compositional semantics. In ACL.",
"links": null
},
"BIBREF12": {
"ref_id": "b12",
"title": "Tipsem (english and spanish): Evaluating crfs and semantic roles in tempeval-2",
"authors": [
{
"first": "Hector",
"middle": [],
"last": "Llorens",
"suffix": ""
},
{
"first": "Estela",
"middle": [],
"last": "Saquete",
"suffix": ""
},
{
"first": "Borja",
"middle": [],
"last": "Navarro",
"suffix": ""
}
],
"year": 2010,
"venue": "Proceedings of the 5th International Workshop on Semantic Evaluation",
"volume": "",
"issue": "",
"pages": "284--291",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Hector Llorens, Estela Saquete, and Borja Navarro. 2010. Tipsem (english and spanish): Evaluating crfs and semantic roles in tempeval-2. In Proceedings of the 5th International Workshop on Semantic Evalu- ation, pages 284-291.",
"links": null
},
"BIBREF13": {
"ref_id": "b13",
"title": "Robust temporal processing of news",
"authors": [
{
"first": "Inderjeet",
"middle": [],
"last": "Mani",
"suffix": ""
},
{
"first": "George",
"middle": [],
"last": "Wilson",
"suffix": ""
}
],
"year": 2000,
"venue": "ACL",
"volume": "",
"issue": "",
"pages": "69--76",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Inderjeet Mani and George Wilson. 2000. Robust tem- poral processing of news. In ACL, pages 69-76, Hong Kong.",
"links": null
},
"BIBREF14": {
"ref_id": "b14",
"title": "A framework for temporal resolution",
"authors": [
{
"first": "G",
"middle": [],
"last": "Puscasu",
"suffix": ""
}
],
"year": 2004,
"venue": "LREC",
"volume": "",
"issue": "",
"pages": "1901--1904",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "G. Puscasu. 2004. A framework for temporal resolu- tion. In LREC, pages 1901-1904.",
"links": null
},
"BIBREF15": {
"ref_id": "b15",
"title": "Elements of Symbolic Logic",
"authors": [
{
"first": "Hans",
"middle": [],
"last": "Reichenbach",
"suffix": ""
}
],
"year": 1947,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Hans Reichenbach. 1947. Elements of Symbolic Logic. Macmillan, New York.",
"links": null
},
"BIBREF16": {
"ref_id": "b16",
"title": "Terseo: Temporal expression resolution system applied to event ordering",
"authors": [
{
"first": "E",
"middle": [],
"last": "Saquete",
"suffix": ""
},
{
"first": "R",
"middle": [],
"last": "Muoz",
"suffix": ""
},
{
"first": "P",
"middle": [],
"last": "Martnez-Barco",
"suffix": ""
}
],
"year": 2003,
"venue": "Text, Speech and Dialogue",
"volume": "",
"issue": "",
"pages": "220--228",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "E. Saquete, R. Muoz, and P. Martnez-Barco. 2003. Terseo: Temporal expression resolution system ap- plied to event ordering. In Text, Speech and Dia- logue, pages 220-228.",
"links": null
},
"BIBREF17": {
"ref_id": "b17",
"title": "Heideltime: High quality rule-based extraction and normalization of temporal expressions",
"authors": [
{
"first": "Jannik",
"middle": [],
"last": "Str\u00f6tgen",
"suffix": ""
},
{
"first": "Michael",
"middle": [],
"last": "Gertz",
"suffix": ""
}
],
"year": 2010,
"venue": "Proceedings of the 5th International Workshop on Semantic Evaluation",
"volume": "",
"issue": "",
"pages": "321--324",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Jannik Str\u00f6tgen and Michael Gertz. 2010. Heideltime: High quality rule-based extraction and normaliza- tion of temporal expressions. In Proceedings of the 5th International Workshop on Semantic Evaluation, Sem-Eval, pages 321-324.",
"links": null
},
"BIBREF18": {
"ref_id": "b18",
"title": "TRIPS and TRIOS system for TempEval-2: Extracting temporal information from text",
"authors": [
{
"first": "Naushad",
"middle": [],
"last": "Uzzaman",
"suffix": ""
},
{
"first": "F",
"middle": [],
"last": "James",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Allen",
"suffix": ""
}
],
"year": 2010,
"venue": "Proceedings of the 5th International Workshop on Semantic Evaluation",
"volume": "",
"issue": "",
"pages": "276--283",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Naushad UzZaman and James F. Allen. 2010. TRIPS and TRIOS system for TempEval-2: Extracting tem- poral information from text. In Proceedings of the 5th International Workshop on Semantic Evaluation, Sem-Eval, pages 276-283.",
"links": null
},
"BIBREF19": {
"ref_id": "b19",
"title": "Semeval-2010 task 13: TempEval-2",
"authors": [
{
"first": "Marc",
"middle": [],
"last": "Verhagen",
"suffix": ""
},
{
"first": "Roser",
"middle": [],
"last": "Sauri",
"suffix": ""
},
{
"first": "Tommaso",
"middle": [],
"last": "Caselli",
"suffix": ""
},
{
"first": "James",
"middle": [],
"last": "Pustejovsky",
"suffix": ""
}
],
"year": 2010,
"venue": "Proceedings of the 5th International Workshop on Semantic Evaluation",
"volume": "",
"issue": "",
"pages": "57--62",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Marc Verhagen, Roser Sauri, Tommaso Caselli, and James Pustejovsky. 2010. Semeval-2010 task 13: TempEval-2. In Proceedings of the 5th Interna- tional Workshop on Semantic Evaluation, pages 57- 62, Uppsala, Sweden.",
"links": null
},
"BIBREF20": {
"ref_id": "b20",
"title": "Uc3m system: Determining the extent, type and value of time expressions in tempeval-2",
"authors": [
{
"first": "Juli\u00e1n",
"middle": [
"Moreno"
],
"last": "Mar\u00eda Teresa Vicente-D\u00edez",
"suffix": ""
},
{
"first": "Paloma",
"middle": [],
"last": "Schneider",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Mart\u00ednez",
"suffix": ""
}
],
"year": 2010,
"venue": "proceedings of the Semantic Evaluation",
"volume": "",
"issue": "",
"pages": "Up-- psala",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Mar\u00eda Teresa Vicente-D\u00edez, Juli\u00e1n Moreno Schneider, and Paloma Mart\u00ednez. 2010. Uc3m system: De- termining the extent, type and value of time expres- sions in tempeval-2. In proceedings of the Semantic Evaluation-2 (Semeval 2010), ACL Conference, Up- psala (Sweden).",
"links": null
},
"BIBREF21": {
"ref_id": "b21",
"title": "A syntaxbased statistical translation model",
"authors": [
{
"first": "Kenji",
"middle": [],
"last": "Yamada",
"suffix": ""
},
{
"first": "Kevin",
"middle": [],
"last": "Knight",
"suffix": ""
}
],
"year": 2001,
"venue": "ACL",
"volume": "",
"issue": "",
"pages": "523--530",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Kenji Yamada and Kevin Knight. 2001. A syntax- based statistical translation model. In ACL, pages 523-530.",
"links": null
},
"BIBREF22": {
"ref_id": "b22",
"title": "Learning to parse database queries using inductive logic programming",
"authors": [
{
"first": "M",
"middle": [],
"last": "John",
"suffix": ""
},
{
"first": "Raymond",
"middle": [
"J"
],
"last": "Zelle",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Mooney",
"suffix": ""
}
],
"year": 1996,
"venue": "AAAI/IAAI",
"volume": "",
"issue": "",
"pages": "1050--1055",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "John M. Zelle and Raymond J. Mooney. 1996. Learn- ing to parse database queries using inductive logic programming. In AAAI/IAAI, pages 1050-1055, Portland, OR.",
"links": null
},
"BIBREF23": {
"ref_id": "b23",
"title": "Learning to map sentences to logical form: Structured classification with probabilistic categorial grammars",
"authors": [
{
"first": "Luke",
"middle": [
"S"
],
"last": "Zettlemoyer",
"suffix": ""
},
{
"first": "Michael",
"middle": [],
"last": "Collins",
"suffix": ""
}
],
"year": 2005,
"venue": "UAI",
"volume": "",
"issue": "",
"pages": "658--666",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Luke S. Zettlemoyer and Michael Collins. 2005. Learning to map sentences to logical form: Struc- tured classification with probabilistic categorial grammars. In UAI, pages 658-666. AUAI Press.",
"links": null
},
"BIBREF24": {
"ref_id": "b24",
"title": "Online learning of relaxed CCG grammars for parsing to logical form",
"authors": [
{
"first": "Luke",
"middle": [
"S"
],
"last": "Zettlemoyer",
"suffix": ""
},
{
"first": "Michael",
"middle": [],
"last": "Collins",
"suffix": ""
}
],
"year": 2007,
"venue": "EMNLP-CoNLL",
"volume": "",
"issue": "",
"pages": "678--687",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Luke S. Zettlemoyer and Michael Collins. 2007. On- line learning of relaxed CCG grammars for parsing to logical form. In EMNLP-CoNLL, pages 678-687.",
"links": null
}
},
"ref_entries": {
"FIGREF0": {
"num": null,
"text": "k over types, and a function f over the semantic values corresponding to v j and v k : \u03c4 j and \u03c4 k . Two classes of bracketing features are extracted: features are extracted over the types of nonterminals being combined (v j and v k ), and over the top-level semantic derivation of the nonterminals (f , \u03c4 j , and \u03c4 k ).",
"type_str": "figure",
"uris": null
},
"FIGREF1": {
"num": null,
"text": "An example parse of Friday of this week, along with the features extracted from the parse.",
"type_str": "figure",
"uris": null
},
"FIGREF2": {
"num": null,
"text": "(b) are examples of such features. Lexical Features The second large class of features extracted are lexicalized features. These are primarily used for tagging phrases with pretermi-nals; however, they are also relevant in incorporating cues from the yield of Nil spans. To illustrate, a week and the week have very different meanings, despite differing by only their Nil tagged tokens.",
"type_str": "figure",
"uris": null
},
"TABREF1": {
"type_str": "table",
"num": null,
"html": null,
"content": "<table><tr><td>Type</td><td>Example Instances</td></tr><tr><td>Range</td><td>Past, Future, Yesterday,</td></tr><tr><td/><td>Tomorrow, Today, Reference,</td></tr><tr><td/><td>Year(n), Century(n)</td></tr><tr><td>Duration</td><td>Second, Minute, Hour,</td></tr><tr><td/><td>Day, Week, Month, Quarter,</td></tr><tr><td/><td>Year, Decade, Century</td></tr></table>",
"text": "Sequence Friday, January, . . . DayOfMonth, DayOfWeek, . . . EveryDay, EveryWeek, . . ."
},
"TABREF2": {
"type_str": "table",
"num": null,
"html": null,
"content": "<table><tr><td>Function</td><td>Description</td></tr><tr><td>shiftLeft</td><td>Shift a</td></tr></table>",
"text": "The content-bearing preterminals of the grammar, arranged by their types. Note that the Sequence type contains more elements than enumerated here; however, only a few of each characteristic type are shown here for brevity. Range left by a Duration shiftRight Shift a Range right by a Duration shrinkBegin Take the first Duration of a Range shrinkEnd Take the last Duration of a Range catLeft Take the Duration after a Range catRight Take the Duration before a Range moveLeft1 Shift a Sequence left by 1 moveRight1 Shift a Sequence right by 1 n th x of y Take the n th element in y approximate Make a Duration approximate"
},
"TABREF3": {
"type_str": "table",
"num": null,
"html": null,
"content": "<table/>",
"text": "The functional preterminals of the grammar. The name and a brief description of the function are given; the functions are most easily interpreted as operations on either an interval or sequence. All operations on Ranges can equivalently be applied to Sequences."
},
"TABREF4": {
"type_str": "table",
"num": null,
"html": null,
"content": "<table><tr><td>FRI</td><td/></tr><tr><td>Input (w,t) ( Friday of this week , August 6 2013 ) Latent parse FRI \u2229 EveryWeek FRI Friday EveryWeek Nil of this EveryWeek week Output \u03c4 * August 9 2013 15. &lt; (a) Friday 1. &lt; Nil of this 2. &lt; EveryWeek week 6. &lt; EveryWeek Nil EveryWeek 7. &lt; FRI \u2229 EveryWeek FRI EveryWeek</td><td>(b)</td></tr></table>",
"text": "b). FRI , Friday > Nil , of > 3. < Nil , this > 4. < Nil , of this > 5. < nil bias > EveryWeek , week > Nil of , EveryWeek > 8. < Nil this , EveryWeek > 9. < Nil of this , EveryWeek > 10. < Nil of , Sequence > 11. < Nil this , Sequence > 12. < Nil of this , Sequence > 13. < Nil , Sequence > 14. < Nil , EveryWeek > Sequence , Sequence > 16. < Intersect , FRI , EveryWeek > 17. < root valid >"
},
"TABREF5": {
"type_str": "table",
"num": null,
"html": null,
"content": "<table><tr><td/><td>Train</td><td>Test</td></tr><tr><td>System</td><td colspan=\"2\">Type Value Type Value</td></tr><tr><td>GUTime</td><td colspan=\"2\">0.72 0.46 0.80 0.42</td></tr><tr><td>SUTime HeidelTime</td><td colspan=\"2\">0.85 0.69 0.94 0.71 0.80 0.67 0.85 0.71</td></tr><tr><td colspan=\"3\">ParsingTime 0.90 0.72 0.88 0.72</td></tr><tr><td>OurSystem</td><td colspan=\"2\">0.94 0.81 0.91 0.76</td></tr></table>",
"text": "). As above, features are extracted for both n-grams and for each word in the phrase. For example, combining of this and week would yield features"
},
"TABREF6": {
"type_str": "table",
"num": null,
"html": null,
"content": "<table><tr><td colspan=\"4\">: English results for TempEval-2 attribute</td></tr><tr><td colspan=\"4\">scores for our system and four previous systems.</td></tr><tr><td colspan=\"4\">The scores are calculated using gold extents, forc-</td></tr><tr><td colspan=\"4\">ing an interpretation for each parse.</td></tr><tr><td/><td/><td>Train</td><td>Test</td></tr><tr><td>System</td><td colspan=\"3\">Type Value Type Value</td></tr><tr><td>UC3M</td><td>-</td><td>-</td><td>0.79 0.72</td></tr><tr><td colspan=\"4\">OurSystem 0.90 0.84 0.92 0.76</td></tr></table>",
"text": ""
},
"TABREF7": {
"type_str": "table",
"num": null,
"html": null,
"content": "<table><tr><td/><td/><td>Train</td><td/><td>Test</td></tr><tr><td>Language</td><td colspan=\"4\"># examples Type Value # examples Type Value</td></tr><tr><td>English</td><td>1052</td><td>0.94 0.81</td><td>156</td><td>0.91 0.76</td></tr><tr><td>Spanish</td><td>1092</td><td>0.90 0.84</td><td>198</td><td>0.92 0.76</td></tr><tr><td>Italian</td><td>523</td><td>0.89 0.85</td><td>126</td><td>0.84 0.38</td></tr><tr><td>Chinese \u2020 Chinese (clean) \u2020 Korean \u2020 French \u2020</td><td>744 659 247 206</td><td>0.95 0.65 0.97 0.73 0.83 0.67 0.78 0.76</td><td>190 143 91 83</td><td>0.87 0.48 0.97 0.60 0.82 0.42 0.78 0.35</td></tr><tr><td/><td/><td colspan=\"3\">Italian The Italian dataset consists of 523 train-ing examples, and 126 test examples. Evaluation</td></tr><tr><td/><td/><td colspan=\"3\">was identical to English and Spanish.</td></tr><tr><td/><td/><td colspan=\"3\">Chinese The Chinese dataset consists of 744 training examples, and 190 test examples. Of</td></tr><tr><td/><td/><td colspan=\"3\">these, only 659 training and 143 test examples had</td></tr><tr><td/><td/><td colspan=\"3\">a temporal value marked; the remaining examples</td></tr><tr><td/><td/><td colspan=\"3\">had a type but no value, and are therefore impossi-</td></tr><tr><td/><td/><td colspan=\"3\">ble to predict. Results are also reported on a clean</td></tr><tr><td/><td/><td colspan=\"3\">corpus with these impossible examples omitted.</td></tr><tr><td/><td/><td colspan=\"3\">The Chinese, Korean, and French corpora had</td></tr><tr><td/><td/><td colspan=\"3\">noticeable inconsistencies in the TIMEX3 anno-</td></tr><tr><td/><td/><td colspan=\"3\">tation. Thus, evaluations are reported according</td></tr></table>",
"text": "Spanish The Spanish dataset consists of 1092 training examples, and 198 test examples. Evaluation was identical to the English, with the heuristic TIMEX3 conversion adapted somewhat."
}
}
}
}