ACL-OCL / Base_JSON /prefixN /json /N04 /N04-1013.json
Benjamin Aw
Add updated pkl file v3
6fa4bc9
{
"paper_id": "N04-1013",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T14:45:20.312110Z"
},
"title": "Speed and Accuracy in Shallow and Deep Stochastic Parsing",
"authors": [
{
"first": "Ronald",
"middle": [
"M"
],
"last": "Kaplan",
"suffix": "",
"affiliation": {},
"email": ""
},
{
"first": "Stefan",
"middle": [],
"last": "Riezler",
"suffix": "",
"affiliation": {},
"email": ""
},
{
"first": "Tracy",
"middle": [],
"last": "Holloway King",
"suffix": "",
"affiliation": {},
"email": ""
},
{
"first": "John",
"middle": [
"T"
],
"last": "Maxwell",
"suffix": "",
"affiliation": {},
"email": ""
},
{
"first": "Alexander",
"middle": [],
"last": "Vasserman",
"suffix": "",
"affiliation": {},
"email": ""
},
{
"first": "Richard",
"middle": [],
"last": "Crouch",
"suffix": "",
"affiliation": {},
"email": ""
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "This paper reports some experiments that compare the accuracy and performance of two stochastic parsing systems. The currently popular Collins parser is a shallow parser whose output contains more detailed semanticallyrelevant information than other such parsers. The XLE parser is a deep-parsing system that couples a Lexical Functional Grammar to a loglinear disambiguation component and provides much richer representations theory. We measured the accuracy of both systems against a gold standard of the PARC 700 dependency bank, and also measured their processing times. We found the deep-parsing system to be more accurate than the Collins parser with only a slight reduction in parsing speed. 1 1 This research has been funded in part by contract # MDA904-03-C-0404 awarded from the Advanced Research and Development Activity, Novel Intelligence from Massive Data program. We would like to thank Chris Culy whose original experiments inspired this research.",
"pdf_parse": {
"paper_id": "N04-1013",
"_pdf_hash": "",
"abstract": [
{
"text": "This paper reports some experiments that compare the accuracy and performance of two stochastic parsing systems. The currently popular Collins parser is a shallow parser whose output contains more detailed semanticallyrelevant information than other such parsers. The XLE parser is a deep-parsing system that couples a Lexical Functional Grammar to a loglinear disambiguation component and provides much richer representations theory. We measured the accuracy of both systems against a gold standard of the PARC 700 dependency bank, and also measured their processing times. We found the deep-parsing system to be more accurate than the Collins parser with only a slight reduction in parsing speed. 1 1 This research has been funded in part by contract # MDA904-03-C-0404 awarded from the Advanced Research and Development Activity, Novel Intelligence from Massive Data program. We would like to thank Chris Culy whose original experiments inspired this research.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "In applications that are sensitive to the meanings expressed by natural language sentences, it has become common in recent years simply to incorporate publicly available statistical parsers. A state-of-the-art statistical parsing system that enjoys great popularity in research systems is the parser described in Collins (1999) (henceforth \"the Collins parser\"). This system not only is frequently used for off-line data preprocessing, but also is included as a black-box component for applications such as document summarization (Daume and Marcu, 2002) , information extraction (Miller et al., 2000) , machine translation (Yamada and Knight, 2001) , and question answering (Harabagiu et al., 2001 ). This is be-cause the Collins parser shares the property of robustness with other statistical parsers, but more than other such parsers, the categories of its parse-trees make grammatical distinctions that presumably are useful for meaningsensitive applications. For example, the categories of the Model 3 Collins parser distinguish between heads, arguments, and adjuncts and they mark some longdistance dependency paths; these distinctions can guide application-specific postprocessors in extracting important semantic relations.",
"cite_spans": [
{
"start": 313,
"end": 327,
"text": "Collins (1999)",
"ref_id": "BIBREF1"
},
{
"start": 530,
"end": 553,
"text": "(Daume and Marcu, 2002)",
"ref_id": "BIBREF3"
},
{
"start": 579,
"end": 600,
"text": "(Miller et al., 2000)",
"ref_id": "BIBREF12"
},
{
"start": 623,
"end": 648,
"text": "(Yamada and Knight, 2001)",
"ref_id": "BIBREF17"
},
{
"start": 674,
"end": 697,
"text": "(Harabagiu et al., 2001",
"ref_id": "BIBREF7"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "In contrast, state-of-the-art parsing systems based on deep grammars mark explicitly and in much more detail a wider variety of syntactic and semantic dependencies and should therefore provide even better support for meaning-sensitive applications. But common wisdom has it that parsing systems based on deep linguistic grammars are too difficult to produce, lack coverage and robustness, and also have poor run-time performance. The Collins parser is thought to be accurate and fast and thus to represent a reasonable trade-off between \"good-enough\" output, speed, and robustness. This paper reports on some experiments that put this conventional wisdom to an empirical test. We investigated the accuracy of recovering semantically-relevant grammatical dependencies from the tree-structures produced by the Collins parser, comparing these dependencies to gold-standard dependencies which are available for a subset of 700 sentences randomly drawn from section 23 of the Wall Street Journal (see King et al. (2003) ).",
"cite_spans": [
{
"start": 996,
"end": 1014,
"text": "King et al. (2003)",
"ref_id": "BIBREF9"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "We compared the output of the XLE system, a deep-grammar-based parsing system using the English Lexical-Functional Grammar previously constructed as part of the Pargram project (Butt et al., 2002) , to the same gold standard. This system incorporates sophisticated ambiguity-management technology so that all possible syntactic analyses of a sentence are computed in an efficient, packed representation (Maxwell and Kaplan, 1993) . In accordance with LFG theory, the output includes not only standard context-free phrase-structure trees but also attribute-value matrices (LFG's f(unctional) structures) that explicitly encode predicate-argument relations and other meaningful properties. XLE selects the most probable analysis from the potentially large candidate set by means of a stochastic disambiguation component based on a log-linear (a.k.a. maximum-entropy) probability model (Riezler et al., 2002) . The stochastic component is also \"ambiguity-enabled\" in the sense that the computations for statistical estimation and selection of the most probable analyses are done efficiently by dynamic programming, avoiding the need to unpack the parse forests and enumerate individual analyses. The underlying parsing system also has built-in robustness mechanisms that allow it to parse strings that are outside the scope of the grammar as a shortest sequence of wellformed \"fragments\". Furthermore, performance parameters that bound parsing and disambiguation work can be tuned for efficient but accurate operation.",
"cite_spans": [
{
"start": 177,
"end": 196,
"text": "(Butt et al., 2002)",
"ref_id": "BIBREF0"
},
{
"start": 403,
"end": 429,
"text": "(Maxwell and Kaplan, 1993)",
"ref_id": "BIBREF11"
},
{
"start": 571,
"end": 590,
"text": "(LFG's f(unctional)",
"ref_id": null
},
{
"start": 883,
"end": 905,
"text": "(Riezler et al., 2002)",
"ref_id": "BIBREF16"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "As part of our assessment, we also measured the parsing speed of the two systems, taking into account all stages of processing that each system requires to produce its output. For example, since the Collins parser depends on a prior part-of-speech tagger (Ratnaparkhi, 1996) , we included the time for POS tagging in our Collins measurements. XLE incorporates a sophisticated finite-state morphology and dictionary lookup component, and its time is part of the measure of XLE performance.",
"cite_spans": [
{
"start": 255,
"end": 274,
"text": "(Ratnaparkhi, 1996)",
"ref_id": "BIBREF14"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Performance parameters of both the Collins parser and the XLE system were adjusted on a heldout set consisting of a random selection of 1/5 of the PARC 700 dependency bank; experimental results were then based on the other 560 sentences. For Model 3 of the Collins parser, a beam size of 1000, and not the recommended beam size of 10000, was found to optimize parsing speed at little loss in accuracy. On the same heldout set, parameters of the stochastic disambiguation system and parameters for parsing performance were adjusted for a Core and a Complete version of the XLE system, differing in the size of the constraint-set of the underlying grammar.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "For both XLE and the Collins parser we wrote conversion programs to transform the normal (tree or fstructure) output into the corresponding relations of the dependency bank. This conversion was relatively straightforward for LFG structures (King et al., 2003) . However, a certain amount of skill and intuition was required to provide a fair conversion of the Collins trees: we did not want to penalize configurations in the Collins trees that encoded alternative but equally legitimate representations of the same linguistic properties (e.g. whether auxiliaries are encoded as main verbs or aspect features), but we also did not want to build into the conversion program transformations that compensate for information that Collins cannot provide without ap-pealing to additional linguistic resources (such as identifying the subjects of infinitival complements). We did not include the time for dependency conversion in our measures of performance.",
"cite_spans": [
{
"start": 240,
"end": 259,
"text": "(King et al., 2003)",
"ref_id": "BIBREF9"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "The experimental results show that stochastic parsing with the Core LFG grammar achieves a better F-score than the Collins parser at a roughly comparable parsing speed. The XLE system achieves 12% reduction in error rate over the Collins parser, that is 77.6% F-score for the XLE system versus 74.6% for the Collins parser, at a cost in parsing time of a factor of 1.49.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "The grammar used for this experiment was developed in the ParGram project (Butt et al., 2002) . It uses LFG as a formalism, producing c(onstituent)-structures (trees) and f(unctional)-structures (attribute value matrices) as output. The c-structures encode constituency and linear order. F-structures encode predicate-argument relations and other grammatical information, e.g., number, tense, statement type. The XLE parser was used to produce packed representations, specifying all possible grammar analyses of the input.",
"cite_spans": [
{
"start": 74,
"end": 93,
"text": "(Butt et al., 2002)",
"ref_id": "BIBREF0"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Stochastic Parsing with LFG 2.1 Parsing with Lexical-Functional Grammar",
"sec_num": "2"
},
{
"text": "In our system, tokenization and morphological analysis are performed by finite-state transductions arranged in a compositional cascade. Both the tokenizer and the morphological analyzer can produce multiple outputs. For example, the tokenizer will optionaly lowercase sentence initial words, and the morphological analyzer will produce walk +Verb +Pres +3sg and walk +Noun +Pl for the input form walks. The resulting tokenized and morphologically analyzed strings are presented to the symbolic LFG grammar.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Stochastic Parsing with LFG 2.1 Parsing with Lexical-Functional Grammar",
"sec_num": "2"
},
{
"text": "The grammar can parse input that has XML delimited named entity markup: <company>Columbia Savings</company> is a major holder of so-called junk bonds. To allow the grammar to parse this markup, the tokenizer includes an additional tokenization of the strings whereby the material between the XML markup is treated as a single token with a special morphological tag (+NamedEntity). As a fall back, the tokenization that the string would have received without that markup is also produced. The named entities have a single multiword predicate. This helps in parsing both because it means that no internal structure has to be built for the predicate and because predicates that would otherwise be unrecognized by the grammar can be parsed (e.g., Cie. Financiere de Paribas). As described in section 5, it was also important to use named entity markup in these experiments to more fairly match the analyses in the PARC 700 dependency bank.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Stochastic Parsing with LFG 2.1 Parsing with Lexical-Functional Grammar",
"sec_num": "2"
},
{
"text": "To increase robustness, the standard grammar is aug-mented with a FRAGMENT grammar. This allows sentences to be parsed as well-formed chunks specified by the grammar, in particular as Ss, NPs, PPs, and VPs, with unparsable tokens possibly interspersed. These chunks have both c-structures and f-structures corresponding to them. The grammar has a fewest-chunk method for determining the correct parse. The grammar incorporates a version of Optimality Theory that allows certain (sub)rules in the grammar to be prefered or disprefered based on OT marks triggered by the (sub)rule (Frank et al., 1998) . The Complete version of the grammar uses all of the (sub)rules in a multi-pass system that depends on the ranking of the OT marks in the rules. For example, topicalization is disprefered, but the topicalization rule will be triggered if no other parse can be built. A one-line rewrite of the Complete grammar creates a Core version of the grammar that moves the majority of the OT marks into the NOGOOD space. This effectively removes the (sub)rules that they mark from the grammar. So, for example, in the Core grammar there is no topicalization rule, and sentences with topics will receive a FRAGMENT parse. This single-pass Core grammar is smaller than the Complete grammar and hence is faster.",
"cite_spans": [
{
"start": 579,
"end": 599,
"text": "(Frank et al., 1998)",
"ref_id": "BIBREF4"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Stochastic Parsing with LFG 2.1 Parsing with Lexical-Functional Grammar",
"sec_num": "2"
},
{
"text": "The XLE parser also allows the user to adjust performance parameters bounding the amount of work that is done in parsing for efficient but accurate operation. XLE's ambiguity management technology takes advantage of the fact that relatively few f-structure constraints apply to constituents that are far apart in the c-structure, so that sentences are typically parsed in polynomial time even though LFG parsing is known to be an NP-complete problem. But the worst-case exponential behavior does begin to appear for some constructions in some sentences, and the computational effort is limited by a SKIMMING mode whose onset is controlled by a user-specified parameter. When skimming, XLE will stop processing the subtree of a constituent whenever the amount of work exceeds that user-specified limit. The subtree is discarded, and the parser will move on to another subtree. This guarantees that parsing will be finished within reasonable limits of time and memory but at a cost of possibly lower accuracy if it causes the best analysis of a constituent to be discarded. As a separate parameter, XLE also lets the user limit the length of medial constituents, i.e., constituents that do not appear at the beginning or the end of a sentence (ignoring punctuation). The rationale behind this heuristic is to limit the weight of constituents in the middle of the sentence but still to allow sentence-final heavy constituents. This discards constituents in a somewhat more principled way as it tries to capture the psycholinguistic tendency to avoid deep center-embedding. When limiting the length of medial constituents, cubictime parsing is possible for sentences up to that length, even with a deep, non-context-free grammar, and linear parsing time is possible for sentences beyond that length.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Stochastic Parsing with LFG 2.1 Parsing with Lexical-Functional Grammar",
"sec_num": "2"
},
{
"text": "The Complete grammar achieved 100% coverage of section 23 as unseen unlabeled data: 79% as full parses, 21% FRAGMENT and/or SKIMMED parses.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Stochastic Parsing with LFG 2.1 Parsing with Lexical-Functional Grammar",
"sec_num": "2"
},
{
"text": "The stochastic disambiguation model we employ defines an exponential (a.k.a. log-linear or maximum-entropy) probability model over the parses of the LFG grammar. The advantage of this family of probability distributions is that it allows the user to encode arbitrary properties of the parse trees as feature-functions of the probability model, without the feature-functions needing to be independent and non-overlapping. The general form of conditional exponential models is as follows:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Dynamic Programming for Estimation and Stochastic Disambiguation",
"sec_num": "2.2"
},
{
"text": "p \u03bb (x|y) = Z \u03bb (y) \u22121 e \u03bb\u2022f (x)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Dynamic Programming for Estimation and Stochastic Disambiguation",
"sec_num": "2.2"
},
{
"text": "where Z \u03bb (y) = x\u2208X(y) e \u03bb\u2022f (x) is a normalizing constant over the set X(y) of parses for sentence y, \u03bb is a vector of log-parameters, f is a vector of featurevalues, and \u03bb \u2022 f (x) is a vector dot product denoting the (log-)weight of parse x. Dynamic-programming algorithms that allow the efficient estimation and searching of log-linear models from a packed parse representation without enumerating an exponential number of parses have been recently presented by Miyao and Tsujii (2002) and Geman and Johnson (2002) . These algorithms can be readily applied to the packed and/or-forests of Maxwell and Kaplan (1993) , provided that each conjunctive node is annotated with feature-values of the loglinear model. In the notation of Miyao and Tsujii (2002) , such a feature forest \u03a6 is defined as a tuple C, D, r, \u03b3, \u03b4 where C is a set of conjunctive nodes, D is a set of disjunctive nodes, r \u2208 C is the root node, \u03b3 : D \u2192 2 C is a conjunctive daughter function, and \u03b4 : C \u2192 2 D is a disjunctive daughter function.",
"cite_spans": [
{
"start": 465,
"end": 488,
"text": "Miyao and Tsujii (2002)",
"ref_id": "BIBREF13"
},
{
"start": 493,
"end": 517,
"text": "Geman and Johnson (2002)",
"ref_id": "BIBREF5"
},
{
"start": 592,
"end": 617,
"text": "Maxwell and Kaplan (1993)",
"ref_id": "BIBREF11"
},
{
"start": 732,
"end": 755,
"text": "Miyao and Tsujii (2002)",
"ref_id": "BIBREF13"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Dynamic Programming for Estimation and Stochastic Disambiguation",
"sec_num": "2.2"
},
{
"text": "A dynamic-programming solution to the problem of finding most probable parses is to compute the weight \u03c6 d of each disjunctive node as the maximum weight of its conjunctive daugher nodes, i.e.,",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Dynamic Programming for Estimation and Stochastic Disambiguation",
"sec_num": "2.2"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "\u03c6 d = max c\u2208\u03b3(d) \u03c6 c",
"eq_num": "(1)"
}
],
"section": "Dynamic Programming for Estimation and Stochastic Disambiguation",
"sec_num": "2.2"
},
{
"text": "and to recursively define the weight \u03c6 c of a conjunctive node as the product of the weights of all its descendant disjunctive nodes and of its own weight:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Dynamic Programming for Estimation and Stochastic Disambiguation",
"sec_num": "2.2"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "\u03c6 c = d\u2208\u03b4(c) \u03c6 d e \u03bb\u2022f (c)",
"eq_num": "(2)"
}
],
"section": "Dynamic Programming for Estimation and Stochastic Disambiguation",
"sec_num": "2.2"
},
{
"text": "Keeping a trace of the maximally weighted choices in a computaton of the weight \u03c6 r of the root conjunctive node r allows us to efficiently recover the most probable parse of a sentence from the packed representation of its parses. The same formulae can be employed for an efficient calculation of probabilistic expectations of featurefunctions for the statistical estimation of the parameters \u03bb. Replacing the maximization in equation 1 by a summation defines the inside weight of disjunctive node. Correspondingly, equation 2 denotes the inside weight of a conjunctive node. The outside weight \u03c8 c of a conjunctive node is defined as the outside weight of its disjunctive mother node(s):",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Dynamic Programming for Estimation and Stochastic Disambiguation",
"sec_num": "2.2"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "\u03c8 c = {d|c\u2208\u03b3(d)} \u03c8 d",
"eq_num": "(3)"
}
],
"section": "Dynamic Programming for Estimation and Stochastic Disambiguation",
"sec_num": "2.2"
},
{
"text": "The outside weight of a disjunctive node is the sum of the product of the outside weight(s) of its conjunctive mother(s), the weight(s) of its mother(s), and the inside weight(s) of its disjunctive sister(s):",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Dynamic Programming for Estimation and Stochastic Disambiguation",
"sec_num": "2.2"
},
{
"text": "\u03c8 d = {c|d\u2208\u03b4(c)} {\u03c8 c e \u03bb\u2022f (c) {d |d \u2208\u03b4(c),d =d} \u03c6 d } (4)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Dynamic Programming for Estimation and Stochastic Disambiguation",
"sec_num": "2.2"
},
{
"text": "From these formulae, the conditional expectation of a feature-function f i can be computed from a chart with root node r for a sentence y in the following way:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Dynamic Programming for Estimation and Stochastic Disambiguation",
"sec_num": "2.2"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "x\u2208X(y) e \u03bb\u2022f (x) f i (x) Z \u03bb (y) = c\u2208C \u03c6 c \u03c8 c f i (c) \u03c6 r",
"eq_num": "(5)"
}
],
"section": "Dynamic Programming for Estimation and Stochastic Disambiguation",
"sec_num": "2.2"
},
{
"text": "Formula 5 is used in our system to compute expectations for discriminative Bayesian estimation from partially labeled data using a first-order conjugate-gradient routine. For a more detailed description of the optimization problem and the feature-functions we use for stochastic LFG parsing see Riezler et al. (2002) . We also employed a combined 1 regularization and feature selection technique described in Riezler and Vasserman (2004) that considerably speeds up estimation and guarantees small feature sets for stochastic disambiguation. In the experiments reported in this paper, however, dynamic programming is crucial for efficient stochastic disambiguation, i.e. to efficiently find the most probable parse from a packed parse forest that is annotated with feature-values. There are two operations involved in stochastic disambiguation, namely calculating feature-values from a parse forest and calculating node weights from a feature forest.",
"cite_spans": [
{
"start": 295,
"end": 316,
"text": "Riezler et al. (2002)",
"ref_id": "BIBREF16"
},
{
"start": 409,
"end": 437,
"text": "Riezler and Vasserman (2004)",
"ref_id": "BIBREF15"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Dynamic Programming for Estimation and Stochastic Disambiguation",
"sec_num": "2.2"
},
{
"text": "Clearly, the first one is more expensive, especially for the extraction of values for non-local feature-functions over large charts. To control the cost of this computation, our stochastic disambiguation system includes a user-specified parameter for bounding the amount of work that is done in calculating feature-values. When the user-specified threshold for feature-value calculation is reached, this computation is discontinued, and the dynamic programming calculation for most-probable-parse search is computed from the current feature-value annotation of the parse forest. Since feature-value computation proceeds incrementally over the feature forest, i.e. for each node that is visited all feature-functions that apply to it are evaluated, a complete feature annotation can be guaranteed for the part of the and/or-forest that is visited until discontinuation. As discussed below, these parameters were set on a held-out portion of the PARC700 which was also used to set the Collins parameters.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Dynamic Programming for Estimation and Stochastic Disambiguation",
"sec_num": "2.2"
},
{
"text": "In the experiments reported in this paper, we used a threshold on feature-extraction that allowed us to cut off feature-extraction in 3% of the cases at no loss in accuracy. Overall, feature extraction and weight calculation accounted for 5% of the computation time in combined parsing and stochastic selection.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Dynamic Programming for Estimation and Stochastic Disambiguation",
"sec_num": "2.2"
},
{
"text": "We used the PARC 700 Dependency Bank (DEPBANK) as the gold standard in our experiments. The DEPBANK consists of dependency annotations for 700 sentences that were randomly extracted from section 23 of the UPenn Wall Street Journal (WSJ) treebank. As described by (King et al., 2003) , the annotations were boot-strapped by parsing the sentences with a LFG grammar and transforming the resulting f-structures to a collection of dependency triples in the DEPBANK format. To prepare a true gold standard of dependencies, the tentative set of dependencies produced by the robust parser was then corrected and extended by human validators 2 . In this format each triple specifies that a particular relation holds between a head and either another head or a feature value, for example, that the SUBJ relation holds between the heads run and dog in the sentence The dog ran. Average sentence length of sentences in DEPBANK is 19.8 words, and the average number of dependencies per sentence is 65.4. The corpus is freely available for research and evaluation, as are documentation and tools for displaying and pruning structures. 3 In our experiments we used a Reduced version of the DEPBANK, including just the minimum set of dependencies necessary for reading out the central semantic relations and properties of a sentence. We tested against this Reduced gold standard to establish accuracy on a lower bound of the information that a meaning-sensitive application would require. The Reduced version contained all the argument and adjunct dependencies shown in Fig. 1 , and a few selected semantically-relevant features, as shown in Fig. 2 . The features in Fig. 2 cause it was felt that they were fundamental to the meaning of the sentences, and in fact they are required by the semantic interpreter we have used in a knowledge-based application . As a concrete example, the dependency list in Fig. 3 is the Reduced set corresponding to the following sentence:",
"cite_spans": [
{
"start": 263,
"end": 282,
"text": "(King et al., 2003)",
"ref_id": "BIBREF9"
},
{
"start": 1122,
"end": 1123,
"text": "3",
"ref_id": null
}
],
"ref_spans": [
{
"start": 1555,
"end": 1562,
"text": "Fig. 1",
"ref_id": "FIGREF0"
},
{
"start": 1628,
"end": 1634,
"text": "Fig. 2",
"ref_id": "FIGREF1"
},
{
"start": 1653,
"end": 1659,
"text": "Fig. 2",
"ref_id": "FIGREF1"
},
{
"start": 1890,
"end": 1896,
"text": "Fig. 3",
"ref_id": "FIGREF2"
}
],
"eq_spans": [],
"section": "The Gold-Standard Dependency Bank",
"sec_num": "3"
},
{
"text": "He reiterated his opposition to such funding, but expressed hope of a compromise.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Gold-Standard Dependency Bank",
"sec_num": "3"
},
{
"text": "An additional feature of the DEPBANK that is relevant to our comparisons is that dependency heads are represented by their standard citation forms (e.g. the verb swam in a sentence appears as swim in its dependencies).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Gold-Standard Dependency Bank",
"sec_num": "3"
},
{
"text": "We believe that most applications will require a conversion to canonical citation forms so that semantic relations can be mapped into application-specific databases or ontologies. The predicates of LFG f-structures are already represented as citation forms; for a fair comparison we ran the leaves of the Collins tree through the same stemmer modules as part of the tree-to-dependency translation. We also note that proper names appear in the DEP-BANK as single multi-word expressions without any internal structure. That is, there are no dependencies holding among the parts of people names (A. Boyd Simpson), company names (Goldman, Sachs & Co) , and organization names (Federal Reserve). This multiword analysis was chosen because many applications do not require the internal structure of names, and the identification of named entities is now typically carried out by a separate non-syntactic pre-processing module. This was captured for the LFG parser by using named entity markup and for the Collins parser by creating complex word forms with a single POS tag (section 5).",
"cite_spans": [
{
"start": 625,
"end": 646,
"text": "(Goldman, Sachs & Co)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "The Gold-Standard Dependency Bank",
"sec_num": "3"
},
{
"text": "conj(coord\u223c0, express\u223c3) conj(coord\u223c0, reiterate\u223c1) coord form(coord\u223c0, but) stmt type(coord\u223c0, declarative) obj(reiterate\u223c1, opposition\u223c6) subj(reiterate\u223c1, pro\u223c7) tense(reiterate\u223c1, past) obj(express\u223c3, hope\u223c15) subj(express\u223c3, pro\u223c7) tense(express\u223c3, past) adjunct(opposition\u223c6, to\u223c11) num(opposition\u223c6, sg) poss(opposition\u223c6, pro\u223c19) num(pro\u223c7, sg) pron form(pro\u223c7, he) obj(to\u223c11, funding\u223c13) adjunct(funding\u223c13, such\u223c45) num(funding\u223c13, sg) adjunct(hope\u223c15, of\u223c46) num(hope\u223c15, sg) num(pro\u223c19, sg) pron form(pro\u223c19, he) adegree(such\u223c45, positive) obj(of\u223c46, compromise\u223c54) det form(compromise\u223c54, a) num(compromise\u223c54, sg) ",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Gold-Standard Dependency Bank",
"sec_num": "3"
},
{
"text": "A conversion routine was required for each system to transform its output so that it could be compared to the DEPBANK dependencies. While it is relatively straightforward to convert LFG f-structures to the dependency bank format because the f-structure is effectively a dependency format, it is more difficult to transform the output trees of the Model 3 Collins parser in a way that fairly allocates both credits and penalties.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conversion to Dependency Bank Format",
"sec_num": "4"
},
{
"text": "We discarded the LFG tree structures and used a general rewriting system previously developed for machine translation to rewrite the relevant f-structure attributes as dependencies (see King et al. (2003) ). The rewritings involved some deletions of irrelevant features, some systematic manipulations of the analyses, and some trivial respellings. The deletions involved features produced by the grammar but not included in the PARC 700 such as negative values of PASS, PERF, and PROG and the feature MEASURE used to mark measure phrases. The manipulations are more interesting and are necessary to map systematic differences between the analyses in the grammar and those in the dependency bank. For example, coordination is treated as a set by the LFG grammar but as a single COORD dependency with several CONJ relations in the dependency bank. Finally, the trivial rewritings were used to, for example, change STMT-TYPE decl in the grammar to STMT-TYPE declarative in the dependency bank. For the Reduced version of the PARC 700 substantially more features were deleted.",
"cite_spans": [
{
"start": 186,
"end": 204,
"text": "King et al. (2003)",
"ref_id": "BIBREF9"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "LFG Conversion",
"sec_num": null
},
{
"text": "An abbreviated representation of the Collins tree for the example above is shown in Fig. 4 . In this display we have eliminated the head lexical items that appear redundantly at all the nonterminals in a head chain, instead indicating by a single number which daughter is the head. Thus, S\u02dc2 indicates that the head of the main clause is its second daughter, the VP, and its head is its first VP daughter. Indirectly, then, the lexical head of the S is the first verb reiterated.",
"cite_spans": [],
"ref_spans": [
{
"start": 84,
"end": 90,
"text": "Fig. 4",
"ref_id": "FIGREF3"
}
],
"eq_spans": [],
"section": "Collins Model 3 Conversion",
"sec_num": null
},
{
"text": "(TOP\u223c1 (S\u223c2 (NP-A\u223c1 (NPB\u223c1 He/PRP)) (VP\u223c1 (VP\u223c1 reiterated/VBD (NP-A\u223c1 (NPB\u223c2 his/PRP$ opposition/NN) (PP\u223c1 to/TO (NPB\u223c2 such/JJ funding/NN)))) but/CC (VP\u223c1 expressed/VBD (NP-A\u223c1 (NPB\u223c1 hope/NN) (PP\u223c1 of/IN (NP-A\u223c1 (NPB\u223c2 a/DT compromise/NN)))))))) The Model 3 output in this example includes standard phrase structure categories, indications of the heads, and the additional -A marker to distinguish arguments from adjuncts. The terminal nodes of this tree are inflected forms, and the first phase of our conversion replaces them with their citation forms (the verbs reiterate and express, and the decapitalized and standardized he for He and his). We also adjust for systematic differences in the choice of heads. The first conjunct tends to be marked as the head of a coordination in Model 3 output, whereas the dependency bank has a more symmetric representation: it introduces a new COORD head and connects that up to the conjunction, and it uses a separate CONJ relation for each of the coordinated items. Similarly, Model 3 identifies the syntactic markers to and that as the heads of complements, whereas the dependency bank treats these as selectional features and marks the main predicate of the complements as the head. These adjustments are carried out without penalty. We also compensate for the differences in the representation of auxiliaries: Model 3 treats these as main verbs with embedded complements instead of the PERF, PROG, and PASSIVE features of the DEP-BANK, and our conversion flattens the trees so that the features can be read off.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Collins Model 3 Conversion",
"sec_num": null
},
{
"text": "The dependencies are read off after these and a few other adjustments are made. NPs under VPs are read off either as objects or adjuncts, depending on whether or not the NP is annotated with the argument indicator (-A) as in this example; the -A presumably would be missing in a sentence like John arrived Friday, and Friday would be treated as an ADJUNCT. Similarly, NP-As under S are read off as subject. In this example, however, this principle of conversion does not lead to a match with the dependency bank: in the DEPBANK grammatical relations that are factored out of conjoined structures are distributed back into those structures, to establish the correct semantic dependencies (in this case, that he is the subject of both reiterate and express and not of the introduced coord). We avoided the temptation of building coordinate distribution into the conversion routine because, first, it is not always obvious from the Model 3 output when distribution should take place, and second, that would be a first step towards building into the conversion routine the deep lexical and syntactic knowledge (essentially the functional component of our LFG grammar) that the shallow approach explicitly discounts 4 .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Collins Model 3 Conversion",
"sec_num": null
},
{
"text": "For the same reasons our conversion routine does not identify the subjects of infinitival complements with particular arguments of matrix verbs. The Model 3 trees provide no indication of how this is to be done, and in many cases the proper assignment depends on lexical information about specific predicates (to capture, for example, the well-known contrast between promise and persuade).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Collins Model 3 Conversion",
"sec_num": null
},
{
"text": "Model 3 trees also provide information about certain long-distance dependencies, by marking with -g annotations the path between a filler and a gap and marking the gap by an explicit TRACE in the terminal string. The filler itself is not clearly identified, but our conversion treats all WH categories under SBAR as potential fillers and attempts to propagate them down the gap-chain to link them up to appropriate traces. In sum, it is not a trivial matter to convert a Model 3 tree to an appropriate set of dependency relations, and the process requires a certain amount of intuition and skill. For our experiments we tried to define a conversion that gives appropriate credit to the dependencies that can be read from the trees without relying on an undue amount of sophisticated linguistic knowledge 5 .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Collins Model 3 Conversion",
"sec_num": null
},
{
"text": "We conducted our experiments by preparing versions of the test sentences in the form appropriate to each system. We used a configuration of the XLE parser that expects sentences conforming to ordinary text conventions to appear in a file separated by double line-feeds. A certain amount of effort was required to remove the part-ofspeech tags and labeled brackets of the WSJ corpus in a way that restored the sentences to a standard English format (for example, to remove the space between wo and n't that remains when the POS tags are removed). Since the PARC 700 treats proper names as multiword expressions, we then augmented the input strings with XML markup of the named entities. These are parsed by the grammar as described in section 2. We used manual named entity markup for this experiment because our intent is to measure parsing technology independent of either the time or errors of an automatic named-entity extractor. However, in other experiments with an automatic finite-state extractor, we have found that the time for named-entity recognition is negligible (on the order of seconds across the entire corpus) and makes relatively few errors, so that the results reported here are good approximations of what might be expected in more realistic situations.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experiments",
"sec_num": "5"
},
{
"text": "As input to the Collins parser, we used the part-ofspeech tagged version of section 23 that was provided with the parser. From this we extracted the 700 sentences in the PARC 700. We then modified them to produce named entity input so that the parses would match the PARC 700. This was done by putting underscores between the parts of the named entity and changing the final part of speech tag to the appropriate one (usually NNP) if necessary. (The number of words indicated at the beginning of the input string was also reduced accordingly.) An example is shown in (1).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experiments",
"sec_num": "5"
},
{
"text": "After parsing, the underscores were converted to spaces to match the PARC 700 predicates.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experiments",
"sec_num": "5"
},
{
"text": "Before the final evaluation, 1/5 of the PARC 700 dependency bank was randomly extracted as a heldout set. This set was used to adjust the performance parameters of the XLE system and the Collins parser so as to optimize parsing speed without losing accuracy. For example, the limit on the length of medial phrases was set to 20 words for the XLE system (see Sec. 2), and a regularizer penalty of 10 was found optimal for the 1 prior used in stochastic disambiguation. For the Collins parser, a beam size of 1000 was found to improve speed considerably at little cost in accuracy. Furthermore, the np-bracketing flag (npbflag) was set to 0 to produce an extended set of NP levels for improved argument/adjunct distinction 6 . The final evaluation was done on the remaining 560 examples. Timing results are reported in seconds of CPU time 7 . POS tagging of the input to the Collins parser took 6 seconds and this was added to the timing result of the Collins parser. Time spent for finite-state morphology and dictionary lookup for XLE is part of the measure of its timing performance. We did not include the time for dependency extraction or stemming the Collins output. Table 1 shows timing and accuracy results for the Reduced dependency set. The parser settings compared are Model 3 of the Collins parser adjusted to beam size 1000, and the Core and Complete versions of the XLE system, differing in the size of the grammar's constraintset. Clearly, both versions of the XLE system achieve a significant reduction in error rate over the Collins parser (12% for the core XLE system and 20% for the complete system) at an increase in parsing time of a factor of only 1.49 for the core XLE system. The complete version gives an overall improvement in F-score of 5% over the Collins parser at a cost of a factor of 5 in parsing time. ",
"cite_spans": [],
"ref_spans": [
{
"start": 1171,
"end": 1178,
"text": "Table 1",
"ref_id": "TABREF2"
}
],
"eq_spans": [],
"section": "Experiments",
"sec_num": "5"
},
{
"text": "We presented some experiments that compare the accuracy and performance of two stochastic parsing systems, the shallow Collins parser and the deep-grammar-based XLE system. We measured the accuracy of both systems against a gold standard derived from the PARC 700 dependency bank, and also measured their processing times. Contrary to conventional wisdom, we found that the shallow system was not substantially faster than the deep parser operating on a core grammar, while the deep system was significantly more accurate. Furthermore, extending the grammar base of the deep system results in much better accuracy at a cost of a factor of 5 in speed.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion",
"sec_num": "6"
},
{
"text": "Our experiment is comparable to recent work on reading off Propbank-style (Kingsbury and Palmer, 2002) predicate-argument relations from gold-standard treebank trees and automatic parses of the Collins parser. Gildea and Palmer (2002) report F-score results in the 55% range for argument and boundary recognition based on automatic parses. From this perspective, the nearly 75% F-score that is achieved for our deterministic rewriting of Collins' trees into dependencies is remarkable, even if the results are not directly comparable. Our scores and Gildea and Palmer's are both substantially lower than the 90% typically cited for evaluations based on labeled or unlabeled bracketing, suggesting that extracting semantically relevant dependencies is a more difficult, but we think more valuable, task.",
"cite_spans": [
{
"start": 74,
"end": 102,
"text": "(Kingsbury and Palmer, 2002)",
"ref_id": "BIBREF10"
},
{
"start": 210,
"end": 234,
"text": "Gildea and Palmer (2002)",
"ref_id": "BIBREF6"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion",
"sec_num": "6"
},
{
"text": "The resulting test set is thus unseen to the grammar and stochastic disambiguation system used in our experiments. This is indicated by the fact that the upperbound of F-score for the best matching parses for the experiment grammar is in the range of 85%, not 100%.3 http://www2.parc.com/istl/groups/nltt/fsbank/",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "However, we did explore a few of these additional transformations and found only marginal F-score increases.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "The results of this conversion are available at http://www2.parc.com/istl/groups/nltt/fsbank/ (1) Sen. NNP Christopher NNP Dodd NNP \u2212\u2192 Sen. Christopher Dodd NNP",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "A beam size of 10000 as used inCollins (1999) improved the F-score on the heldout set only by .1% at an increase of parsing time by a factor of 3. Beam sizes lower than 1000 decreased the heldout F-score significantly.7 All experiments were run on one CPU of a dual processor AMD Opteron 244 with 1.8 GHz and 4GB main memory. Loading times are included in CPU times.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
}
],
"back_matter": [],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "The parallel grammar project",
"authors": [
{
"first": "Miriam",
"middle": [],
"last": "Butt",
"suffix": ""
},
{
"first": "Helge",
"middle": [],
"last": "Dyvik",
"suffix": ""
},
{
"first": "Tracy",
"middle": [
"Holloway"
],
"last": "King",
"suffix": ""
},
{
"first": "Hiroshi",
"middle": [],
"last": "Masuichi",
"suffix": ""
},
{
"first": "Christian",
"middle": [],
"last": "Rohrer",
"suffix": ""
}
],
"year": 2002,
"venue": "Proceedings of COL-ING2002",
"volume": "",
"issue": "",
"pages": "1--7",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Miriam Butt, Helge Dyvik, Tracy Holloway King, Hi- roshi Masuichi, and Christian Rohrer. 2002. The parallel grammar project. In Proceedings of COL- ING2002, Workshop on Grammar Engineering and Evaluation, pages 1-7.",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "Head-Driven Statistical Models for Natural Language Parsing",
"authors": [
{
"first": "Michael",
"middle": [],
"last": "Collins",
"suffix": ""
}
],
"year": 1999,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Michael Collins. 1999. Head-Driven Statistical Models for Natural Language Parsing. Ph.D. thesis, Univer- sity of Pennsylvania.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "Scalability of redundancy detection in focused document collections",
"authors": [
{
"first": "D",
"middle": [],
"last": "Crouch",
"suffix": ""
},
{
"first": "C",
"middle": [],
"last": "Condoravdi",
"suffix": ""
},
{
"first": "R",
"middle": [],
"last": "Stolle",
"suffix": ""
},
{
"first": "T",
"middle": [
"H"
],
"last": "King",
"suffix": ""
},
{
"first": "V",
"middle": [],
"last": "De Paiva",
"suffix": ""
},
{
"first": "J",
"middle": [],
"last": "Everett",
"suffix": ""
},
{
"first": "D",
"middle": [],
"last": "Bobrow",
"suffix": ""
}
],
"year": 2002,
"venue": "Proceedings of Scalable Natural Language Understanding",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "D. Crouch, C. Condoravdi, R. Stolle, T.H. King, V. de Paiva, J. Everett, and D. Bobrow. 2002. Scal- ability of redundancy detection in focused document collections. In Proceedings of Scalable Natural Lan- guage Understanding, Heidelberg.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "A noisy-channel model for document compression",
"authors": [
{
"first": "Hal",
"middle": [],
"last": "Daume",
"suffix": ""
},
{
"first": "Daniel",
"middle": [],
"last": "Marcu",
"suffix": ""
}
],
"year": 2002,
"venue": "Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics (ACL'02)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Hal Daume and Daniel Marcu. 2002. A noisy-channel model for document compression. In Proceedings of the 40th Annual Meeting of the Association for Com- putational Linguistics (ACL'02), Philadelphia, PA.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "Optimality theory style constraint ranking in large-scale LFG grammars",
"authors": [
{
"first": "Anette",
"middle": [],
"last": "Frank",
"suffix": ""
},
{
"first": "Tracy",
"middle": [
"H"
],
"last": "King",
"suffix": ""
},
{
"first": "Jonas",
"middle": [],
"last": "Kuhn",
"suffix": ""
},
{
"first": "John",
"middle": [],
"last": "Maxwell",
"suffix": ""
}
],
"year": 1998,
"venue": "Proceedings of the Third LFG Conference",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Anette Frank, Tracy H. King, Jonas Kuhn, and John Maxwell. 1998. Optimality theory style constraint ranking in large-scale LFG grammars. In Proceedings of the Third LFG Conference.",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "Dynamic programming for parsing and estimation of stochatic unification-based grammars",
"authors": [
{
"first": "Stuart",
"middle": [],
"last": "Geman",
"suffix": ""
},
{
"first": "Mark",
"middle": [],
"last": "Johnson",
"suffix": ""
}
],
"year": 2002,
"venue": "Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics (ACL'02)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Stuart Geman and Mark Johnson. 2002. Dynamic programming for parsing and estimation of stochatic unification-based grammars. In Proceedings of the 40th Annual Meeting of the Association for Compu- tational Linguistics (ACL'02), Philadelphia, PA.",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "The necessity of parsing for predicate argument recognition",
"authors": [
{
"first": "Dan",
"middle": [],
"last": "Gildea",
"suffix": ""
},
{
"first": "Martha",
"middle": [],
"last": "Palmer",
"suffix": ""
}
],
"year": 2002,
"venue": "Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics (ACL'02)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Dan Gildea and Martha Palmer. 2002. The necessity of parsing for predicate argument recognition. In Pro- ceedings of the 40th Annual Meeting of the Association for Computational Linguistics (ACL'02), Philadelphia.",
"links": null
},
"BIBREF7": {
"ref_id": "b7",
"title": "The role of lexico-semantic feedback in open-domain textual question-answering",
"authors": [
{
"first": "Sanda",
"middle": [],
"last": "Harabagiu",
"suffix": ""
},
{
"first": "Dan",
"middle": [],
"last": "Moldovan",
"suffix": ""
},
{
"first": "Marius",
"middle": [],
"last": "Pa\u015fca",
"suffix": ""
},
{
"first": "Rada",
"middle": [],
"last": "Mihalcea",
"suffix": ""
},
{
"first": "Mihai",
"middle": [],
"last": "Surdeanu",
"suffix": ""
},
{
"first": "R\u0203zvan",
"middle": [],
"last": "Bunescu",
"suffix": ""
},
{
"first": "Roxana",
"middle": [],
"last": "G\u00eerju",
"suffix": ""
},
{
"first": "Vasile",
"middle": [],
"last": "Rus",
"suffix": ""
},
{
"first": "Paul",
"middle": [],
"last": "Mor\u0203rescu",
"suffix": ""
}
],
"year": 2001,
"venue": "Proceedings of the 39th",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Sanda Harabagiu, Dan Moldovan, Marius Pa\u015fca, Rada Mihalcea, Mihai Surdeanu, R\u0203zvan Bunescu, Roxana G\u00eerju, Vasile Rus, and Paul Mor\u0203rescu. 2001. The role of lexico-semantic feedback in open-domain tex- tual question-answering. In Proceedings of the 39th",
"links": null
},
"BIBREF8": {
"ref_id": "b8",
"title": "Annual Meeting and 10th Conference of the European Chapter of the Asssociation for Computational Linguistics (ACL'01)",
"authors": [],
"year": null,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Annual Meeting and 10th Conference of the European Chapter of the Asssociation for Computational Lin- guistics (ACL'01), Toulouse, France.",
"links": null
},
"BIBREF9": {
"ref_id": "b9",
"title": "The PARC 700 dependency bank",
"authors": [
{
"first": "Tracy",
"middle": [
"H"
],
"last": "King",
"suffix": ""
},
{
"first": "Richard",
"middle": [],
"last": "Crouch",
"suffix": ""
},
{
"first": "Stefan",
"middle": [],
"last": "Riezler",
"suffix": ""
},
{
"first": "Mary",
"middle": [],
"last": "Dalrymple",
"suffix": ""
},
{
"first": "Ronald",
"middle": [
"M"
],
"last": "Kaplan",
"suffix": ""
}
],
"year": 2003,
"venue": "Proceedings of the Workshop on \"Linguistically Interpreted Corpora\" at the 10th Conference of the European Chapter of the Association for Computational Linguistics (LINC'03)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Tracy H. King, Richard Crouch, Stefan Riezler, Mary Dalrymple, and Ronald M. Kaplan. 2003. The PARC 700 dependency bank. In Proceedings of the Work- shop on \"Linguistically Interpreted Corpora\" at the 10th Conference of the European Chapter of the Asso- ciation for Computational Linguistics (LINC'03), Bu- dapest, Hungary.",
"links": null
},
"BIBREF10": {
"ref_id": "b10",
"title": "From treebank to propbank",
"authors": [
{
"first": "Paul",
"middle": [],
"last": "Kingsbury",
"suffix": ""
},
{
"first": "Martha",
"middle": [],
"last": "Palmer",
"suffix": ""
}
],
"year": 2002,
"venue": "Proceedings of the 3rd International Conference on Language Resources and Evaluation (LREC'02)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Paul Kingsbury and Martha Palmer. 2002. From tree- bank to propbank. In Proceedings of the 3rd Interna- tional Conference on Language Resources and Evalu- ation (LREC'02), Las Palmas, Spain.",
"links": null
},
"BIBREF11": {
"ref_id": "b11",
"title": "The interface between phrasal and functional constraints",
"authors": [
{
"first": "John",
"middle": [],
"last": "Maxwell",
"suffix": ""
},
{
"first": "Ron",
"middle": [],
"last": "Kaplan",
"suffix": ""
}
],
"year": 1993,
"venue": "Computational Linguistics",
"volume": "19",
"issue": "4",
"pages": "571--589",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "John Maxwell and Ron Kaplan. 1993. The interface be- tween phrasal and functional constraints. Computa- tional Linguistics, 19(4):571-589.",
"links": null
},
"BIBREF12": {
"ref_id": "b12",
"title": "A novel use of statistical parsing to extract information from text",
"authors": [
{
"first": "Scott",
"middle": [],
"last": "Miller",
"suffix": ""
},
{
"first": "Heidi",
"middle": [],
"last": "Fox",
"suffix": ""
},
{
"first": "Lance",
"middle": [],
"last": "Ramshaw",
"suffix": ""
},
{
"first": "Ralph",
"middle": [],
"last": "Weischedel",
"suffix": ""
}
],
"year": 2000,
"venue": "Proceedings of the 1st Conference of the North American Chapter of the Association for Computational Linguistics",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Scott Miller, Heidi Fox, Lance Ramshaw, and Ralph Weischedel. 2000. A novel use of statistical parsing to extract information from text. In Proceedings of the 1st Conference of the North American Chapter of the Association for Computational Linguistics (ANLP- NAACL 2000), Seattle, WA.",
"links": null
},
"BIBREF13": {
"ref_id": "b13",
"title": "Maximum entropy estimation for feature forests",
"authors": [
{
"first": "Yusuke",
"middle": [],
"last": "Miyao",
"suffix": ""
},
{
"first": "Jun'ichi",
"middle": [],
"last": "Tsujii",
"suffix": ""
}
],
"year": 2002,
"venue": "Proceedings of the Human Language Technology Conference (HLT'02)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Yusuke Miyao and Jun'ichi Tsujii. 2002. Maximum entropy estimation for feature forests. In Proceed- ings of the Human Language Technology Conference (HLT'02), San Diego, CA.",
"links": null
},
"BIBREF14": {
"ref_id": "b14",
"title": "A maximum entropy model for part-of-speech tagging",
"authors": [
{
"first": "Adwait",
"middle": [],
"last": "Ratnaparkhi",
"suffix": ""
}
],
"year": 1996,
"venue": "Proceedings of EMNLP-1",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Adwait Ratnaparkhi. 1996. A maximum entropy model for part-of-speech tagging. In Proceedings of EMNLP- 1.",
"links": null
},
"BIBREF15": {
"ref_id": "b15",
"title": "Gradient feature testing and 1 regularization for maximum entropy parsing",
"authors": [
{
"first": "Stefan",
"middle": [],
"last": "Riezler",
"suffix": ""
},
{
"first": "Alexander",
"middle": [],
"last": "Vasserman",
"suffix": ""
}
],
"year": 2004,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Stefan Riezler and Alexander Vasserman. 2004. Gradi- ent feature testing and 1 regularization for maximum entropy parsing. Submitted for publication.",
"links": null
},
"BIBREF16": {
"ref_id": "b16",
"title": "Parsing the Wall Street Journal using a Lexical-Functional Grammar and discriminative estimation techniques",
"authors": [
{
"first": "Stefan",
"middle": [],
"last": "Riezler",
"suffix": ""
},
{
"first": "Tracy",
"middle": [
"H"
],
"last": "King",
"suffix": ""
},
{
"first": "Ronald",
"middle": [
"M"
],
"last": "Kaplan",
"suffix": ""
},
{
"first": "Richard",
"middle": [],
"last": "Crouch",
"suffix": ""
},
{
"first": "John",
"middle": [
"T"
],
"last": "Maxwell",
"suffix": ""
},
{
"first": "Mark",
"middle": [],
"last": "Johnson",
"suffix": ""
}
],
"year": 2002,
"venue": "Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics (ACL'02)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Stefan Riezler, Tracy H. King, Ronald M. Kaplan, Richard Crouch, John T. Maxwell, and Mark John- son. 2002. Parsing the Wall Street Journal using a Lexical-Functional Grammar and discriminative esti- mation techniques. In Proceedings of the 40th Annual Meeting of the Association for Computational Linguis- tics (ACL'02), Philadelphia, PA.",
"links": null
},
"BIBREF17": {
"ref_id": "b17",
"title": "A syntax-based statistical translation model",
"authors": [
{
"first": "Kenji",
"middle": [],
"last": "Yamada",
"suffix": ""
},
{
"first": "Kevin",
"middle": [],
"last": "Knight",
"suffix": ""
}
],
"year": 2001,
"venue": "Proceedings of the 39th Annual Meeting and 10th Conference of the European Chapter of the Asssociation for Computational Linguistics (ACL'01)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Kenji Yamada and Kevin Knight. 2001. A syntax-based statistical translation model. In Proceedings of the 39th Annual Meeting and 10th Conference of the Eu- ropean Chapter of the Asssociation for Computational Linguistics (ACL'01), Toulouse, France.",
"links": null
}
},
"ref_entries": {
"FIGREF0": {
"uris": null,
"type_str": "figure",
"num": null,
"text": "Grammatical functions in DEPBANK."
},
"FIGREF1": {
"uris": null,
"type_str": "figure",
"num": null,
"text": "Selected features for Reduced DEPBANK ."
},
"FIGREF2": {
"uris": null,
"type_str": "figure",
"num": null,
"text": "Reduced dependency relations for He reiterated his opposition to such funding, but expressed hope of a compromise."
},
"FIGREF3": {
"uris": null,
"type_str": "figure",
"num": null,
"text": "Collins Model 3 tree for He reiterated his opposition to such funding, but expressed hope of a compromise."
},
"TABREF2": {
"content": "<table><tr><td/><td>time</td><td colspan=\"2\">prec. rec. F-score</td></tr><tr><td>LFG core</td><td colspan=\"2\">298.88 79.1 76.2</td><td>77.6</td></tr><tr><td colspan=\"2\">LFG complete 985.3</td><td>79.4 79.8</td><td>79.6</td></tr><tr><td>Collins 1000</td><td>199.6</td><td>78.3 71.2</td><td>74.6</td></tr></table>",
"text": "Timing and accuracy results for Collins parser and Complete and Core versions of XLE system on Reduced version of PARC 700 dependency bank.",
"type_str": "table",
"num": null,
"html": null
}
}
}
}