| { |
| "paper_id": "P97-1003", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T09:15:33.064489Z" |
| }, |
| "title": "Three Generative, Lexicalised Models for Statistical Parsing", |
| "authors": [ |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Collins", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "University of Pennsylvania Philadelphia", |
| "location": { |
| "postCode": "19104", |
| "region": "PA", |
| "country": "U.S.A" |
| } |
| }, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "In this paper we first propose a new statistical parsing model, which is a generative model of lexicalised context-free grammar. We then extend the model to include a probabilistic treatment of both subcategorisation and wh-movement. Results on Wall Street Journal text show that the parser performs at 88.1/87.5% constituent precision/recall, an average improvement of 2.3% over (Collins 96).", |
| "pdf_parse": { |
| "paper_id": "P97-1003", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "In this paper we first propose a new statistical parsing model, which is a generative model of lexicalised context-free grammar. We then extend the model to include a probabilistic treatment of both subcategorisation and wh-movement. Results on Wall Street Journal text show that the parser performs at 88.1/87.5% constituent precision/recall, an average improvement of 2.3% over (Collins 96).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Generative models of syntax have been central in linguistics since they were introduced in (Chomsky 57). Each sentence-tree pair (S,T) in a language has an associated top-down derivation consisting of a sequence of rule applications of a grammar. These models can be extended to be statistical by defining probability distributions at points of non-determinism in the derivations, thereby assigning a probability 7)(S, T) to each (S, T) pair. Probabilistic context free grammar (Booth and Thompson 73) was an early example of a statistical grammar. A PCFG can be lexicalised by associating a headword with each non-terminal in a parse tree; thus far, (Magerman 95; Jelinek et al. 94) and (Collins 96) , which both make heavy use of lexical information, have reported the best statistical parsing performance on Wall Street Journal text. Neither of these models is generative, instead they both estimate 7)(T] S) directly.", |
| "cite_spans": [ |
| { |
| "start": 478, |
| "end": 501, |
| "text": "(Booth and Thompson 73)", |
| "ref_id": null |
| }, |
| { |
| "start": 665, |
| "end": 683, |
| "text": "Jelinek et al. 94)", |
| "ref_id": null |
| }, |
| { |
| "start": 688, |
| "end": 700, |
| "text": "(Collins 96)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "This paper proposes three new parsing models. Model 1 is essentially a generative version of the model described in (Collins 96) . In Model 2, we extend the parser to make the complement/adjunct distinction by adding probabilities over subcategorisation frames for head-words. In Model 3 we give a probabilistic treatment of wh-movement, which This research was supported by ARPA Grant N6600194-C6043.", |
| "cite_spans": [ |
| { |
| "start": 116, |
| "end": 128, |
| "text": "(Collins 96)", |
| "ref_id": null |
| }, |
| { |
| "start": 325, |
| "end": 343, |
| "text": "wh-movement, which", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "is derived from the analysis given in Generalized Phrase Structure Grammar (Gazdar et al. 95) . The work makes two advances over previous models: First, Model 1 performs significantly better than (Collins 96) , and Models 2 and 3 give further improvements --our final results are 88.1/87.5% constituent precision/recall, an average improvement of 2.3% over (Collins 96) . Second, the parsers in (Collins 96) and (Magerman 95; Jelinek et al. 94) produce trees without information about whmovement or subcategorisation. Most NLP applications will need this information to extract predicateargument structure from parse trees.", |
| "cite_spans": [ |
| { |
| "start": 75, |
| "end": 93, |
| "text": "(Gazdar et al. 95)", |
| "ref_id": null |
| }, |
| { |
| "start": 196, |
| "end": 208, |
| "text": "(Collins 96)", |
| "ref_id": null |
| }, |
| { |
| "start": 357, |
| "end": 369, |
| "text": "(Collins 96)", |
| "ref_id": null |
| }, |
| { |
| "start": 395, |
| "end": 407, |
| "text": "(Collins 96)", |
| "ref_id": null |
| }, |
| { |
| "start": 412, |
| "end": 425, |
| "text": "(Magerman 95;", |
| "ref_id": null |
| }, |
| { |
| "start": 426, |
| "end": 444, |
| "text": "Jelinek et al. 94)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In the remainder of this paper we describe the 3 models in section 2, discuss practical issues in section 3, give results in section 4, and give conclusions in section 5.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The Three Parsing Models", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2", |
| "sec_num": null |
| }, |
| { |
| "text": "In general, a statistical parsing model defines the conditional probability, 7)(T] S), for each candidate parse tree T for a sentence S. The parser itself is an algorithm which searches for the tree, Tb~st, that maximises 7~(T I S). A generative model uses the observation that maximising 7V(T, S) is equivalent to maximising 7~(T ] S): 1", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model 1", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "Tbe,t = argm~xT~(TlS) = argmTax ?~(T,S) ~(s) = arg m~x 7~(T, S)", |
| "eq_num": "(1)" |
| } |
| ], |
| "section": "Model 1", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "7~(T, S) is then estimated by attaching probabilities to a top-down derivation of the tree. In a PCFG, for a tree derived by n applications of context-free re-write rules LHSi ~ RHSi, 1 < i < n,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model 1", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "7~(T,S) = H 7)(RHSi I LHSi) (2) i=l..n", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model 1", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "The re-write rules are either internal to the tree, where LHS is a non-terminal and RHS is a string", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model 1", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "7~(T,S)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model 1", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "17~(S) is constant, hence maximising ~ is equivalent to maximising \"P(T, S).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model 1", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "i S(bought) NP(w~ought ) t VB/~Np m JJ NN NNP I I I ooks) Last week Marks I 1 bought NNP f Brooks TOP -> S(bought) S(bought) -> NP(week) NP(week) -> JJ(Last) NP (Marks) -> NNP (Marks) VP (bought) -> VB (bought) NP (Brooks) -> NNP (Brooks) NP(Marks) VP(bought) NN(week)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "TOP", |
| "sec_num": null |
| }, |
| { |
| "text": "Figure 1: A lexicalised parse tree, and a list of the rules it contains. For brevity we omit the POS tag associated with each word.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "NP(Brooks)", |
| "sec_num": null |
| }, |
| { |
| "text": "of one or more non-terminals; or lexical, where LHS is a part of speech tag and RHS is a word.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "NP(Brooks)", |
| "sec_num": null |
| }, |
| { |
| "text": "A PCFG can be lexicalised 2 by associating a word w and a part-of-speech (POS) tag t with each nonterminal X in the tree. Thus we write a nonterminal as X(x), where x = (w,t), and X is a constituent label. Each rule now has the form3:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "NP(Brooks)", |
| "sec_num": null |
| }, |
| { |
| "text": "P(h) -> Ln(In)...ni(ll)H(h)Rl(rl)...Rm(rm) (3)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "NP(Brooks)", |
| "sec_num": null |
| }, |
| { |
| "text": "H is the head-child of the phrase, which inherits the head-word h from its parent P. L1...L~ and R1...Rm are left and right modifiers of H. Either n or m may be zero, and n = m = 0 for unary rules. Figure 1 shows a tree which will be used as an example throughout this paper.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 198, |
| "end": 206, |
| "text": "Figure 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "NP(Brooks)", |
| "sec_num": null |
| }, |
| { |
| "text": "The addition of lexical heads leads to an enormous number of potential rules, making direct estimation of ?)(RHS { LHS) infeasible because of sparse data problems. We decompose the generation of the RHS of a rule such as (3), given the LHS, into three steps --first generating the head, then making the independence assumptions that the left and right modifiers are generated by separate 0th-order markov processes 4:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "NP(Brooks)", |
| "sec_num": null |
| }, |
| { |
| "text": "1. Generate the head constituent label of the phrase, with probability 7)H(H I P, h).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "NP(Brooks)", |
| "sec_num": null |
| }, |
| { |
| "text": "2. Generate modifiers to the right of the head with probability 1-Ii=1..m+1 ~n(Ri(ri) { P, h, H). R,~+l(r,~+l) is defined as STOP -the STOP symbol is added to the vocabulary of nonterminals, and the model stops generating right modifiers when it is generated. 2We find lexical heads in Penn treebank data using rules which are similar to those used by (Magerman 95; Jelinek et al. 94) .", |
| "cite_spans": [ |
| { |
| "start": 366, |
| "end": 384, |
| "text": "Jelinek et al. 94)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "NP(Brooks)", |
| "sec_num": null |
| }, |
| { |
| "text": "SWith the exception of the top rule in the tree, which has the form TOP --+ H(h).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "NP(Brooks)", |
| "sec_num": null |
| }, |
| { |
| "text": "4An exception is the first rule in the tree, T0P -+ H (h), which has probability Prop (H, hlTOP ) We have made the 0 th order markov assumptions", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "NP(Brooks)", |
| "sec_num": null |
| }, |
| { |
| "text": "7~,(Li(li) { H, P, h, L1 (ll)...Li-1 (/i-1)) = P~(Li(li) { H,P,h) (4) Pr (Ri (ri) { H, P, h, R1 (rl)...R~-1 (ri-1 )) = ?~r(Ri(ri) { H, P, h) (5)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "NP(Brooks)", |
| "sec_num": null |
| }, |
| { |
| "text": "but in general the probabilities could be conditioned on any of the preceding modifiers. In fact, if the derivation order is fixed to be depth-first --that is, each modifier recursively generates the sub-tree below it before the next modifier is generated -then the model can also condition on any structure below the preceding modifiers. For the moment we exploit this by making the approximations", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "NP(Brooks)", |
| "sec_num": null |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "7~l( Li(li ) { H, P, h, Ll ( ll )...Li_l (l~_l ) ) = ?)l(ni(li) l H, P,h, distancez(i -1)) (6) ?)r( ai(ri) ] H, P, h, R1 (rl)...Ri-1 (ri-l ) ) = ?~T(Ri(ri) [ H,P.h, distancer(i -1))", |
| "eq_num": "(7)" |
| } |
| ], |
| "section": "NP(Brooks)", |
| "sec_num": null |
| }, |
| { |
| "text": "where distancez and distancer are functions of the surface string from the head word to the edge of the constituent (see figure 2) . The distance measure is the same as in (Collins 96) , a vector with the following 3 elements: (1) is the string of zero length? (Allowing the model to learn a preference for rightbranching structures); (2) does the string contain a verb? (Allowing the model to learn a preference for modification of the most recent verb). (3) Does the string contain 0, 1, 2 or > 2 commas? (where a comma is anything tagged as \",\" or \":\").", |
| "cite_spans": [ |
| { |
| "start": 172, |
| "end": 184, |
| "text": "(Collins 96)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 121, |
| "end": 130, |
| "text": "figure 2)", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "NP(Brooks)", |
| "sec_num": null |
| }, |
| { |
| "text": "P(h) distance -I", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "NP(Brooks)", |
| "sec_num": null |
| }, |
| { |
| "text": "Figure 2: The next child, Ra(r3), is generated with probability 7~(R3(r3) [ P,H, h, distancer(2) ). The distance is a function of the surface string from the word after h to the last word of R2, inclusive. In principle the model could condition on any structure dominated by H, R1 or R2.", |
| "cite_spans": [ |
| { |
| "start": 74, |
| "end": 96, |
| "text": "[ P,H, h, distancer(2)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "NP(Brooks)", |
| "sec_num": null |
| }, |
| { |
| "text": "The tree in figure 1 is an example of the importance of the complement/adjunct distinction. It would be useful to identify \"Marks\" as a subject, and \"Last week\" as an adjunct (temporal modifier), but this distinction is not made in the tree, as both NPs are in the same position 5 (sisters to a VP under an S node). From here on we will identify complements by attaching a \"-C\" suffix to non-terminals --figure 3 gives an example tree.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model 2: The complement/adjunct distinction and subcategorisation", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "TOP 1 S(bought) NP(w~ought) Last week Marks VBD NP-C(Brooks) I l", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model 2: The complement/adjunct distinction and subcategorisation", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "bought Brooks Figure 3 : A tree with the \"-C\" suffix used to identify complements. \"Marks\" and \"Brooks\" are in subject and object position respectively. \"Last week\" is an adjunct.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 14, |
| "end": 22, |
| "text": "Figure 3", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Model 2: The complement/adjunct distinction and subcategorisation", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "A post-processing stage could add this detail to the parser output, but we give two reasons for making the distinction while parsing: First, identifying complements is complex enough to warrant a probabilistic treatment. Lexical information is needed 5Except \"Marks\" is closer to the VP, but note that \"Marks\" is also the subject in \"Marks last week bought Brooks\".", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model 2: The complement/adjunct distinction and subcategorisation", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "--for example, knowledge that \"week '' is likely to be a temporal modifier. Knowledge about subcategorisation preferences --for example that a verb takes exactly one subject --is also required. These problems are not restricted to NPs, compare \"The spokeswoman said (SBAR that the asbestos was dangerous)\" vs. \"Bonds beat short-term investments (SBAR because the market is down)\", where an SBAR headed by \"that\" is a complement, but an SBAI:t headed by \"because\" is an adjunct.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model 2: The complement/adjunct distinction and subcategorisation", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "The second reason for making the complement/adjunct distinction while parsing is that it may help parsing accuracy. The assumption that complements are generated independently of each other often leads to incorrect parses --see figure 4 for further explanation.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model 2: The complement/adjunct distinction and subcategorisation", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Adjuncts in the Penn Treebank", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Identifying Complements and", |
| "sec_num": "2.2.1" |
| }, |
| { |
| "text": "We add the \"-C\" suffix to all non-terminals in training data which satisfy the following conditions:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Identifying Complements and", |
| "sec_num": "2.2.1" |
| }, |
| { |
| "text": "1. The non-terminal must be: (1) an NP, SBAR, or S whose parent is an S; (2) an NP, SBAR, S, or VP whose parent is a VP; or (3) an S whose parent is an SBAR.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Identifying Complements and", |
| "sec_num": "2.2.1" |
| }, |
| { |
| "text": "2. The non-terminal must not have one of the following semantic tags: ADV, VOC, BNF, DIR, EXT, LOC, MNR, TMP, CLR or PRP. See (Marcus et al. 94) for an explanation of what these tags signify. For example, the NP \"Last week\" in figure 1 would have the TMP (temporal) tag; and the SBAR in \"(SBAR because the market is down)\", would have the ADV (adverbial) tag.", |
| "cite_spans": [ |
| { |
| "start": 126, |
| "end": 144, |
| "text": "(Marcus et al. 94)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Identifying Complements and", |
| "sec_num": "2.2.1" |
| }, |
| { |
| "text": "In addition, the first child following the head of a prepositional phrase is marked as a complement.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Identifying Complements and", |
| "sec_num": "2.2.1" |
| }, |
| { |
| "text": "The model could be retrained on training data with the enhanced set of non-terminals, and it might learn the lexical properties which distinguish complements and adjuncts (\"Marks\" vs \"week\", or \"that\" vs. \"because\"). However, it would still suffer from the bad independence assumptions illustrated in figure 4. To solve these kinds of problems, the generative process is extended to include a probabilistic choice of left and right subcategorisation frames:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Probabilities over Subcategorisation Frames", |
| "sec_num": "2.2.2" |
| }, |
| { |
| "text": "1. Choose a head H with probability ~H(H [P, h) . Figure 4 : Two examples where the assumption that modifiers are generated independently of each other leads to errors.", |
| "cite_spans": [ |
| { |
| "start": 41, |
| "end": 47, |
| "text": "[P, h)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 50, |
| "end": 58, |
| "text": "Figure 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Probabilities over Subcategorisation Frames", |
| "sec_num": "2.2.2" |
| }, |
| { |
| "text": "In (1) the probability of generating both \"Dreyfus\" and \"fund\" as subjects, 7~(NP-C(Dreyfus) I S,VP,was) * T'(NP-C(fund) I S,VP,was) is unreasonably high. (2) is similar: 7 ~ (NP-C (bill), VP-C (funding) I VP, VB, was) = P(NP-C (bill) I VP, VB, was) * 7~(VP-C (funding) I VP, VB, was) is a bad independence assumption.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Probabilities over Subcategorisation Frames", |
| "sec_num": "2.2.2" |
| }, |
| { |
| "text": "Prc (RCIP, H,h ) . Each subcat frame is a multiset 6 specifying the complements which the head requires in its left or right modifiers.", |
| "cite_spans": [ |
| { |
| "start": 4, |
| "end": 16, |
| "text": "(RCIP, H,h )", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Probabilities over Subcategorisation Frames", |
| "sec_num": "2.2.2" |
| }, |
| { |
| "text": "- abilities 7)l(Li, li I H, P, h, distancet(i -1), LC) and 7~r (R~, ri I H, P, h, distancer(i -1), RC) re-", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Generate the left and right modifiers with prob", |
| "sec_num": "3." |
| }, |
| { |
| "text": "spectively. Thus the subcat requirements are added to the conditioning context. As complements are generated they are removed from the appropriate subcat multiset. Most importantly, the probability of generating the STOP symbol will be 0 when the subcat frame is non-empty, and the probability of generating a complement will be 0 when it is not in the subcat frame; thus all and only the required complements will be generated. Here the head initially decides to take a single NP-C (subject) to its left, and no complements ~A rnultiset, or bag, is a set which may contain duplicate non-terminal labels. Example 3 The store (SBAR which Marks bought Brooks Brothers/tom TRACE)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Generate the left and right modifiers with prob", |
| "sec_num": "3." |
| }, |
| { |
| "text": "It might be possible to write rule-based patterns which identify traces in a parse tree. However, we argue again that this task is best integrated into the parser: the task is complex enough to warrant a probabilistic treatment, and integration may help parsing accuracy. A couple of complexities are that modification by an SBAR does not always involve extraction (e.g., \"the fact (SBAR that besoboru is Figure 5 : A +gap feature can be added to non-terminals to describe NP extraction. The top-level NP initially generates an SBAR modifier, but specifies that it must contain an NP trace by adding the +gap feature. The gap is then passed down through the tree, until it is discharged as a TRACE complement to the right of bought.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 405, |
| "end": 413, |
| "text": "Figure 5", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Generate the left and right modifiers with prob", |
| "sec_num": "3." |
| }, |
| { |
| "text": "played with a ball and a bat)\"), and it is not uncommon for extraction to occur through several constituents, (e.g., \"The changes (SBAR that he said the government was prepared to make TRACE)\").", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Generate the left and right modifiers with prob", |
| "sec_num": "3." |
| }, |
| { |
| "text": "The second reason for an integrated treatment of traces is to improve the parameterisation of the model. In particular, the subcategorisation probabilities are smeared by extraction. In examples 1, 2 and 3 above 'bought' is a transitive verb, but without knowledge of traces example 2 in training data will contribute to the probability of 'bought' being an intransitive verb.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Generate the left and right modifiers with prob", |
| "sec_num": "3." |
| }, |
| { |
| "text": "Formalisms similar to GPSG (Gazdar et al. 95) handle NP extraction by adding a gap feature to each non-terminal in the tree, and propagating gaps through the tree until they are finally discharged as a trace complement (see figure 5) . In extraction cases the Penn treebank annotation co-indexes a TRACE with the WHNP head of the SBAR, so it is straightforward to add this information to trees in training data.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 224, |
| "end": 233, |
| "text": "figure 5)", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Generate the left and right modifiers with prob", |
| "sec_num": "3." |
| }, |
| { |
| "text": "Given that the LHS of the rule has a gap, there are 3 ways that the gap can be passed down to the RHS:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Generate the left and right modifiers with prob", |
| "sec_num": "3." |
| }, |
| { |
| "text": "Head The gap is passed to the head of the phrase, as in rule (3) in figure 5.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Generate the left and right modifiers with prob", |
| "sec_num": "3." |
| }, |
| { |
| "text": "Left, Right The gap is passed on recursively to one of the left or right modifiers of the head, or is discharged as a trace argument to the left/right of the head. In rule (2) it is passed on to a right modifier, the S complement. In rule (4) a trace is generated to the right of the head VB.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Generate the left and right modifiers with prob", |
| "sec_num": "3." |
| }, |
| { |
| "text": "We specify a parameter 7~c (GIP, h, H) where G is either Head, Left or Right. The generative process is extended to choose between these cases after generating the head of the phrase. The rest of the phrase is then generated in different ways depending on how the gap is propagated: In the Head case the left and right modifiers are generated as normal. In the Left, Right cases a gap requirement is added to either the left or right SUBCAT variable. This requirement is fulfilled (and removed from the subcat list) when a trace or a modifier non-terminal which has the +gap feature is gener- Li, Iti, P, H, w, t, A, LC L,, lti, P, H, t, A, LC LI, lti It~ (H I P, w, t) , e2 = 7~H(H I P, t), and e3 = PH(H I P). A is the distance measure.", |
| "cite_spans": [ |
| { |
| "start": 27, |
| "end": 38, |
| "text": "(GIP, h, H)", |
| "ref_id": null |
| }, |
| { |
| "start": 656, |
| "end": 669, |
| "text": "(H I P, w, t)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Generate the left and right modifiers with prob", |
| "sec_num": "3." |
| }, |
| { |
| "text": ":ulfills both the S-C and +gap requirements in RC.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Generate the left and right modifiers with prob", |
| "sec_num": "3." |
| }, |
| { |
| "text": "In rule (4) Right is chosen again. Note that generation of trace satisfies both the NP-C and +gap subcat requirements. Table 1 shows the various levels of back-off for each type of parameter in the model. Note that we decompose \"PL(Li(lwi,lti) I P, H,w,t,~,LC) (where lwi and Iti are the word and POS tag generated with non-terminal Li, A is the distance measure)", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 119, |
| "end": 126, |
| "text": "Table 1", |
| "ref_id": "TABREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Generate the left and right modifiers with prob", |
| "sec_num": "3." |
| }, |
| { |
| "text": "into the product 79L1(Li(lti) I P, H,w,t, Zx,LC) x 7~ L2 (lwi ILi, lti, 19, H, w, t, A, LC) , and then smooth these two probabilities separately (Jason Eisner, p.c.). In each case 7 the final estimate is e----Ale1", |
| "cite_spans": [ |
| { |
| "start": 30, |
| "end": 48, |
| "text": "I P, H,w,t, Zx,LC)", |
| "ref_id": null |
| }, |
| { |
| "start": 57, |
| "end": 66, |
| "text": "(lwi ILi,", |
| "ref_id": null |
| }, |
| { |
| "start": 67, |
| "end": 71, |
| "text": "lti,", |
| "ref_id": null |
| }, |
| { |
| "start": 72, |
| "end": 75, |
| "text": "19,", |
| "ref_id": null |
| }, |
| { |
| "start": 76, |
| "end": 78, |
| "text": "H,", |
| "ref_id": null |
| }, |
| { |
| "start": 79, |
| "end": 81, |
| "text": "w,", |
| "ref_id": null |
| }, |
| { |
| "start": 82, |
| "end": 84, |
| "text": "t,", |
| "ref_id": null |
| }, |
| { |
| "start": 85, |
| "end": 87, |
| "text": "A,", |
| "ref_id": null |
| }, |
| { |
| "start": 88, |
| "end": 91, |
| "text": "LC)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Smoothing and Unknown Words", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "+ (1 -&l)(A2e2 + (1 -&2)ea)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Smoothing and Unknown Words", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "where ex, e2 and e3 are maximum likelihood estimates with the context at levels 1, 2 and 3 in the table, and ,kl, ,k2 and )~3 are smoothing parameters where 0 _< ,ki _< 1. All words occurring less than 5 times in training data, and words in test data which rExcept cases L2 and R2, which have 4 levels, so that e = ~let + (1 --*X1)()~2e2 + (1 -,~2)(&3e3 + (1 -~3)e4)).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Smoothing and Unknown Words", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "have never been seen in training, are replaced with the \"UNKNOWN\" token. This allows the model to robustly handle the statistics for rare or new words.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Smoothing and Unknown Words", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "Part of speech tags are generated along with the words in this model. When parsing, the POS tags allowed for each word are limited to those which have been seen in training data for that word. For unknown words, the output from the tagger described in (Ratnaparkhi 96 ) is used as the single possible tag for that word. A CKY style dynamic programming chart parser is used to find the maximum probability tree for each sentence (see figure 6 ).", |
| "cite_spans": [ |
| { |
| "start": 252, |
| "end": 267, |
| "text": "(Ratnaparkhi 96", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 433, |
| "end": 441, |
| "text": "figure 6", |
| "ref_id": "FIGREF5" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Part of Speech Tagging and Parsing", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "The parser was trained on sections 02 -21 of the Wall Street Journal portion of the Penn Treebank (Marcus et al. 93) (approximately 40,000 sentences), and tested on section 23 (2,416 sentences). We use the PAR.SEVAL measures (Black et al. 91) Crossing Brackets ----number of constituents which violate constituent boundaries with a constituent in the treebank parse.", |
| "cite_spans": [ |
| { |
| "start": 225, |
| "end": 242, |
| "text": "(Black et al. 91)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "4" |
| }, |
| { |
| "text": "For a constituent to be 'correct' it must span the same set of words (ignoring punctuation, i.e. all tokens tagged as commas, colons or quotes) and have the same label s as a constituent in the treebank parse. Table 2 shows the results for Models 1, 2 and 3. The precision/recall of the traces found by Model 3 was 93.3%/90.1% (out of 436 cases in section 23 of the treebank), where three criteria must be met for a trace to be \"correct\": (1) it must be an argument to the correct head-word; (2) it must be in the correct position in relation to that head word (preceding or following); (3) it must be dominated by the correct non-terminal label. For example, in figure 5 the trace is an argument to bought, which it follows, and it is dominated by a VP. Of the 436 cases, 342 were string-vacuous extraction from subject position, recovered with 97.1%/98.2% precision/recall; and 94 were longer distance cases, recovered with 76%/60.6% precision/recall 9", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 210, |
| "end": 217, |
| "text": "Table 2", |
| "ref_id": "TABREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Model 1 is similar in structure to (Collins 96) -the major differences being that the \"score\" for each bigram dependency is 7't(L{,liIH, P, h, distancet) 8(Magerman 95) collapses ADVP and PRT to the same label, for comparison we also removed this distinction when calculating scores.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Comparison to previous work", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "9We exclude infinitival relative clauses from these figures, for example \"I called a plumber TRACE to fix the sink\" where 'plumber' is co-indexed with the trace subject of the infinitival. The algorithm scored 41%/18% precision/recall on the 60 cases in section 23 --but infinitival relatives are extremely difficult even for human annotators to distinguish from purpose clauses (in this case, the infinitival could be a purpose clause modifying 'called') (Ann Taylor, p.c.) rather than Pz (Li, P, H I li, h, distancel) , and that there are the additional probabilities of generating the head and the STOP symbols for each constituent. However, Model 1 has some advantages which may account for the improved performance. The model in (Collins 96) is deficient, that is for most sentences S, Y~T 7~( T ] S) < 1, because probability mass is lost to dependency structures which violate the hard constraint that no links may cross. For reasons we do not have space to describe here, Model 1 has advantages in its treatment of unary rules and the distance measure. The generative model can condition on any structure that has been previously generated --we exploit this in models 2 and 3 --whereas (Collins 96 sparse data problems, or problems with coverage (a rule which has never been seen in training may be required for a test data sentence). The complement/adjunct distinction and traces increase the number of rules, compounding this problem. (Eisner 96) proposes 3 dependency models, and gives results that show that a generative model similar to Model 1 performs best of the three. However, a pure dependency model omits non-terminal information, which is important. For example, \"hope\" is likely to generate a VP(T0) modifier (e.g., I hope [VP to sleep]) whereas \"'require\" is likely to generate an S(T0) modifier (e.g., I require IS Jim to sleep]), but omitting non-terminals conflates these two cases, giving high probability to incorrect structures such as \"I hope [Jim to sleep]\" or \"I require [to sleep]\". (Alshawi 96) extends a generative dependency model to include an additional state variable which is equivalent to having non-terminals --his suggestions may be close to our models 1 and 2, but he does not fully specify the details of his model, and doesn't give results for parsing accuracy. (Miller et al. 96) describe a model where the RHS of a rule is generated by a Markov process, although the process is not head-centered. They increase the set of non-terminals by adding semantic labels rather than by adding lexical head-words.", |
| "cite_spans": [ |
| { |
| "start": 490, |
| "end": 519, |
| "text": "(Li, P, H I li, h, distancel)", |
| "ref_id": null |
| }, |
| { |
| "start": 734, |
| "end": 746, |
| "text": "(Collins 96)", |
| "ref_id": null |
| }, |
| { |
| "start": 1193, |
| "end": 1204, |
| "text": "(Collins 96", |
| "ref_id": null |
| }, |
| { |
| "start": 1444, |
| "end": 1455, |
| "text": "(Eisner 96)", |
| "ref_id": null |
| }, |
| { |
| "start": 2307, |
| "end": 2325, |
| "text": "(Miller et al. 96)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Comparison to previous work", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "(Magerman 95; Jelinek et al. 94) describe a history-based approach which uses decision trees to estimate 7a(T[S). Our models use much less sophisticated n-gram estimation methods, and might well benefit from methods such as decision-tree estimation which could condition on richer history than just surface distance.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Comparison to previous work", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "There has recently been interest in using dependency-based parsing models in speech recognition, for example (Stolcke 96) . It is interesting to note that Models 1, 2 or 3 could be used as language models. The probability for any sentence can be estimated as P(S) = ~~.TP(T,S), or (making a Viterbi approximation for efficiency reasons) as 7)(S) .~ P (Tb~st, S) . We intend to perform experiments to compare the perplexity of the various models, and a structurally similar 'pure' PCFG 1\u00b0.", |
| "cite_spans": [ |
| { |
| "start": 109, |
| "end": 121, |
| "text": "(Stolcke 96)", |
| "ref_id": null |
| }, |
| { |
| "start": 351, |
| "end": 361, |
| "text": "(Tb~st, S)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Comparison to previous work", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "This paper has proposed a generative, lexicalised, probabilistic parsing model. We have shown that linguistically fundamental ideas, namely subcategorisation and wh-movement, can be given a statistical interpretation. This improves parsing performance, and, more importantly, adds useful information to the parser's output.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "I would like to thank Mitch Marcus, Jason Eisner, Dan Melamed and Adwait Ratnaparkhi for many useful discussions, and comments on earlier versions of this paper. This work has also benefited greatly from suggestions and advice from Scott Miller.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgements", |
| "sec_num": "6" |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Head Automata and Bilingual Tiling: Translation with Minimal Representations", |
| "authors": [ |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Alshawi", |
| "suffix": "" |
| } |
| ], |
| "year": 1996, |
| "venue": "Proceedings of the 3~th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "167--176", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "H. Alshawi. 1996. Head Automata and Bilingual Tiling: Translation with Minimal Representa- tions. Proceedings of the 3~th Annual Meeting of the Association for Computational Linguistics, pages 167-176.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "A Procedure for Quantitatively Comparing the Syntactic Coverage of English Grammars", |
| "authors": [ |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Black", |
| "suffix": "" |
| } |
| ], |
| "year": 1991, |
| "venue": "Proceedings of the February 1991 DARPA Speech and Natural Language Workshop. 1\u00b0Thanks to one of the anonymous reviewers for suggesting these experiments", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "E. Black et al. 1991. A Procedure for Quantita- tively Comparing the Syntactic Coverage of En- glish Grammars. Proceedings of the February 1991 DARPA Speech and Natural Language Workshop. 1\u00b0Thanks to one of the anonymous reviewers for sug- gesting these experiments.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Applying Probability Measures to Abstract Languages", |
| "authors": [ |
| { |
| "first": "T", |
| "middle": [ |
| "L" |
| ], |
| "last": "Booth", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [ |
| "A" |
| ], |
| "last": "Thompson", |
| "suffix": "" |
| } |
| ], |
| "year": 1973, |
| "venue": "IEEE Transactions on Computers, C", |
| "volume": "22", |
| "issue": "5", |
| "pages": "442--450", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "T. L. Booth and R. A. Thompson. 1973. Applying Probability Measures to Abstract Languages. IEEE Transactions on Computers, C-22(5), pages 442- 450.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Parsing with Context-Free Grammars and Word Statistics", |
| "authors": [ |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Charniak", |
| "suffix": "" |
| } |
| ], |
| "year": 1995, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "E. Charniak. 1995. Parsing with Context-Free Gram- mars and Word Statistics. Technical Report CS- 95-28, Dept. of Computer Science, Brown Univer- sity.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Syntactic Structures", |
| "authors": [ |
| { |
| "first": "N", |
| "middle": [], |
| "last": "Chomsky", |
| "suffix": "" |
| } |
| ], |
| "year": 1957, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "N. Chomsky. 1957. Syntactic Structures, Mouton, The Hague.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "A New Statistical Parser Based on Bigram Lexical Dependencies", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [ |
| "J" |
| ], |
| "last": "Collins", |
| "suffix": "" |
| } |
| ], |
| "year": 1996, |
| "venue": "Proceedings o/ the 34th Annual Meeting o/ the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "184--191", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "M. J. Collins. 1996. A New Statistical Parser Based on Bigram Lexical Dependencies. Proceedings o/ the 34th Annual Meeting o/ the Association for Computational Linguistics, pages 184-191.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Three New Probabilistic Models for Dependency Parsing: An Exploration", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Eisner", |
| "suffix": "" |
| } |
| ], |
| "year": 1996, |
| "venue": "Proceedings o/ COLING-96", |
| "volume": "", |
| "issue": "", |
| "pages": "340--345", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "J. Eisner. 1996. Three New Probabilistic Models for Dependency Parsing: An Exploration. Proceed- ings o/ COLING-96, pages 340-345.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Generalized Phrase Structure Grammar", |
| "authors": [ |
| { |
| "first": "G", |
| "middle": [], |
| "last": "Gazdar", |
| "suffix": "" |
| }, |
| { |
| "first": "E", |
| "middle": [ |
| "H" |
| ], |
| "last": "Klein", |
| "suffix": "" |
| }, |
| { |
| "first": "G", |
| "middle": [ |
| "K" |
| ], |
| "last": "Pullum", |
| "suffix": "" |
| }, |
| { |
| "first": "I", |
| "middle": [ |
| "A" |
| ], |
| "last": "Sag", |
| "suffix": "" |
| } |
| ], |
| "year": 1985, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "G. Gazdar, E.H. Klein, G.K. Pullum, I.A. Sag. 1985. Generalized Phrase Structure Grammar. Harvard University Press.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Decision Tree Parsing using a Hidden Derivation Model", |
| "authors": [ |
| { |
| "first": "F", |
| "middle": [], |
| "last": "Jelinek", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Lafferty", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Magerman", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Mercer", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Ratnaparkhi", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Roukos", |
| "suffix": "" |
| } |
| ], |
| "year": 1994, |
| "venue": "Proceedings o/ the 1994 Human Language Technology Workshop", |
| "volume": "", |
| "issue": "", |
| "pages": "272--277", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "F. Jelinek, J. Lafferty, D. Magerman, R. Mercer, A. Ratnaparkhi, S. Roukos. 1994. Decision Tree Pars- ing using a Hidden Derivation Model. Proceedings o/ the 1994 Human Language Technology Work- shop, pages 272-277.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Statistical Decision-Tree Models for Parsing", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Magermaa", |
| "suffix": "" |
| } |
| ], |
| "year": 1995, |
| "venue": "Proceedings o/ the 33rd Annual Meeting o] the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "276--283", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "D. Magermaa. 1995. Statistical Decision-Tree Mod- els for Parsing. Proceedings o/ the 33rd Annual Meeting o] the Association for Computational Linguistics, pages 276-283.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Building a Large Annotated Corpus of English: the Penn Treebank", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Marcus", |
| "suffix": "" |
| }, |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Santorini", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Marcinkiewicz", |
| "suffix": "" |
| } |
| ], |
| "year": 1993, |
| "venue": "Computational Linguistics", |
| "volume": "19", |
| "issue": "2", |
| "pages": "313--330", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "M. Marcus, B. Santorini and M. Marcinkiewicz. 1993. Building a Large Annotated Corpus of En- glish: the Penn Treebank. Computational Linguis- tics, 19(2):313-330.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "The Penn Treebank: Annotating Predicate Argument Structure", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Marcus", |
| "suffix": "" |
| }, |
| { |
| "first": "G", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [ |
| "A" |
| ], |
| "last": "Marcinkiewicz", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Macintyre", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Bies", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Ferguson", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Katz", |
| "suffix": "" |
| }, |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Schasberger", |
| "suffix": "" |
| } |
| ], |
| "year": 1994, |
| "venue": "Proceedings of the 1994 Human Language Technology Workshop", |
| "volume": "", |
| "issue": "", |
| "pages": "110--115", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "M. Marcus, G. Kim, M. A. Marcinkiewicz, R. MacIntyre, A. Bies, M. Ferguson, K. Katz, B. Schasberger. 1994. The Penn Treebank: Annotat- ing Predicate Argument Structure. Proceedings of the 1994 Human Language Technology Workshop, pages 110~115.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "A Fully Statistical Approach to Natural Language Interfaces", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Miller", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Staliard", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Schwartz", |
| "suffix": "" |
| } |
| ], |
| "year": 1996, |
| "venue": "Proceedings o/ the 34th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "55--61", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "S. Miller, D. Staliard and R. Schwartz. 1996. A Fully Statistical Approach to Natural Language Interfaces. Proceedings o/ the 34th Annual Meeting of the Association for Computational Linguistics, pages 55-61.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "A Maximum Entropy Model for Part-Of-Speech Tagging", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Ratnaparkhi", |
| "suffix": "" |
| } |
| ], |
| "year": 1996, |
| "venue": "Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "A. Ratnaparkhi. 1996. A Maximum Entropy Model for Part-Of-Speech Tagging. Conference on Em- pirical Methods in Natural Language Processing.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Linguistic Dependency Modeling", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Stolcke", |
| "suffix": "" |
| } |
| ], |
| "year": 1996, |
| "venue": "Proceedings of ICSLP 96, Fourth International Conference on Spoken Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "A. Stolcke. 1996. Linguistic Dependency Modeling. Proceedings of ICSLP 96, Fourth International Conference on Spoken Language Processing.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "type_str": "figure", |
| "uris": null, |
| "num": null, |
| "text": "Generate modifiers to the left of the head with probability rL=l..n+ l ?) L ( L~( li ) l P, h, H), where Ln+l (ln+l) = STOP. For example, the probability of the rule S(bought) -> NP(week) NP(Marks) YP(bought)would be estimated as 7~h(YP I S,bought) x ~l(NP(Marks) I S,YP,bought) x 7~,(NP(week) { S,VP,bought) x 7~z(STOP I S,VP,bought) x ~r(STOP I S, VP, bought)" |
| }, |
| "FIGREF1": { |
| "type_str": "figure", |
| "uris": null, |
| "num": null, |
| "text": "2. Choose left and right subcat frames, LC and RC, with probabilities 7)~c(LC [ P, H, h) and" |
| }, |
| "FIGREF2": { |
| "type_str": "figure", |
| "uris": null, |
| "num": null, |
| "text": "The probability of the phrase S(bought)-> NP(week) NP-C(Marks) VP(bought)is now: NP-C(Marks) IS ,VP,bought, {NP-C}) x 7:~I(NP(week) I S ,VP ,bought, {}) x 7)l(STOe I S ,ve ,bought, {}) \u00d7 Pr(STOP I S, VP,bought, {})" |
| }, |
| "FIGREF3": { |
| "type_str": "figure", |
| "uris": null, |
| "num": null, |
| "text": "to its right. NP-C(Marks) is immediately generated as the required subject, and NP-C is removed from LC, leaving it empty when the next modifier, NP(week) is generated. The incorrect structures in figure 4 should now have low probability because ~Ic({NP-C,NP-C} [ S,VP,bought) and \"Prc({NP-C,VP-C} I VP,VB,was) are small. 2.3 Model 3: Traces and Wh-Movement Another obstacle to extracting predicate-argument structure from parse trees is wh-movement. This section describes a probabilistic treatment of extraction from relative clauses. Noun phrases are most often extracted from subject position, object position, or from within PPs: Example 1 The store (SBAR which TRACE bought Brooks Brothers) Example 2 The store (SBAR which Marks bought TRACE)" |
| }, |
| "FIGREF5": { |
| "type_str": "figure", |
| "uris": null, |
| "num": null, |
| "text": "ated. For example, Rule (2), SBAR(that) (+gap) -> WHNP(that) S-C(bought) (+gap), has probability~h (WHNP I SBAR, that) \u00d7 7~G (Right I SBAR, WHNP, that) x T~LC({} I SBAR,WHNP,that) x T'Rc({S-C} [ SBAR,WHNP, that) x 7~R (S-C (bought) (+gap) [ SBAR, WHNP, that, {S-C,+gap}) x 7~R(STOP I SBAR,WHNP,that, {}) x PC (STOP I SBAR, WHNP, that, { }) Rule (4), VP(bought) (+gap) -> VB(bought) TRACE NP (week), has probability 7~h(VB I VP,bought) x PG(Right I VP,bought,VB) x PLC({} I VP,bought,VB) x ~PRc({NP-C} I vP,bought,VB) x 7~R(TRACE I VP,bought,VB, {NP-C, +gap}) x PR(NP(week) I VP,bought ,VB, {}) \u00d7 7)L(STOP I VP,bought,VB, {}) x 7~R (STOP I VP ,bought ,VB, {}) In rule (2) Right is chosen, so the +gap requirement is added to RC. Generation of S-C(bought)(+gap) The life of a constituent in the chart." |
| }, |
| "FIGREF6": { |
| "type_str": "figure", |
| "uris": null, |
| "num": null, |
| "text": "x Y x ~R(Ri(ri) I P,H,...) (+) means a constituent is complete (i.e. it includes the stop probabilities), (-) means a constituent is incomplete. (a) a new constituent is started by projecting a complete rule upwards; (b) the constituent then takes left and right modifiers (or none if it is unary). (c) finally, STOP probabilities are added to complete the constituent. Back-off \"PH(H I\"-) Pa(G I ...) PL~(Li(It,) I..-) Level PLc(LC t ...) Pm(Ri(rti) I...) 7)Rc(RC I ...)" |
| }, |
| "TABREF0": { |
| "content": "<table/>", |
| "num": null, |
| "text": "", |
| "type_str": "table", |
| "html": null |
| }, |
| "TABREF2": { |
| "content": "<table><tr><td>Labeled Recall -~ number o/ correct constituents in proposed parse</td></tr><tr><td>number of constituents in treebank parse</td></tr></table>", |
| "num": null, |
| "text": "Results on Section 23 of the WSJ Treebank. LR/LP = labeled recall/precision. CBs is the average number of crossing brackets per sentence. 0 CBs, < 2 CBs are the percentage of sentences with 0 or < 2 crossing brackets respectively.", |
| "type_str": "table", |
| "html": null |
| } |
| } |
| } |
| } |