ACL-OCL / Base_JSON /prefixP /json /P91 /P91-1030.json
Benjamin Aw
Add updated pkl file v3
6fa4bc9
{
"paper_id": "P91-1030",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T09:03:30.271278Z"
},
"title": "STRUCTURAL AMBIGUITY AND LEXICAL RELATIONS",
"authors": [
{
"first": "Donald",
"middle": [],
"last": "Hindle",
"suffix": "",
"affiliation": {
"laboratory": "AT&T Bell Laboratories",
"institution": "",
"location": {
"addrLine": "600 Mountain Avenue Murray Hill",
"postCode": "07974",
"region": "NJ"
}
},
"email": ""
},
{
"first": "Mats",
"middle": [],
"last": "Rooth",
"suffix": "",
"affiliation": {
"laboratory": "AT&T Bell Laboratories",
"institution": "",
"location": {
"addrLine": "600 Mountain Avenue Murray Hill",
"postCode": "07974",
"region": "NJ"
}
},
"email": ""
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "We propose that ambiguous prepositional phrase attachment can be resolved on the basis of the relative strength of association of the preposition with noun and verb, estimated on the basis of word distribution in a large corpus. This work suggests that a distributional approach can be effective in resolving parsing problems that apparently call for complex reasoning.",
"pdf_parse": {
"paper_id": "P91-1030",
"_pdf_hash": "",
"abstract": [
{
"text": "We propose that ambiguous prepositional phrase attachment can be resolved on the basis of the relative strength of association of the preposition with noun and verb, estimated on the basis of word distribution in a large corpus. This work suggests that a distributional approach can be effective in resolving parsing problems that apparently call for complex reasoning.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "Prepositional phrase attachment is the canonical case of structural ambiguity, as in the time worn example,",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": null
},
{
"text": "(1) I saw the man with the telescope The existence of such ambiguity raises problems for understanding and for language models. It looks like it might require extremely complex computation to determine what attaches to what. Indeed, one recent proposal suggests that resolving attachment ambiguity requires the construction of a discourse model in which the entities referred to in a text must be reasoned about (Altmann and Steedman 1988) . Of course, if attachment ambiguity demands reference to semantics and discourse models, there is little hope in the near term of building computational models for unrestricted text to resolve the ambiguity.",
"cite_spans": [
{
"start": 412,
"end": 439,
"text": "(Altmann and Steedman 1988)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": null
},
{
"text": "There have been several structure-based proposals about ambiguity resolution in the literature; they are particularly attractive because they are simple and don't demand calculations in the semantic or discourse domains. The two main ones are:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Structure based ambiguity resolution",
"sec_num": null
},
{
"text": "\u2022 Right Association -a constituent tends to attach to another constituent immediately to its right (Kimball 1973 ).",
"cite_spans": [
{
"start": 99,
"end": 112,
"text": "(Kimball 1973",
"ref_id": "BIBREF5"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Structure based ambiguity resolution",
"sec_num": null
},
{
"text": "\u2022 Minimal Attachment -a constituent tends to attach so as to involve the fewest additional syntactic nodes (Frazier 1978) .",
"cite_spans": [
{
"start": 107,
"end": 121,
"text": "(Frazier 1978)",
"ref_id": "BIBREF3"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Structure based ambiguity resolution",
"sec_num": null
},
{
"text": "For the particular case we are concerned with, attachment of a prepositional phrase in a verb + object context as in sentence (1), these two principles -at least in the version of syntax that Frazier assumes -make opposite predictions: Right Association predicts noun attachment, while Minimal Attachment predicts verb attachment.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Structure based ambiguity resolution",
"sec_num": null
},
{
"text": "Psycholinguistic work on structure-based strategies is primarily concerned with modeling the time course of parsing and disambiguation, and proponents of this approach explicitly acknowledge that other information enters into determining a final parse. Still, one can ask what information is relevant to determining a final parse, and it seems that in this domain structure-based disambiguation is not a very good predictor. A recent study of attachment of prepositional phrases in a sample of written responses to a \"Wizard of Oz\" travel information experiment shows that neither Right Association nor Minimal Attachment account for more than 55% of the cases (Whittemore et al. 1990) . And experiments by Taraban and McClelland (1988) show that the structural models are not in fact good predictors of people's behavior in resolving ambiguity.",
"cite_spans": [
{
"start": 661,
"end": 685,
"text": "(Whittemore et al. 1990)",
"ref_id": "BIBREF9"
},
{
"start": 707,
"end": 736,
"text": "Taraban and McClelland (1988)",
"ref_id": "BIBREF8"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Structure based ambiguity resolution",
"sec_num": null
},
{
"text": "Resolving ambiguity through lexical associations Whittemore et al. (1990) found lexical preferences to be the key to resolving attachment ambiguity. Similarly, Taraban and McClelland found lexical content was key in explaining people's behavior. Various previous proposals for guiding attachment disambiguation by the lexical content of specific words have appeared (e.g. Ford, Bresnan, and Kaplan 1982; Marcus 1980) . Unfortunately, it is not clear where the necessary information about lexical preferences is to be found. In the Whittemore et al. study, the judgement of attachment preferences had to be made by hand for exactly the cases that their study covered; no precompiled list of lexical preferences was available. Thus, we are posed with the problem: how can we get a good list of lexical preferences.",
"cite_spans": [
{
"start": 49,
"end": 73,
"text": "Whittemore et al. (1990)",
"ref_id": "BIBREF9"
},
{
"start": 372,
"end": 403,
"text": "Ford, Bresnan, and Kaplan 1982;",
"ref_id": "BIBREF2"
},
{
"start": 404,
"end": 416,
"text": "Marcus 1980)",
"ref_id": "BIBREF6"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Structure based ambiguity resolution",
"sec_num": null
},
{
"text": "Our proposal is to use cooccurrence of with prepositions in text as an indicator of lexical preference. Thus, for example, the preposition to occurs frequently in the context send NP --, i.e., after the object of the verb send, and this is evidence of a lexical association of the verb send with to. Similarly, from occurs frequently in the context withdrawal --, and this is evidence of a lexical association of the noun withdrawal with the preposition from. Of course, this kind of association is, unlike lexical selection, a symmetric notion. Cooccurrence provides no indication of whether the verb is selecting the preposition or vice versa. We will treat the association as a property of the pair of words. It is a separate matter, which we unfortunately cannot pursue here, to assign the association to a particular linguistic licensing relation. The suggestion which we want to explore is that the association revealed by textual distributionwhether its source is a complementation relation, a modification relation, or something else -gives us information needed to resolve the prepositional attachment.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Structure based ambiguity resolution",
"sec_num": null
},
{
"text": "A 13 million word sample of Associated Press new stories from 1989 were automatically parsed by the Fidditch parser (Hindle 1983) , using Church's part of speech analyzer as a preprocessor (Church 1988) . From the syntactic analysis provided by the parser for each sentence, we extracted a table containing all the heads of all noun phrases. For each noun phrase head, we recorded the following preposition if any occurred (ignoring whether or not the parser attached the preposition to the noun phrase), and the preceding verb if the noun phrase was the object of that verb. Thus, we generated a table with entries including those shown in Table 1 .",
"cite_spans": [
{
"start": 116,
"end": 129,
"text": "(Hindle 1983)",
"ref_id": "BIBREF4"
},
{
"start": 189,
"end": 202,
"text": "(Church 1988)",
"ref_id": "BIBREF1"
}
],
"ref_spans": [
{
"start": 641,
"end": 648,
"text": "Table 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "Discovering Lexical Association in Text",
"sec_num": null
},
{
"text": "In Table 1 , example (a) represents a passivized instance of the verb blame followed by the prepo- sition for. Example (b) is an instance of a noun phrase whose head is money; this noun phrase is not an object of any verb, but is followed by the preposition for. Example (c) represents an instance of a noun phrase with head noun development which neither has a following preposition nor is the object of a verb. Example (d) is an instance of a noun phrase with head government, which is the object of the verb control but is followed by no preposition. Example (j) represents an instance of the ambiguity we are concerned with resolving: a noun phrase (head is concession), which is the object of a verb (grant), followed by a preposition",
"cite_spans": [],
"ref_spans": [
{
"start": 3,
"end": 10,
"text": "Table 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "Discovering Lexical Association in Text",
"sec_num": null
},
{
"text": "From the 13 million word sample, 2,661,872 noun phrases were identified. Of these, 467,920 were recognized as the object of a verb, and 753,843 were followed by a preposition. Of the noun phrase objects identified, 223,666 were ambiguous verb-noun-preposition triples.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "(to).",
"sec_num": null
},
{
"text": "Of course, the table of verbs, nouns and prepositions does not directly tell us what the strength lexical associations are. There are three potential sources of noise in the model. First, the parser in some cases gives us false analyses. Second, when a preposition follows a noun phrase (or verb), it may or may not be structurally related to that noun phrase (or verb). (In our terms, it may attach to that noun phrase or it may attach somewhere else). And finally, even if we get accurate attachment information, it may be that fre-quency of cooccurrence is not a good indication of strength of attachment. We will proceed to build the model of lexical association strength, aware of these sources of noise.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Estimating attachment preferences",
"sec_num": null
},
{
"text": "We want to use the verb-noun-preposition table to derive a table of bigrams, where the first term is a noun or verb, and the second term is an associated preposition (or no preposition). To do this we need to try to assign each preposition that occurs either to the noun or to the verb that it occurs with. In some cases it is fairly certain that the preposition attaches to the noun or the verb; in other cases, it is far less certain. Our approach is to assign the clear cases first, then to use these to decide the unclear cases that can be decided, and finally to arbitrarily assign the remaining cases. The procedure for assigning prepositions in our sample to noun or verb is as follows:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Estimating attachment preferences",
"sec_num": null
},
{
"text": "1. No Preposition -if there is no preposition, the noun or verb is simply counted with the null preposition. (cases (c-h) in Table 1 ).",
"cite_spans": [],
"ref_spans": [
{
"start": 125,
"end": 132,
"text": "Table 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "Estimating attachment preferences",
"sec_num": null
},
{
"text": "2. Sure Verb Attach 1 -preposition is attached to the verb if the noun phrase head is a pronoun. (i in Table 1) 3. Sure Verb Attach 2 -preposition is attached to the verb if the verb is passivized (unless the preposition is by. The instances of by following a passive verb were left unassigned.) (a in Table 1) 4. Sure Noun Attach -preposition is attached to the noun, if the noun phrase occurs in a context where no verb could license the prepositional phrase (i.e., the noun phrase is in subject or pre-verbal position.) (b, if pre-verbal)",
"cite_spans": [],
"ref_spans": [
{
"start": 103,
"end": 111,
"text": "Table 1)",
"ref_id": null
},
{
"start": 302,
"end": 310,
"text": "Table 1)",
"ref_id": null
}
],
"eq_spans": [],
"section": "Estimating attachment preferences",
"sec_num": null
},
{
"text": "5. Ambiguous Attach 1 -Using the table of attachment so far, if a t-score for the ambiguity (see below) is greater than 2.1 or less than -2.1, then assign the preposition according to the t-score. Iterate through the ambiguous triples until all such attachments are done. (j and k may be assigned)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Estimating attachment preferences",
"sec_num": null
},
{
"text": "6. Ambiguous Attach 2 -for the remaining ambiguous triples, split the attachment between the noun and the verb, assigning .5 to the noun and .5 to the verb. (j and k may be assigned)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Estimating attachment preferences",
"sec_num": null
},
{
"text": "7. Unsure Attach -for the remaining pairs (all of which are either attached to the preceding noun or to some unknown element), assign them to the noun. (b, if following a verb)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Estimating attachment preferences",
"sec_num": null
},
{
"text": "This procedure gives us a table of bigrams representing our guess about what prepositions associate with what nouns or verbs, made on the basis of the distribution of verbs nouns and prepositions in our corpus.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Estimating attachment preferences",
"sec_num": null
},
{
"text": "The procedure for guessing attachment Given the table of bigrams, derived as described above, we can define a simple procedure for determining the attachment for an instance of verbnoun-preposition ambiguity. Consider the example of sentence (2), where we have to choose the attachment given verb send, noun soldier, and preposition into.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Estimating attachment preferences",
"sec_num": null
},
{
"text": "(2) Moscow sent more than 100,000 soldiers into Afganistan ...",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Estimating attachment preferences",
"sec_num": null
},
{
"text": "The idea is to contrast the probability with which into occurs with the noun soldier (P(into [ soldier)) with the probability with which into occurs with the verb send (P(into [ send)). A tscore is an appropriate way to make this contrast (see Church et al. to appear) . In general, we want to calculate the contrast between the conditional probability of seeing a particular preposition given a noun with the conditional probability of seeing that preposition given a verb.",
"cite_spans": [
{
"start": 244,
"end": 268,
"text": "Church et al. to appear)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Estimating attachment preferences",
"sec_num": null
},
{
"text": "t= ~/a2(P(prep I noun)) + ~2(e(prep I verb))",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "P(prep [ noun) -P(prep [ verb)",
"sec_num": null
},
{
"text": "We use the \"Expected Likelihood Estimate\" (Church et al., to appear) to estimate the probabilities, in order to adjust for small frequencies; that is, given a noun and verb, we simply add 1/2 to all bigram frequency counts involving a preposition that occurs with either the noun or the verb, and then recompute the unigram frequencies. This method leaves the order of t-scores nearly intact, though their magnitude is inflated by about 30%. To compensate for this, the 1.65 threshold for significance at the 95% level should be adjusted up to about 2.15.",
"cite_spans": [
{
"start": 42,
"end": 68,
"text": "(Church et al., to appear)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "P(prep [ noun) -P(prep [ verb)",
"sec_num": null
},
{
"text": "Consider how we determine attachment for sentence (2). We use a t-score derived from the adjusted frequencies in our corpus to decide whether the prepositional phrase into Afganistan is attached to the verb (root) send/V or to the noun (root) soldier/N. In our corpus, soldier/N has an adjusted frequency of 1488.5, and send/V has an adjusted frequency of 1706.5; soldier/N occurred in 32 distinct preposition contexts, and send/Via 60 distinct preposition contexts; f(send/V into) = 84, f(soidier/N into) = 1.5. From this we calculate the t-score as follows: 1 decide that into should attach to send rather than to soldier. Of the 84 send/V into bigrams, 10 were assigned by steps 2 and 3 ('sure attachements').",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "P(prep [ noun) -P(prep [ verb)",
"sec_num": null
},
{
"text": "t- P(wlsoldier/ N ) -P(wlsend/ V) ~/a2(P(wlsoidier/N)) + c~2(P(wlsend/ V)) l(soldier/N into)+ll2 .f(send/V into)+l/2 f(soidierlN)+V/2 -- /(send/V)+V/2 \\//(,oldier/N into)+l/2 /(send/V into)+l[2 (f(soldierlN)+V/2)2 + (/(send/V",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "P(prep [ noun) -P(prep [ verb)",
"sec_num": null
},
{
"text": "To evaluate the performance of this procedure, first the two authors graded a set of verb-nounpreposition triples as follows. From the AP new stories, we randomly selected 1000 test sentences in which the parser identified an ambiguous verbnoun-preposition triple. (These sentences were selected from stories included in the 13 million word sample, but the particular sentences were excluded from the calculation of lexical associations.) For every such triple, each author made a judgement of the correct attachment on the basis of the three words alone (forced choice -preposition attaches to noun or verb). This task is in essence the one that we will give the computer -i.e., to judge the attachment without any more information than the preposition and the head of the two possible attachment sites, the noun and the verb. This gave us two sets of judgements to compare the algorithm's performance to. ",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Testing Attachment Preference",
"sec_num": null
},
{
"text": "We also wanted a standard of correctness for these test sentences. To derive this standard, we together judged the attachment for the 1000 triples a second time, this time using the full sentence context. It turned out to be a surprisingly difficult task to assign attachment preferences for the test sample. Of course, many decisions were straightforward; sometimes it is clear that a prepositional phrase is and argument of a noun or verb. But more than 10% of the sentences seemed problematic to at least one author. There are several kinds of constructions where the attachment decision is not clear theoretically. These include idioms (3-4), light verb constructions (5), small clauses (6).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Judging correct attachment",
"sec_num": null
},
{
"text": "(3) But over time, misery has given way to mending.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Judging correct attachment",
"sec_num": null
},
{
"text": "(4) The meeting will take place in Quanrico (5) Bush has said he would not make cuts in Social Security (6) Sides said Francke kept a .38-caliber revolver in his car's glove compartment",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Judging correct attachment",
"sec_num": null
},
{
"text": "We chose always to assign light verb constructions to noun attachment and small clauses to verb attachment.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Judging correct attachment",
"sec_num": null
},
{
"text": "Another source of difficulty arose from cases where there seemed to be a systematic ambiguity in attachment. 7shows a systematic locative ambiguity: if you frequent a bar and the bar is in a place, the frequenting event is arguably in the same place. Sentence (8) shows a systematic benefactive ambiguity: if you arrange something for someone, then the thing arranged is also for them. The ambiguity in (9) arises from the fact that if someone is one of the joint agents in the signing of an agreement, that person is likely to be a party to the agreement. In general, we call an attachment systematically ambiguous when, given our understanding of the semantics, situations which make the interpretation of one of the attachments true always (or at least usually) also validate the interpretation of the other attachment.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Judging correct attachment",
"sec_num": null
},
{
"text": "It seems to us that this difficulty in assigning attachment decisions is an important fact that deserves further exploration. If it is difficult to decide what licenses a prepositional phrase a significant proportion of the time, then we need to develop language models that appropriately capture this vagueness. For our present purpose, we decided to force an attachment choice in all cases, in some cases making the choice on the bases of an unanalyzed intuition.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Judging correct attachment",
"sec_num": null
},
{
"text": "In addition to the problematic cases, a significant number (120) of the 1000 triples identified automatically as instances of the verb-objectpreposition configuration turned out in fact to be other constructions. These misidentifications were mostly due to parsing errors, and in part due to our underspecifying for the parser exactly what configuration to identify. Examples of these misidentifications include: identifying the subject of the complement clause of say as its object, as in (10), which was identified as (say ministers from); misparsing two constituents as a single object noun phrase, as in (11), which was identified as (make subject to); and counting non-object noun phrases as the object as in (12), identified as (get hell out_oJ). Of course these errors are folded into the calculation of associations. No doubt our bigram model would be better if we could eliminate these items, but many of them represent parsing errors that cannot readily be identified by the parser, so we proceed with these errors included in the bigrams.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Judging correct attachment",
"sec_num": null
},
{
"text": "After agreeing on the 'correct' attachment for the sample of 1000 triples, we are left with 880 verb-noun-preposition triples (having discarded the 120 parsing errors). Of these, 586 are noun attachments and 294 verb attachments.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Judging correct attachment",
"sec_num": null
},
{
"text": "First, consider how the simple structural attachment preference schemas perform at predicting the Judge 1 I i i i i 4.9 i LA 557 323 85.4 65.9 78.3 Now consider the performance of our attachment procedure for the 880 standard test sentences. Table 2 shows the performance for the two human judges and for the lexical association attachment procedure.",
"cite_spans": [],
"ref_spans": [
{
"start": 242,
"end": 249,
"text": "Table 2",
"ref_id": "TABREF2"
}
],
"eq_spans": [],
"section": "Evaluating performance",
"sec_num": null
},
{
"text": "First, we note that the task of judging attachment on the basis of verb, noun and preposition alone is not easy. The human judges had overall error rates of 10-15%. (Of course this is considerably better than always choosing noun attachment.) The lexical association procedure based on t-scores is somewhat worse than the human judges, with an error rate of 22%, but this also is an improvement over simply choosing the nearest attachment site.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Evaluating performance",
"sec_num": null
},
{
"text": "If we restrict the lexical association procedure to choose attachment only in cases where its confidence is greater than about 95% (i.e., where t is greater than 2.1), we get attachment judgements on 607 of the 880 test sentences, with an overall error rate of 15% (Table 3) . On these same sentences, the human judges also showed slight improvement. Table 3 : Performance on the test sentences for 2 human judges and the lexical association procedure (LA) for test triples where t > 2.1 cording to the nature of the relationship underlying the attachment. In many cases, the decision was difficult. Even the argument/adjunct distinction showed many gray cases between clear participants in an action (arguments) and clear temporal modifiers (adjuncts). We made rough best guesses to partition the cases into the following categories: argument, adjunct, idiom, small clause, locative ambiguity, systematic ambiguity, light verb. With this set of categories, 84 of the 880 cases remained so problematic that we assigned them to category other. Table 4 shows the performance of the lexical attachment procedure for these classes of relations. Even granting the roughness of the categorization, some clear patterns emerge. Our approach is quite successful at attaching arguments correctly; this represents some confirmation that the associations derived from the AP sample are indeed the kind of associations previous research has suggested are relevant to determining attachment. The procedure does better on arguments than on adjuncts, and in fact performs rather poorly on adjuncts of verbs (chiefly time and manner phrases). The remaining cases are all hard in some way, and the performance tends to be worse on these cases, showing clearly for a more elaborated model.",
"cite_spans": [],
"ref_spans": [
{
"start": 265,
"end": 274,
"text": "(Table 3)",
"ref_id": null
},
{
"start": 351,
"end": 358,
"text": "Table 3",
"ref_id": null
},
{
"start": 1043,
"end": 1050,
"text": "Table 4",
"ref_id": "TABREF5"
}
],
"eq_spans": [],
"section": "Evaluating performance",
"sec_num": null
},
{
"text": "The initial steps of our procedure constructed a table of frequencies with entries f(z,p), where z is a noun or verb root string, and p is a preposition string. These primitives might be too coarse, in that they do not distinguish different senses of a preposition, noun, or verb. For instance, the tem-porM use of in in the phrase in December is identified with a locative use in Teheran. As a result, the procedure LA necessarily makes the same attach- We experimented with the first problem by substituting an abstract preposition in,MONTH for all occurrences of in with a month name as an object. While the tuple reopen embassy in~oMONTH was correctly pushed in the direction of a verb attachment (-1.34), in other cases errors were introduced, and there was no compelling general improvement in performance. In tuples of the form drop/grow/increase percent inJ~MONTH , derived from examples such as (16), the preposition was incorrectly attached to the noun percent.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Sense Conflations",
"sec_num": null
},
{
"text": "(16) Output at mines and oil wells dropped 1.8 percent in February (17) ,1.8 percent was dropped by output at mines and oil wells",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Sense Conflations",
"sec_num": null
},
{
"text": "We suspect that this reveals a problem with our estimation procedure, not for instance a paucity of data. Part of the problem may be the fact that adverbial noun phrase headed by percent in 16does not passivize or pronominalize, so that there are no sure verb attachment cases directly corresponding to these uses of scalar motion verbs.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Sense Conflations",
"sec_num": null
},
{
"text": "The idea that lexical preference is a key factor in resolving structural ambiguity leads us naturally to ask whether existing dictionaries can provide useful information for disambiguation. There are reasons to anticipate difficulties in this regard. Typically, dictionaries have concentrated on the 'interesting' phenomena of English, tending to ignore mundane lexical associations. However, the Collins Cobuild English Language Dictionary (Sinclair et al. 1987 ) seems particularly appropriate for comparing with the AP sample for several reasons: it was compiled on the basis of a large text corpus, and thus may be less subject to idiosyncrasy than more arbitrarily constructed works; and it provides, in a separate field, a direct indication of prepositions typically associated with many nouns and verbs. Nevertheless, even for Cobuild, we expect to find more concentration on, for example, idioms and closely bound arguments, and less attention to the adjunct relations which play a significant role in determining attachment preferences.",
"cite_spans": [
{
"start": 441,
"end": 462,
"text": "(Sinclair et al. 1987",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Comparison with a Dictionary",
"sec_num": null
},
{
"text": "From a machine-readable version of the dictionary, we extracted a list of 1535 nouns associated with a particular preposition, and of 1193 verbs associated with a particular preposition after an object noun phrase. These 2728 associations are many fewer than the number of associations found in the AP sample. (see Table 5 .)",
"cite_spans": [],
"ref_spans": [
{
"start": 315,
"end": 322,
"text": "Table 5",
"ref_id": "TABREF6"
}
],
"eq_spans": [],
"section": "Comparison with a Dictionary",
"sec_num": null
},
{
"text": "Of course, most of the preposition association pairs from the AP sample end up being nonsignificant; of the 88,860 pairs, fewer than half (40,869) occur with a frequency greater than 1, and only 8337 have a t-score greater than 1.65. So our sample gives about three times as many significant preposition associations as the COBUILD dictionary. Note however, as Table 5 : Count of noun and verb associations for COBUILD and the AP sample there are over 20,000 nouns, over 5000 verbs, and over 90 prepositions.) On the other hand, the lack of overlap for so many cases -assuming that the dictionary and the significant bigrams actually record important preposition associations -indicates that 1) our sample is too small, and 2) the dictionary coverage is widely scattered. First, we note that the dictionary chooses attachments in 182 cases of the 880 test sentences. Seven of these are cases where the dictionary finds an association between the preposition and both the noun and the verb. In these cases, of course, the dictionary provides no information to help in choosing the correct attachment.",
"cite_spans": [],
"ref_spans": [
{
"start": 361,
"end": 368,
"text": "Table 5",
"ref_id": "TABREF6"
}
],
"eq_spans": [],
"section": "Comparison with a Dictionary",
"sec_num": null
},
{
"text": "Looking at the 175 cases where the dictionary finds one and only one association for the preposition, we can ask how well it does in predicting the correct attachment. Here the results are no better than our human judges or than our bigram procedure. Of the 175 cases, in 25 cases the dictionary finds a verb association when the correct association is with the noun. In 3 cases, the dictionary finds a noun association when the correct association is with the verb. Thus, overall, the dictionary is 86% correct.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Comparison with a Dictionary",
"sec_num": null
},
{
"text": "It is somewhat unfair to use a dictionary as a source of disambiguation information; there is no reason to expect that a dictionary to provide information on all significant associations; it may record only associations that are interesting for some reason (perhaps because they are semantically unpredictable.) Table 6 shows a small sample of verb-preposition associations from the AP sam- Table 6 : Verb-(NP)-Preposition associations in AP sample and COBUILD. pie and from Cobuild. The overlap is considerable, but each source of information provides intuitively important associations that are missing from the other.",
"cite_spans": [],
"ref_spans": [
{
"start": 312,
"end": 319,
"text": "Table 6",
"ref_id": null
},
{
"start": 391,
"end": 398,
"text": "Table 6",
"ref_id": null
}
],
"eq_spans": [],
"section": "Comparison with a Dictionary",
"sec_num": null
},
{
"text": "Our attempt to use lexical associations derived from distribution of lexical items in text shows promising results. Despite the errors in parsing introduced by automatically analyzing text, we are able to extract a good list of associations with prepositions, overlapping significantly with an existing dictionary. This information could easily be incorporated into an automatic parser, and additional sorts of lexical associations could similarly be derived from text. The particular approach to deciding attachment by t-score gives results nearly as good as human judges given the same information. Thus, we conclude that it may not be necessary to resort to a complete semantics or to discourse models to resolve many pernicious cases of attachment ambiguity.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion",
"sec_num": null
},
{
"text": "It is clear however, that the simple model of attachment preference that we have proposed, based only on the verb, noun and preposition, is too weak to make correct attachments in many cases. We need to explore ways to enter more complex calculations into the procedure.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion",
"sec_num": null
},
{
"text": "(13) is a phrase from our corpus, while (14) is a constructed example.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
}
],
"back_matter": [],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "Interaction with context during human sentence processing",
"authors": [
{
"first": "Gerry",
"middle": [],
"last": "Altmman",
"suffix": ""
},
{
"first": "Mark",
"middle": [],
"last": "Steedman",
"suffix": ""
}
],
"year": 1988,
"venue": "Cognition",
"volume": "30",
"issue": "",
"pages": "191--238",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Altmman, Gerry, and Mark Steedman. 1988. Interac- tion with context during human sentence process- ing. Cognition, 30, 191-238.",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "A stochastic parts program and noun phrase parser for unrestricted text",
"authors": [
{
"first": "Kenneth",
"middle": [
"W"
],
"last": "Church",
"suffix": ""
}
],
"year": 1988,
"venue": "Proceedings of the Second Conference on Applied Natural Language Processing",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Church, Kenneth W. 1988. A stochastic parts program and noun phrase parser for unrestricted text, Proceedings of the Second Conference on Applied Natural Language Processing, Austin, Texas. Church, Kenneth W., William A. Gale, Patrick Hanks, and Donald Hindle. (to appear). Using statistics in lexical analysis, in Zernik (ed.) Lexical acqui- sition: using on-line resources to build a lexicon.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "The Mental Represen. tation o.f Grammatical Relations",
"authors": [
{
"first": "Marilyn",
"middle": [],
"last": "Ford",
"suffix": ""
},
{
"first": "Joan",
"middle": [],
"last": "Bresnan",
"suffix": ""
},
{
"first": "Ronald",
"middle": [
"M"
],
"last": "Kaplan",
"suffix": ""
}
],
"year": 1982,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Ford, Marilyn, Joan Bresnan and Ronald M. Kaplan. 1982. A competence based theory of syntactic clo- sure, in Bresnan, J. (ed.) The Mental Represen. tation o.f Grammatical Relations. MIT Press.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "On comprehending sentences: Syntactic parsing strategies",
"authors": [
{
"first": "L",
"middle": [],
"last": "Frazier",
"suffix": ""
}
],
"year": 1978,
"venue": "University of Connecticut",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Frazier, L. 1978. On comprehending sentences: Syn- tactic parsing strategies. PhD. dissertation, Uni- versity of Connecticut.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "User manual for fidditch, a deterministic parser",
"authors": [
{
"first": "Donald",
"middle": [],
"last": "Hindle",
"suffix": ""
}
],
"year": 1983,
"venue": "Naval Research Laboratory Technical Memorandum",
"volume": "",
"issue": "",
"pages": "7590--142",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Hindle, Donald. 1983. User manual for fidditch, a deterministic parser. Naval Research Laboratory Technical Memorandum 7590-142.",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "Seven principles of surface structure parsing in natural language",
"authors": [
{
"first": "J",
"middle": [],
"last": "Kimball",
"suffix": ""
}
],
"year": 1973,
"venue": "Cognition",
"volume": "2",
"issue": "",
"pages": "15--47",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Kimball, J. 1973. Seven principles of surface structure parsing in natural language, Cognition, 2, 15-47.",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "A theory of syntactic recognition for natural language",
"authors": [
{
"first": "Mitchell",
"middle": [
"P"
],
"last": "Marcus",
"suffix": ""
}
],
"year": 1980,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Marcus, Mitchell P. 1980. A theory of syntactic recog- nition for natural language. MIT Press.",
"links": null
},
"BIBREF8": {
"ref_id": "b8",
"title": "Constituent attachment and thematic role assignment in sentence processing: influences of content-based expectations",
"authors": [
{
"first": "Roman",
"middle": [],
"last": "Taraban",
"suffix": ""
},
{
"first": "James",
"middle": [
"L"
],
"last": "Mcclelland",
"suffix": ""
}
],
"year": 1988,
"venue": "Journal of Memory and Language",
"volume": "27",
"issue": "",
"pages": "597--632",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Taraban, Roman and James L. McClelland. 1988. Constituent attachment and thematic role as- signment in sentence processing: influences of content-based expectations, Journal of Memory and Language, 27, 597-632.",
"links": null
},
"BIBREF9": {
"ref_id": "b9",
"title": "Empirical study of predictive powers of simple attachment schemes for post-modifier prepositional phrases",
"authors": [
{
"first": "Greg",
"middle": [],
"last": "Whittemore",
"suffix": ""
},
{
"first": "Kathleen",
"middle": [],
"last": "Ferrara",
"suffix": ""
},
{
"first": "Hans",
"middle": [],
"last": "Brunnet",
"suffix": ""
}
],
"year": 1990,
"venue": "Proceedings of the ~8th Annual Meeting of the Association for Computational Linguistics",
"volume": "",
"issue": "",
"pages": "23--30",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Whittemore, Greg, Kathleen Ferrara and Hans Brun- net. 1990. Empirical study of predictive powers of simple attachment schemes for post-modifier prepositional phrases. Proceedings of the ~8th An- nual Meeting of the Association for Computa- tional Linguistics, 23-30.",
"links": null
}
},
"ref_entries": {
"FIGREF0": {
"type_str": "figure",
"num": null,
"text": "/2p -I-1706.s+70/2)2 This figure of-8.81 represents a significant association of the preposition into with the verb send, and on this basis, the procedure would (correctly)",
"uris": null
},
"FIGREF1": {
"type_str": "figure",
"num": null,
"text": "(7) ...known to frequent the same bars in one neighborhood. (8) Inaugural officials reportedly were trying to arrange a reunion for Bush and his old submarine buddies ... (9) We have not signed a settlement agreement with them Sentence",
"uris": null
},
"FIGREF2": {
"type_str": "figure",
"num": null,
"text": "Ortega also said deputy foreign ministers from the five governments would meet Tuesday in Managua .... (11) Congress made a deliberate choice to make this commission subject to the open meeting requirements ... (12) Student Union, get the hell out of China!",
"uris": null
},
"TABREF2": {
"type_str": "table",
"num": null,
"text": "",
"content": "<table><tr><td>: Performance on the test sentences for 2</td></tr><tr><td>human judges and the lexical association proce-</td></tr><tr><td>dure (LA).</td></tr><tr><td>outcome in our test set. Right Association, which</td></tr><tr><td>predicts noun attachment, does better, since in</td></tr><tr><td>our sample there are more noun attachments, but</td></tr><tr><td>it still has an error rate of 33%. Minimal Attach.</td></tr><tr><td>meat, interpreted to mean verb attachment, has</td></tr><tr><td>the complementary error rate of 67%. Obviously,</td></tr><tr><td>neither of these procedures is particularly impres-</td></tr><tr><td>sive.</td></tr></table>",
"html": null
},
"TABREF5": {
"type_str": "table",
"num": null,
"text": "",
"content": "<table><tr><td>: Performance of the Lexical attachment</td></tr><tr><td>procedure by underlying relationship</td></tr><tr><td>ment prediction for in December and in Teheran</td></tr><tr><td>occurring in the same context. For instance, LA</td></tr><tr><td>identifies the tuple reopen embassy in as an NP at-</td></tr><tr><td>tachment (t-score 5.02). This is certainly incorrect</td></tr><tr><td>for (13), though not for (14). 2</td></tr><tr><td>(13) Britain reopened the embassy in De-</td></tr><tr><td>cember</td></tr><tr><td>(14) Britain reopened its embassy in</td></tr><tr><td>Teheran</td></tr><tr><td>Similarly, the scalar sense of drop exemplified in</td></tr><tr><td>(15) sponsors a preposition to, while the sense rep-</td></tr><tr><td>resented in drop the idea does not. Identifying the</td></tr><tr><td>two senses may be the reason that LA makes no</td></tr><tr><td>attachment choice for drop resistance to (derived</td></tr><tr><td>from (16)), where the score is -0.18.</td></tr><tr><td>(15) exports are expected to drop a fur-</td></tr><tr><td>ther 1.5 percent to 810,000</td></tr><tr><td>(16) persuade Israeli leaders to drop their</td></tr><tr><td>resistance to talks with the PLO</td></tr></table>",
"html": null
},
"TABREF6": {
"type_str": "table",
"num": null,
"text": "",
"content": "<table><tr><td>shows, the</td></tr><tr><td>overlap is remarkably good, considering the large</td></tr><tr><td>space of possible bigrams. (In our bigram table</td></tr></table>",
"html": null
}
}
}
}