| { |
| "paper_id": "P83-1017", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T09:19:20.827289Z" |
| }, |
| "title": "Sentence Disambiguation by a Shift-Reduce Parsing Technique*", |
| "authors": [ |
| { |
| "first": "Stuart", |
| "middle": [ |
| "M" |
| ], |
| "last": "Shieber", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "Native speakers of English show definite and consistent preferences for certain readings of syntactically ambiguous sentences. A user of a natural-language-processing system would naturally expect it to reflect the same preferences. Thus, such systems must model in some way the linguistic performance as well as the linguistic competence of the native speaker. We have developed a parsing algorithm-a variant of the LALR(I} shift.-reduce algorithm-that models the preference behavior of native speakers for a range of syntactic preference phenomena reported in the psycholinguistic literature, including the recent data on lexical preferences. The algorithm yields the preferred parse deterministically, without building multiple parse trees and choosing among them. As a side effect, it displays appropriate behavior in processing the much discussed garden-path sentences. The parsing algorithm has been implemented and has confirmed the feasibility of our approach to the modeling of these phenomena.", |
| "pdf_parse": { |
| "paper_id": "P83-1017", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "Native speakers of English show definite and consistent preferences for certain readings of syntactically ambiguous sentences. A user of a natural-language-processing system would naturally expect it to reflect the same preferences. Thus, such systems must model in some way the linguistic performance as well as the linguistic competence of the native speaker. We have developed a parsing algorithm-a variant of the LALR(I} shift.-reduce algorithm-that models the preference behavior of native speakers for a range of syntactic preference phenomena reported in the psycholinguistic literature, including the recent data on lexical preferences. The algorithm yields the preferred parse deterministically, without building multiple parse trees and choosing among them. As a side effect, it displays appropriate behavior in processing the much discussed garden-path sentences. The parsing algorithm has been implemented and has confirmed the feasibility of our approach to the modeling of these phenomena.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "For natural language processing systems to be useful, they must assign the same interpretation to a given sentence that a native speaker would, since that is precisely the behavior users will expect.. Consider, for example, the case of ambiguous sentences. Native speakers of English show definite and consistent preferences for certain readings of syntactically ambiguous sentences [Kimball, 1973 , Frazier and Fodor, 1978 , Ford et aL, 1982 . A user of a natural-language-processing system would naturally expect, it to reflect the same preferences. Thus, such systems must model in some way the lineuistie performance as well as the linguistic competence of the native speaker. This idea is certainly not new in the artificial-intelligence literature. The pioneering work of Marcus [Marcus, 1980] is perhaps the best. known example of linguistic-performance modeling in AI. Starting from the hypothesis that ~deterministic\" parsing of English is possible, he demonstrated that certain performance \"This research was supported by the Defense Advanced Research Proiects Agency under Contract NOOO39-80-C-0575 with the Naval Electronic Systems Command. The views and conclusions contained in this document are those of the author and should not be interpreted a.s representative of the oh~cial policies, either expressed or implied, of the Defense Advanced Research Projects Agency or the United States government.", |
| "cite_spans": [ |
| { |
| "start": 383, |
| "end": 397, |
| "text": "[Kimball, 1973", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 398, |
| "end": 423, |
| "text": ", Frazier and Fodor, 1978", |
| "ref_id": null |
| }, |
| { |
| "start": 424, |
| "end": 442, |
| "text": ", Ford et aL, 1982", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 785, |
| "end": 799, |
| "text": "[Marcus, 1980]", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "constraints, e.g., the difl]culty of parsing garden-path sentences, could be modeled. His claim about deterministic parsing was quite strong. Not only was the behavior of the parser required to be deterministic, but, as Marcus claimed,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "The interpreter cannot use some general rule to take a nondeterministic grammar specification and impose arbitrary constraints to convert it to a deterministic specification {unless, of course, there is a general rule which will always lead to the correct decision in such a case). [Marcus, 1980, p.14] We have developed and implemented a parsing system that. given a nondeterministic grammar, forces disambiguation in just the manner Marcus rejected (i.e. t .hrough general rules}; it thereby exhibits the same preference behavior that psycbolinguists have attributed to native speakers of English for a certain range of ambiguities. These include structural ambiguities [Frazier and Fodor, 1978 , Frazier and Fodor, 1980 and lexical preferences [Ford et aL, 1982l , as well as the gardenpath sentences as a side effect. The parsing system is based on the shih.-reduee scheduling technique of Pereira [forthcoming] .", |
| "cite_spans": [ |
| { |
| "start": 282, |
| "end": 302, |
| "text": "[Marcus, 1980, p.14]", |
| "ref_id": null |
| }, |
| { |
| "start": 672, |
| "end": 696, |
| "text": "[Frazier and Fodor, 1978", |
| "ref_id": null |
| }, |
| { |
| "start": 697, |
| "end": 722, |
| "text": ", Frazier and Fodor, 1980", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 747, |
| "end": 765, |
| "text": "[Ford et aL, 1982l", |
| "ref_id": null |
| }, |
| { |
| "start": 894, |
| "end": 915, |
| "text": "Pereira [forthcoming]", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "Our parsing algorithm is a slight variant of LALR{ 1) parsing, and, as such, exhibits the three conditions postulated by Marcus for a deterministic mechanism: it is data-driven, reflects expectations, and has look-ahead. Like Marcus's parser, our parsing system is deterministic. Unlike Marcus's parser, the grammars used by our parser can be ambiguous.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "The parsing system was designed to manifest preferences among ,~tructurally distinct parses of ambiguous sentences. It, does this by building just one parse tree--rather than building multiple parse trees and choosing among them. Like the Marcus parsing system, ours does not do disambiguation requiring \"extensive semantic processing,\" hut, in contrast to Marcus, it does handle such phenomena as PP-attachment insofar as there exist a priori preferences for one attachment over another. By a priori we mean preferences that are exhibited in contexts where pragmatic or plausibility considerations do not tend to favor one reading over the other. Rather than make such value judgments ourselves, we defer to the psycholinguistic literature {specifically [Frazier and Fodor, 1978] , [Frazier and Fodor, 1980] and [Ford et al., 1982] ) for our examples. in which \"for Susan* modifies %he book\" rather than \"bought. \" Frazier and Fodor [1978] note that these are canes in which the higher attachment includes fewer nodes in the parse tree. Ore\" analysis is somewhat different. Ford et al. [10821 present evidence that attachment preferences depend on lexical choice. Thus, the preferred reading for", |
| "cite_spans": [ |
| { |
| "start": 755, |
| "end": 780, |
| "text": "[Frazier and Fodor, 1978]", |
| "ref_id": null |
| }, |
| { |
| "start": 783, |
| "end": 808, |
| "text": "[Frazier and Fodor, 1980]", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 813, |
| "end": 832, |
| "text": "[Ford et al., 1982]", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 914, |
| "end": 940, |
| "text": "\" Frazier and Fodor [1978]", |
| "ref_id": null |
| }, |
| { |
| "start": 1075, |
| "end": 1086, |
| "text": "Ford et al.", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Phenomena to be Modeled", |
| "sec_num": "2." |
| }, |
| { |
| "text": "The woman wanted the dresm on that rock.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lexical Preference", |
| "sec_num": null |
| }, |
| { |
| "text": "has low attachment of the PP, whereas", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lexical Preference", |
| "sec_num": null |
| }, |
| { |
| "text": "The tnoman positioned the dreu on that rack.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lexical Preference", |
| "sec_num": null |
| }, |
| { |
| "text": "has high attachment.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lexical Preference", |
| "sec_num": null |
| }, |
| { |
| "text": "The horse raced pamt the barn fell.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Garden-Path Sentences Grammatical sentences such as", |
| "sec_num": null |
| }, |
| { |
| "text": "seem actually to receive no parse by the native speaker until some sort of \"conscioun parsing\" is done. Following Marcus [Marcus, 1980] , we take this to be a hard failure of the human sentence-processing mechanism.", |
| "cite_spans": [ |
| { |
| "start": 121, |
| "end": 135, |
| "text": "[Marcus, 1980]", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Garden-Path Sentences Grammatical sentences such as", |
| "sec_num": null |
| }, |
| { |
| "text": "It will be seen that all these phenomena axe handled in oux parser by the same general rules. The simple context-free grammar used t (see Appendix I) allows both parses of the ambiguous sentences as well as one for the garden-path sentences. The parser disambiguates the grammar and yields only the preferred structure. The actual output of the parsing system can be found in Appendix II.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Garden-Path Sentences Grammatical sentences such as", |
| "sec_num": null |
| }, |
| { |
| "text": "The parsing system we use is a shift-reduce purser. Shiftreduce parsers [Aho and Johnson, 19741 the parse and a shift-reduce table for guiding the parse, At each step in the parse, the table is used for deciding between two basic types of operations: the shift operation, which adds the next word in the sentence (with its pretcrminal category) to the top of the stack, and the reduce operation, which removes several elements from the top of the stack and replaces them with a new element--for instance, removing an NP and a VP from the top of the stack and replacing them with an S. The state of the parser is also updated in accordance with the shift-reduce table at each stage. The combination of the stack, input, and state of the parser will be called a configuration and will be notated as, for example, 1 NPv IIMar, 110 1", |
| "cite_spans": [ |
| { |
| "start": 72, |
| "end": 95, |
| "text": "[Aho and Johnson, 19741", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Parsing System", |
| "sec_num": "3." |
| }, |
| { |
| "text": "where the stack contains the nonterminals NP and V, the input contains the lexical item Mary and the parser is in state 10.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Parsing System", |
| "sec_num": "3." |
| }, |
| { |
| "text": "By way of example, we demonstrate the operation of the parser (using the grammar of Appendix I) on the oft-cited sentence \"John loves Mary. ~ Initially the stack is empty and no input has been consumed. The parser begins in state 0.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Parsing System", |
| "sec_num": "3." |
| }, |
| { |
| "text": "As elements are shifted to the stack, they axe replaced by their preterminal category.\" T.he shiR-reduce table for the grammar of Appendix I states that in state 0, with a proper noun as the next word in the input, the appropriate action is a shift. the input has been consumed and an S derived. Thus the sentence is grammatical ia the grammar of Appendix I, as expected.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "II ahn 10.. Mar, i0 i", |
| "sec_num": null |
| }, |
| { |
| "text": "The shift-reduce table mentioned above is generated automatically from a context-free grammar by the standard algorithm [Aho and Johnson, 1974] . The parsing alogrithm differs, however, from the standard LALR(1) parsing algorithm in two ways. First, instead of assigning preterminal symbols to words as they are shifted, the algorithm allows the assignment to be delayed if the word is ambiguous among preterminals. When the word is used in a reduction, the appropriate preterminal is assigned.", |
| "cite_spans": [ |
| { |
| "start": 120, |
| "end": 143, |
| "text": "[Aho and Johnson, 1974]", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Differences from the Standard LR Techniques", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "Second, and most importantly, since true LR parsers exist only for unambiguous grammars, the normal algorithm for deriving LALR(1) shift-reduce tables yields a table that may specify conflicting actions under certain configurations. It is through the choice made from the options in a conflict that the preference behavior we desire is engendered.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Differences from the Standard LR Techniques", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "One key advantage of shift-reduce parsing that is critical in our system is the fact that decisions about the structure to be assigned to a phrase are postponed as long as possible. In keeping with this general principle, we extend the algorithm to allow the ~ssignment of a preterminal category to a lexical item to be deferred until a decision is forced upon it, so to speak, by aa encompassing reduction. For instance, we would not want to decide on the preterminal category of the word \"that,\" which can serve as either a determiner (DET) or complementizer (THAT), until some further information is available. Consider the sentences That problem i* important.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Preterminal Delaying", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "That problema are difficult to naive ia important.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Preterminal Delaying", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Instead of a.~signiag a preterminal to ~that,\" we leave open the possibility of assigning either DET or THAT until the first reduction that involves the word. In the first case, this reduction will be by the rule NP ~DET NOM, thus forcing, once and for all, the assignment of DET as preterminal. In the second ease, the DET NOM analysis is disallowed oa the basis of number agreement, so that the first applicable reduction is the COMPS reduction to S, forcing the assignment of THAT as preterminal.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Preterminal Delaying", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Of course, the question arises as to what state the parser goes into after shitting the lexical item ~that.\" The answer is quite straightforward, though its interpretation t,i~ d t,,a the determinism hypothesis is subtle. The simple answer is that the parser enters into a state corresponding to the union of the states entered upon shifting a DET and upon shifting a THAT respectively, in much the same way as the deterministic simulation of a nondeterministic finite automaton enters a ~uniou\" state when faced with a nondeterministic choice. Are we then merely simulating a aoadeterministic machine here. ~ The anss~er is equivocal. Although the implementation acts as a simulator for a nondeterministic machine, the nondeterminism is a priori bounded, given a particular grammar and lexicon. 3 Thus. the nondeterminism could be traded in for a larger, albeit still finite, set of states, unlike the nondeterminism found in other parsing algorithms. Another way of looking at the situation is to note that there is no observable property of the algorithm that would distinguish the operation of the parser from a deterministic one. In some sense, there is no interesting difference between the limited nondeterminism of this parser, and Marcus's notion of strict determinism. In fact, the implementation of Marcus's parser also embodies a bounded nondeterminism in much the same way this parser does.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Preterminal Delaying", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "The differentiating property between this parser and that of Marcus is a slightly different one, namely, the property of qaaM-real-time operation. 4 By quasi-real-time operation, Marcus means that there exists a maximum interval of parser operation for which no output can be generated. If the parser operates for longer than this, it must generate some output. For instance, the parser might be guaranteed to produce output (i.e., structure) at least every three words. However, because preterminal assignment can be delayed indefinitely in pathological grammars, there may exist sentences in such grammars for which arbitrary numbers of words need to be read before output can be produced. It is not clear whether this is a real disadvantage or not, and, if so, whether there are simple adjustments to the algorithm that would result in quasi-real-time behavior. In fact, it is a property of bottom-up parsing in general that quasi-real-time behavior is not guaranteed. Our parser has a less restrictive but similar property, fairneaH, that is, our parser generates output linear in the input, though there is no constant over which output is guaranteed. For a fuller discussion of these properties, see Pereira and Shieber [forthcoming].", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Preterminal Delaying", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "To summarize, preterminal delaying, as an intrinsic part of the algorithm, does not actually change the basic properties of the algorithm in any observable way. Note, however, that preterminal assignments, like reductions, are irrevocable once they are made {as a byproduct of the determinism of the algo-rithm}. Such decisions can therefore lead to garden paths, as they do for the sentences presented in Section 3.6.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Preterminal Delaying", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "We now discuss the central feature of the algorithm. namely, the resolution of shift-reduce conflicts.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Preterminal Delaying", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Conflicts arise in two ways: aM/t-reduce conflicts, in which the parser has the option of either shifting a word onto the stack or reducing a set of elements on the stack to a new element; reduce-reduce conflicts, in which reductions by several grammar 3The boundedness comes about because only a finite amount or informatie, n is kept per state (an integer) and the nondeterrninlsm stops at the prcterminat level, so that, the splitting of states does not. propogate, 41 am indebted to Mitch Marcus for this .bservation and the previous comparison with his parser.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Disambiguation Rules", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "rules are possible. The parser uses two rules to resolve these conflicts: 5", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Disambiguation Rules", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "(I) Resolve shift-reduce conflicts by shifting.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Disambiguation Rules", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "(2) Resolve reduce-reduce conflicts by performing the longer reduction.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Disambiguation Rules", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "These two rules suffice to engender the appropriate behavior in the parser for cases of right association and minimal attachment. Though we demonstrate our system primarily with PP-attachment examples, we claim that the rules are generally valid for the phenomena being modeled [Pereira and Shieber, forthcoming] .", |
| "cite_spans": [ |
| { |
| "start": 278, |
| "end": 312, |
| "text": "[Pereira and Shieber, forthcoming]", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Disambiguation Rules", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "Some examples demonstrate these principles. Consider the sentence Joe took the book that I bought for Sum,re.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Some Examples", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "After a certain amount of parsing has beta completed deterministically, the parser will be in the following coniigttration:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Some Examples", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "I NP v that V Ill\u00b0r S,... I", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Some Examples", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "with a shift-reduce confict, since the V can be reduced to a VP/NP \u00b0 or the P can be shifted. The principle* presented would solve the conflict in favor of the shift, thereby leading to the following derivation: The sentence 5The original notion of using a shift-reduce parser and general scheduling principles to handle right association and minlmal attachment, together with the following two rules, are due to Fernando Pereira [Pereira, 1982 [. The formalization of preterminal delaying and the extensions to the Ionic tlpreference cases and garden-path behavior are due to the author. 8The \"slash-category\" analysis of long-distance dependencies used here is loosely based on the work of Gaadar [lggl] . The Appendix 1 grammar does not incorporate the full range of slashed rules, however, but merely a representative selection for illustrative purposes.", |
| "cite_spans": [ |
| { |
| "start": 430, |
| "end": 444, |
| "text": "[Pereira, 1982", |
| "ref_id": null |
| }, |
| { |
| "start": 699, |
| "end": 705, |
| "text": "[lggl]", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Some Examples", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "NP V NP that NP V P l] Su,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Some Examples", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "Joe bou\u00a2ht the book for Su,an.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Some Examples", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "demonstrates resolution of a reduce-reduce conflict. At some point in the parse, the parser is in the following configuration: ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Some Examples", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "[ NP V", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Some Examples", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "To handle the lexical-preferenee examples, we extend the second rule slightly. Preterminal-word pairs can be stipulated as either weak or strong. The second rule becomes (2} Resolve reduce-reduce conflicts by performing the longest reduction with the stroncest &ftmost stack element. 7", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lexical Preference", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "Therefore, if it is assumed that the lexicon encodes the information that the triadic form of ~ant\" iV2 in the sample grammar) and the dyadic form of ~position\" (V1) are both weak, we can see the operation of the shift-reduce parser on the ~dress on that rack\" sentences of Section 2. Both sentences are similar in form and will thus have a similar configuration when the reduce-reduce conflict arises. For example, the first sentence will be in the following configuration:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lexical Preference", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "t NP wanted NP PP i[ 120 i", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lexical Preference", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "In this case, the longer reduction would require assignment of the preterminat category V2 to ~ant,\" which is the weak form: thus, the shorter reduction will be preferred, leading to the derivation: 7Note that, strength takes precedence over length.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lexical Preference", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "I NP", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lexical Preference", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "In the ca~e in which the verb is \"positioned,\" however, the longer reduction does not yield the weak form of the verb; it will therefore be invoked, reslting in the structure: As a side effect of these conflict resolution rules, certain sentences in the language of the grammar will receive no parse by the parsing system just discussed. These sentences are apparently the ones classified as \"garden-path\" sentences, a class that humans also have great difficulty parsing. Marcus's conjecture that such difficulty stems from a hard failure of the normal sentence-processing mechanism is directly modeled by the parsing system presented here.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lexical Preference", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "[", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lexical Preference", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "For instance, the sentence", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lexical Preference", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "The horse raced past the barn fell exhibits a reduce-reduce conflict before the last word. If the participial form of \"raced\" is weak, the finite verb form will be chosen; consequently, \"raced pant the barn\" will be reduced to a VP rather than a participial phrase. The parser will fail shortly, since the correct choice of reduction was not made.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lexical Preference", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "That scaly, deep-sea fish ,hould be underwater i~ important.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Similarly, the sentence", |
| "sec_num": null |
| }, |
| { |
| "text": "will fail. though grammatical. Before the word %hould\" is shifted, a reduce-reduce conflict arises in forming an NP from either \"That scaly, deep-sea l~h\" or \"scaly, deep-sea fish.\" The longer (incorrect} reduction will be performed and the parser will fail.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Similarly, the sentence", |
| "sec_num": null |
| }, |
| { |
| "text": "Other examples, e.g., \"the boy got fat melted,\" or \"the prime number few\" would be handled similarly by the parser, though the sample grammar of Appendix I does not parse them [Pcreira and Shieber, forthcoming].", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Similarly, the sentence", |
| "sec_num": null |
| }, |
| { |
| "text": "To be useful, aatttral-language systems must model the behavior, if not the method, of the native speaker. We have demonstrated that a parser using simple general rules for disambiguating sentences can yield appropriate behavior for a large class of performance phenomena--right a-~soeiation, minimal attachment, lexical preference, and garden-path sentences--and that, morever, it can do so deterministically wit, hour generating all the parses and choosing among them. The parsing system has been implemented and has confirmed the feasibility of ottr approach to the modeling of these phenomena.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "4." |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "The following is the grammar used to test the parting ~ystem descibed in the paper. Not a robust grammar of English by any means, it is presented only for the purpose of establishing that the preference rules yield the correct, results. ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Appendix I. The Test Grammar", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "LR Parsing", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [ |
| "V" |
| ], |
| "last": "Aho", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [ |
| "C" |
| ], |
| "last": "Johnson", |
| "suffix": "" |
| } |
| ], |
| "year": 1974, |
| "venue": "", |
| "volume": "6", |
| "issue": "", |
| "pages": "99--123", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Aho, A.V.. and S.C. Johnson, 1974: \"LR Parsing,\" Computi,, 9 Sur,,eys. Volume 6, Number 2, pp. 99-i24 ISpring).", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "A Competence-Based Theory of Syntactic Closure", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Ford", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Bresnan", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Kaplan", |
| "suffix": "" |
| } |
| ], |
| "year": 1982, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ford, M., J. Bresnan, and R. Kaplan, 1982: \"A Competence- Based Theory of Syntactic Closure,\" in The Mental Representation o/Grammatical Relations, J. Bresnan, ed. (Cambridge, Massachusetts: MIT Press).", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "~I'he Sausage Machine: A New Two-Stage Parsing Model", |
| "authors": [ |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Frazier", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [ |
| "D" |
| ], |
| "last": "Radar", |
| "suffix": "" |
| } |
| ], |
| "year": 1978, |
| "venue": "", |
| "volume": "6", |
| "issue": "", |
| "pages": "291--325", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Frazier, L., and J.D. radar, 1978: ~I'he Sausage Machine: A New Two-Stage Parsing Model,\" Cognition, Volume 6, pp. 291-325.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Is the Human Sentence Parsing Mechanism aa ATN?", |
| "authors": [ |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Frazier", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [ |
| "D" |
| ], |
| "last": "Fodor", |
| "suffix": "" |
| } |
| ], |
| "year": 1980, |
| "venue": "Cognition", |
| "volume": "8", |
| "issue": "", |
| "pages": "411--459", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Frazier, L., and J.D. Fodor, 1980: \"Is the Human Sentence Parsing Mechanism aa ATN?\" Cognition, Volume 8, pp. 411-459.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Unbounded dependencies and coordinate structure", |
| "authors": [ |
| { |
| "first": "G", |
| "middle": [], |
| "last": "Gazdar", |
| "suffix": "" |
| } |
| ], |
| "year": 1981, |
| "venue": "Linquistic Inquiry", |
| "volume": "12", |
| "issue": "", |
| "pages": "105--179", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Gazdar, G., 1981: \"Unbounded dependencies and coordinate structure,\" Linquistic Inquiry, Volume 12, pp. 105-179.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Seven Principles of Surface Structure Parsing in Natural Language", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Kimball", |
| "suffix": "" |
| } |
| ], |
| "year": 1973, |
| "venue": "Cognition", |
| "volume": "2", |
| "issue": "1", |
| "pages": "15--47", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kimball, d., 1973: \"Seven Principles of Surface Structure Parsing in Natural Language,\" Cognition, Volume 2, Number 1, pp. 15-47.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Theory of Syntactic Recognition/or Natural Lanquagc", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Marcus", |
| "suffix": "" |
| } |
| ], |
| "year": 1980, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Marcus, M., 1980: A Theory of Syntactic Recognition/or Natural Lanquagc, (Cambridge, Massachusetts: MIT Press).", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "A New Characterization of Attachment Preferences", |
| "authors": [ |
| { |
| "first": "F", |
| "middle": [ |
| "C N" |
| ], |
| "last": "Pereira", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Forthcoming", |
| "suffix": "" |
| } |
| ], |
| "year": null, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Pereira, F.C.N., forthcoming: \"A New Characterization of Attachment Preferences,\" to appear in D. Dowry, L. Karttunen, and A. gwicky (eds.)", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Natural Language Prate,int. Psyeholingui, tic, Computational, and Theoretical Perspective~", |
| "authors": [], |
| "year": null, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Natural Language Prate,int. Psyeholingui, tic, Computational, and Theoretical Perspective~, Cambridge, England: Cambridge University Press.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "ShiR-Reduce Scheduling and Syntactic Closure/ to appear", |
| "authors": [ |
| { |
| "first": "F", |
| "middle": [ |
| "C N" |
| ], |
| "last": "Pereira", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [ |
| "M" |
| ], |
| "last": "Shieber", |
| "suffix": "" |
| } |
| ], |
| "year": null, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Pereira, F.C.N., and S.M. Shieber, forthcoming: \"ShiR-Reduce Scheduling and Syntactic Closure/ to appear.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "The ATN and the Sausage Machine: Which One is Baloney?\" Caanition", |
| "authors": [ |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Wanner", |
| "suffix": "" |
| } |
| ], |
| "year": 1980, |
| "venue": "", |
| "volume": "8", |
| "issue": "", |
| "pages": "209--225", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wanner, E., 1980: \"The ATN and the Sausage Machine: Which One is Baloney?\" Caanition, Volume 8, pp. '209-225.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "TABREF1": { |
| "text": "axe a very general class of bottom-up parsers characterized by the following architecture. They incorporate a stock for holding constituents built up during", |
| "type_str": "table", |
| "num": null, |
| "html": null, |
| "content": "<table/>" |
| }, |
| "TABREF2": { |
| "text": "The verb and second proper noun axe now shifted, in accordance with the shift-reduce table, exhausting the input, and the proper noun is then reduced to an NP.", |
| "type_str": "table", |
| "num": null, |
| "html": null, |
| "content": "<table><tr><td/><td/><td/><td>The</td></tr><tr><td colspan=\"2\">new configuration, therefore, is</td><td/><td/></tr><tr><td>i</td><td>PNOUN</td><td>lo~e8 Mar~l</td><td colspan=\"2\">i 4 !</td></tr><tr><td colspan=\"4\">The next operation specified is a reduction of the proper noun</td></tr><tr><td colspan=\"2\">to a noun phrase yielding</td><td/><td/></tr><tr><td>,</td><td colspan=\"2\">NP iI loves Mary</td><td>[2</td><td>i</td></tr><tr><td/><td colspan=\"2\">NP v !l Ma,,</td><td>!1o</td></tr><tr><td/><td>v P. ouN il</td><td/><td>!,</td></tr><tr><td/><td>NP V NP i]</td><td/><td>:14</td></tr><tr><td colspan=\"5\">Finally, the verb and noun phrase on the top of the stack are</td></tr><tr><td>reduced to a VP</td><td/><td/><td/></tr><tr><td>i</td><td>NP VP !I II</td><td/><td>! ~6</td><td>l I</td></tr><tr><td colspan=\"5\">which is in turn reduced, together with the subject NP, to an S.</td></tr><tr><td>i</td><td colspan=\"2\">sJl</td><td colspan=\"2\">,'I )</td></tr><tr><td colspan=\"5\">This final configuration is an accepting configuration, since all</td></tr></table>" |
| }, |
| "TABREF6": { |
| "text": "sthe woman [vP positioned [Npthe dress][ppon that rackl]]", |
| "type_str": "table", |
| "num": null, |
| "html": null, |
| "content": "<table><tr><td>3.6 Garden-Path Sentences</td></tr></table>" |
| } |
| } |
| } |
| } |