ACL-OCL / Base_JSON /prefixJ /json /J02 /J02-1005.json
Benjamin Aw
Add updated pkl file v3
6fa4bc9
{
"paper_id": "J02-1005",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T02:44:08.519469Z"
},
"title": "Squibs and Discussions",
"authors": [
{
"first": "Mark",
"middle": [],
"last": "Johnson",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "Brown University",
"location": {}
},
"email": "markjohnson@brown.edu"
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "A data-oriented parsing or DOP model for statistical parsing associates fragments of linguistic representations with numerical weights, where these weights are estimated by normalizing the empirical frequency of each fragment in a training corpus (see Bod [1998] and references cited therein). This note observes that this estimation method is biased and inconsistent; that is, the estimated distribution does not in general converge on the true distribution as the size of the training corpus increases.",
"pdf_parse": {
"paper_id": "J02-1005",
"_pdf_hash": "",
"abstract": [
{
"text": "A data-oriented parsing or DOP model for statistical parsing associates fragments of linguistic representations with numerical weights, where these weights are estimated by normalizing the empirical frequency of each fragment in a training corpus (see Bod [1998] and references cited therein). This note observes that this estimation method is biased and inconsistent; that is, the estimated distribution does not in general converge on the true distribution as the size of the training corpus increases.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "The data-oriented parsing or DOP approach to statistical natural language analysis has attracted considerable attention recently and has been used to produce statistical language models based on various kinds of linguistic representation, as described in Bod (1998) . These models are based on the intuition that statistical generalizations about natural languages should be stated in terms of \"chunks\" or \"fragments\" of linguistic representations. Linguistic representations are produced by combining these fragments, but unlike in stochastic models such as Probabilistic Context-Free Grammars, a single linguistic representation may be generated by several different combinations of fragments. These fragments may be large, permitting DOP models to describe nonlocal dependencies. Usually the fragments used in a DOP model are themselves obtained from a training corpus of linguistic representations. For example, in DOP1 or Tree-DOP the fragments are typically all the connected multinode trees that appear as subgraphs of any tree in the training corpus.",
"cite_spans": [
{
"start": 255,
"end": 265,
"text": "Bod (1998)",
"ref_id": "BIBREF0"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1."
},
{
"text": "This note shows that the estimation procedure standardly used to set the parameters or fragment weights of a DOP model (see, for example, Bod [1998] ) is biased and inconsistent. This means that as sample size increases, the corresponding sequence of probability distributions estimated by this procedure does not converge to the true distribution that generated the training data. Consistency is usually regarded as the minimal requirement any estimation method must satisfy (Breiman 1973; Shao 1999) , and the inconsistency of the standard DOP estimation method suggests it may be worth looking for other estimation methods. Note that while the bulk of DOP research uses the estimation procedure studied here, recently there has been research that has used other estimators for DOP models (Bonnema, Buying, and Scha 1999; Bod 2000) , and it would be interesting to investigate the statistical properties of these estimators as well. ",
"cite_spans": [
{
"start": 138,
"end": 148,
"text": "Bod [1998]",
"ref_id": "BIBREF0"
},
{
"start": 476,
"end": 490,
"text": "(Breiman 1973;",
"ref_id": "BIBREF3"
},
{
"start": 491,
"end": 501,
"text": "Shao 1999)",
"ref_id": "BIBREF5"
},
{
"start": 791,
"end": 823,
"text": "(Bonnema, Buying, and Scha 1999;",
"ref_id": "BIBREF2"
},
{
"start": 824,
"end": 833,
"text": "Bod 2000)",
"ref_id": "BIBREF1"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1."
},
{
"text": "For simplicity, this note focuses on DOP1 or Tree-DOP models, in which linguistic representations are phrase structure trees, but the results carry over to more complex models that use attribute-value feature structure representations such as LFG-DOP. The fragments used in DOP1 are multinode trees whose leaves may be labeled with nonterminals as well as terminals. A derivation starts with a fragment whose root is labeled with the start symbol, and it proceeds by substituting a fragment for the leftmost nonterminal leaf under the constraint that the fragment's root node and the leaf node have the same label. The derivation terminates when there are no nonterminal leaves. Figure 1 depicts three different derivations that yield the same tree. The fragments used in these derivations could have been obtained from a training corpus of trees that contains trees for examples such as Sasha likes motorcycles, Alex eats pizza, and so on.",
"cite_spans": [],
"ref_spans": [
{
"start": 679,
"end": 687,
"text": "Figure 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "DOP1 Models",
"sec_num": "2."
},
{
"text": "In a DOP model, each fragment is associated with a real-valued weight, and the weight of a derivation is the product of the weights of the tree fragments involved. The weight of a representation is the sum of the weights of its derivations, and a probability distribution over linguistic representations is obtained by normalizing the representations' weights. 1 Given a combinatory operation and a fixed set of fragments, a DOP model is a parametric model where the fragment weights are the parameters.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "DOP1 Models",
"sec_num": "2."
},
{
"text": "In DOP1 and DOP models based on it, the weight associated with a fragment is estimated as follows (Bod 1998) . For each tree fragment f , let n(f ) be the number of times it appears in the training corpus, and let F be the set of all tree fragments with the same root as f . Then the weight w(f ) associated with f is",
"cite_spans": [
{
"start": 98,
"end": 108,
"text": "(Bod 1998)",
"ref_id": "BIBREF0"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "DOP1 Models",
"sec_num": "2."
},
{
"text": "w(f ) = n(f ) f \u2208F n(f ) .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "DOP1 Models",
"sec_num": "2."
},
{
"text": "This relative-frequency estimation method has the advantage of simplicity, but as shown in the following sections, it is biased and inconsistent.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "DOP1 Models",
"sec_num": "2."
},
{
"text": "Bias and inconsistency are usually defined for parametric estimation procedures in terms that are not quite appropriate for evaluating the DOP estimation procedure, but their standard definitions (see Shao [1999] for a textbook exposition) will serve as the basis for the definitions adopted below. Let \u0398 be a vector space of real-valued parameters, so that P \u03b8 , \u03b8 \u2208 \u0398 is a probability distribution. In the DOP1 case, \u0398 would be the space of all possible weight assignments to fragments. An estimator \u03c6 is a function from a vector x of n samples to a parameter value \u03c6(x) \u2208 \u0398, and an estimation procedure specifies an estimator \u03c6 n for each sample size n. Let X be a vector of n independent random variables distributed according to P \u03b8 for some \u03b8 \u2208 \u0398. Then \u03c6(X) is also a random variable, ranging over parameter vectors \u0398, with an expected value E \u03b8 (\u03c6(X)). The bias of the estimator \u03c6 at \u03b8 is the difference E \u03b8 (\u03c6(X)) \u2212 \u03b8 between its expected value and the \"true\" parameter value \u03b8 that determines the distribution X. A biased estimator is one with nonzero bias for some value of \u03b8 .",
"cite_spans": [
{
"start": 201,
"end": 212,
"text": "Shao [1999]",
"ref_id": "BIBREF5"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Bias and Inconsistency",
"sec_num": "3."
},
{
"text": "A loss function L is a function from pairs of parameter vectors to the nonnegative reals. Given a sample x drawn from the distribution \u03b8 , L(\u03b8 , \u03c6(x)) measures the \"cost\" or the \"loss\" incurred by the error in the estimate \u03c6(x) of \u03b8 . For example, a standard loss function is the Euclidean distance metric L(\u03b8 , \u03c6(x)) = \u03c6(X) \u2212 \u03b8 2 (note that the results below do not depend on this choice of loss function). The risk of an estimator \u03c6 at \u03b8 is its expected loss E \u03b8 (L(\u03b8 , \u03c6(X)). An estimation procedure is consistent if and only if the limit of the risk of \u03c6 n is 0 as n \u2192 \u221e for all \u03b8 . (There are various different notions of consistency depending on how convergence is defined; however, the DOP1 estimator is not consistent with respect to any of the standard definitions of consistency.)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Bias and Inconsistency",
"sec_num": "3."
},
{
"text": "Strictly speaking, the standard definitions of bias and loss function are not applicable to DOP estimation because there can be two distinct parameter vectors \u03b8 1 , \u03b8 2 for which P \u03b8 1 = P \u03b8 2 even though \u03b8 1 = \u03b8 2 (such a case is presented in the next section). Thus it is more natural to define bias and loss in terms of the probability distributions that the parameters specify, rather than in terms of the parameters themselves. In this paper, an estimator is unbiased iff P E \u03b8 (\u03c6(X)) = P \u03b8 for all \u03b8 ; that is, its expected parameter estimate specifies the same distribution as the true parameters. Similarly, the loss function is the mean squared difference between the \"true\" and estimated distributions; that is, if \u2126 is the event space (in DOP1, the space of all phrase structure trees), then",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Bias and Inconsistency",
"sec_num": "3."
},
{
"text": "L(\u03b8 , \u03c6(x)) = \u03c9\u2208\u2126 P \u03b8 (\u03c9)(P \u03b8 (\u03c9) \u2212 P \u03c6(x) (\u03c9)) 2 .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Bias and Inconsistency",
"sec_num": "3."
},
{
"text": "As before, the risk of an estimator is its expected loss, and an estimation procedure is consistent iff the limit of the expected loss is 0 as n \u2192 \u221e.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Bias and Inconsistency",
"sec_num": "3."
},
{
"text": "This section presents a simple DOP1 model that only generates two trees with probability p and 1 \u2212 p, respectively. The DOP relative frequency estimator is applied to a random sample of size n drawn from this population to estimate the tree weight parameters for the model. The bias and inconsistency of the estimator follow from the fact that these estimated parameters generate the trees with probabilities different from p and 1 \u2212 p. The trees used and their DOP1 fragments are shown in Figure 2.",
"cite_spans": [],
"ref_spans": [
{
"start": 490,
"end": 496,
"text": "Figure",
"ref_id": null
}
],
"eq_spans": [],
"section": "A DOP1 Example",
"sec_num": "4."
},
{
"text": "The trees t 1 , t 2 and their associated fragments f 1 , . . . , f 7 in the DOP1 model. Suppose the \"true\" weights for the fragments f 1 , . . . , f 7 are 0 except for the following fragments:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Figure 2",
"sec_num": null
},
{
"text": "w (f 4 ) = p, w (f 6 ) = 1 \u2212 p, w (f 7 ) = 1.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Figure 2",
"sec_num": null
},
{
"text": "Then P w (t 1 ) = p and P w (t 2 ) = 1 \u2212 p. (Note that exactly the same tree distribution could be obtained by setting w (f 1 ) = p and w (f 5 ) = 1 \u2212 p and all other weights to 0; thus the tree weights are not identifiable.) Then in a sample of size n drawn from the distribution P w the expected number of occurrences of tree t 1 is np and the expected number of occurrences of tree t 2 is n(1 \u2212 p). Thus the expected number of occurrences of the fragments in a sample of size n is E(n(f i )) = np for i = 1, . . . , 4; E(n(f i )) = n(1 \u2212 p) for i = 5, 6; E(n(f 7 )) = n + np.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Figure 2",
"sec_num": null
},
{
"text": "Thus after normalizing, the expected estimated weights for the fragments using the DOP estimator are",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Figure 2",
"sec_num": null
},
{
"text": "E(\u0175(f i )) = p 2 + 2p for i = 1, . . . , 4; E(\u0175(f i )) = 1 \u2212 p 2 + 2p for i = 5, 6;",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Figure 2",
"sec_num": null
},
{
"text": "E(\u0175(f 7 )) = 1.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Figure 2",
"sec_num": null
},
{
"text": "Further calculation shows that P E(\u0175) (t 1 ) = 2p 1 + p ,",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Figure 2",
"sec_num": null
},
{
"text": "The value of P E(\u0175) (t 1 ) as a function of Pw (t 1 ) = p. The identity function p is also plotted for comparison.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Figure 3",
"sec_num": null
},
{
"text": "P E(\u0175) (t 2 ) = 1 \u2212 p 1 + p .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Figure 3",
"sec_num": null
},
{
"text": "Figure 3 shows how P E(\u0175) (t 1 ) varies as a function of P w (t 1 ) = p. The difference P E(\u0175) (t 1 ) \u2212 p reaches a maximum value of approximately 0.17 at p = \u221a 2 \u2212 1. Thus except for p = 0 and p = 1, P E(\u0175) = P w ; that is, the DOP1 estimator is biased.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Figure 3",
"sec_num": null
},
{
"text": "Further, note that the estimated distribution P E(\u0175) does not approach P w as the sample size increases, so the expected loss does not converge to 0 as the sample size n increases. Thus the DOP1 estimator is also inconsistent.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Figure 3",
"sec_num": null
},
{
"text": "The previous section showed that the relative frequency estimation procedure used in DOP1 and related DOP models is biased and inconsistent. Bias is not necessarily a defect in an estimator, and Geman, Bienenstock, and Doursat (1992) argue that it may be desirable to trade variance for bias. However, inconsistency is usually viewed as a fatal flaw of an estimator. Nevertheless, excellent empirical results have been claimed for the DOP1 model, so perhaps there are some circumstances in which inconsistent estimators perform well. Undoubtedly there are other estimation procedures for DOP models that are unbiased and consistent. For example, maximum likelihood estimators are unbiased and consistent across a wide class of models, including, it would seem, all reasonable DOP models (Shao 1999) . Bod (2000) describes a procedure for maximum likelihood estimation of DOP models based on an Expectation Maximization-like algorithm. In addition, Rens Bod (personal communication) points out that because the set of fragments in a DOP1 model includes all of the trees in the training corpus, the maximum likelihood estimator will assign the training corpus trees their empirical frequencies, and assign 0 weight to all other trees. However, this seems to be an overlearning problem rather than a problem with maximum likelihood estimation per se, and standard methods, such as cross-validation or regularization, would seem in principle to be ways to avoid such overlearning. Obviously, empirical investigation would be useful here.",
"cite_spans": [
{
"start": 195,
"end": 233,
"text": "Geman, Bienenstock, and Doursat (1992)",
"ref_id": "BIBREF4"
},
{
"start": 787,
"end": 798,
"text": "(Shao 1999)",
"ref_id": "BIBREF5"
},
{
"start": 801,
"end": 811,
"text": "Bod (2000)",
"ref_id": "BIBREF1"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion",
"sec_num": "5."
},
{
"text": "In DOP1 and similar models, it is not necessary to normalize the representations' weights if the fragments' weights are themselves appropriately normalized.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
}
],
"back_matter": [
{
"text": "I would like to thank Rens Bod, Michael Collins, Eugene Charniak, David McAllester, and the anonymous reviewers for their excellent advice.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Acknowledgments",
"sec_num": null
}
],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "Beyond Grammar: An Experience-Based Theory of Language",
"authors": [
{
"first": "Rens",
"middle": [],
"last": "Bod",
"suffix": ""
}
],
"year": 1998,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Bod, Rens. 1998. Beyond Grammar: An Experience-Based Theory of Language. CSLI Publications, Stanford, CA.",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "Combining semantic and syntactic structure for language modelling",
"authors": [
{
"first": "Rens",
"middle": [],
"last": "Bod",
"suffix": ""
}
],
"year": 2000,
"venue": "Proceedings of the 8th International Conference on Spoken Language Processing",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Bod, Rens. 2000. Combining semantic and syntactic structure for language modelling. In Proceedings of the 8th International Conference on Spoken Language Processing (ICSLP 2000), Beijing.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "A new probability model for data oriented parsing",
"authors": [
{
"first": "Remko",
"middle": [],
"last": "Bonnema",
"suffix": ""
},
{
"first": "Paul",
"middle": [],
"last": "Buying",
"suffix": ""
},
{
"first": "Remko",
"middle": [],
"last": "Scha",
"suffix": ""
}
],
"year": 1999,
"venue": "Proceedings of the 12th Amsterdam Colloquium",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Bonnema, Remko, Paul Buying, and Remko Scha. 1999. A new probability model for data oriented parsing. In Proceedings of the 12th Amsterdam Colloquium, Amsterdam.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "Statistics with a View toward Applications",
"authors": [
{
"first": "Leo",
"middle": [],
"last": "Breiman",
"suffix": ""
}
],
"year": 1973,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Breiman, Leo. 1973. Statistics with a View toward Applications. Houghton Mifflin, Boston.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "Neural networks and the bias/variance dilemma",
"authors": [
{
"first": "Stuart",
"middle": [],
"last": "Geman",
"suffix": ""
},
{
"first": "Elie",
"middle": [],
"last": "Bienenstock",
"suffix": ""
},
{
"first": "Ren\u00e9",
"middle": [],
"last": "Doursat",
"suffix": ""
}
],
"year": 1992,
"venue": "Neural Computation",
"volume": "4",
"issue": "",
"pages": "1--58",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Geman, Stuart, Elie Bienenstock, and Ren\u00e9 Doursat. 1992. Neural networks and the bias/variance dilemma. Neural Computation, 4:1-58.",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "Mathematical Statistics",
"authors": [
{
"first": "",
"middle": [],
"last": "Shao",
"suffix": ""
}
],
"year": 1999,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Shao, Jun. 1999. Mathematical Statistics. Springer-Verlag, New York.",
"links": null
}
},
"ref_entries": {
"FIGREF0": {
"uris": null,
"text": "Depictions of three different derivations of the same tree representation of Alex likes pizza, with arrows indicating the sites of tree fragment substitutions.",
"type_str": "figure",
"num": null
}
}
}
}