| { |
| "paper_id": "S10-1009", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T15:28:02.368177Z" |
| }, |
| "title": "SemEval-2010 Task 12: Parser Evaluation using Textual Entailments", |
| "authors": [ |
| { |
| "first": "Deniz", |
| "middle": [], |
| "last": "Yuret", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Ko\u00e7 Universit\u1e8f Istanbul", |
| "location": { |
| "country": "Turkey" |
| } |
| }, |
| "email": "dyuret@ku.edu.tr" |
| }, |
| { |
| "first": "Ayd\u0131n", |
| "middle": [], |
| "last": "Han", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Ko\u00e7 Universit\u1e8f Istanbul", |
| "location": { |
| "country": "Turkey" |
| } |
| }, |
| "email": "ahan@ku.edu.tr" |
| }, |
| { |
| "first": "Zehra", |
| "middle": [], |
| "last": "Turgut", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Ko\u00e7 Universit\u1e8f Istanbul", |
| "location": { |
| "country": "Turkey" |
| } |
| }, |
| "email": "zturgut@ku.edu.tr" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "Parser Evaluation using Textual Entailments (PETE) is a shared task in the SemEval-2010 Evaluation Exercises on Semantic Evaluation. The task involves recognizing textual entailments based on syntactic information alone. PETE introduces a new parser evaluation scheme that is formalism independent, less prone to annotation error, and focused on semantically relevant distinctions.", |
| "pdf_parse": { |
| "paper_id": "S10-1009", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "Parser Evaluation using Textual Entailments (PETE) is a shared task in the SemEval-2010 Evaluation Exercises on Semantic Evaluation. The task involves recognizing textual entailments based on syntactic information alone. PETE introduces a new parser evaluation scheme that is formalism independent, less prone to annotation error, and focused on semantically relevant distinctions.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Parser Evaluation using Textual Entailments (PETE) is a shared task that involves recognizing textual entailments based on syntactic information alone. Given two text fragments called \"text\" and \"hypothesis\", textual entailment recognition is the task of determining whether the meaning of the hypothesis is entailed (can be inferred) from the text. In contrast with general RTE tasks (Dagan et al., 2009) the PETE task focuses on syntactic entailments:", |
| "cite_spans": [ |
| { |
| "start": 385, |
| "end": 405, |
| "text": "(Dagan et al., 2009)", |
| "ref_id": "BIBREF8" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Text: The man with the hat was tired. Hypothesis-1: The man was tired. (yes) Hypothesis-2: The hat was tired. (no) PETE is an evaluation scheme based on a natural human linguistic competence (i.e. the ability to comprehend sentences and answer simple yes/no questions about them). We believe systems should try to model natural human linguistic competence rather than their dubious competence in artificial tagging tasks.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The PARSEVAL measures introduced nearly two decades ago (Black et al., 1991) still dominate the field of parser evaluation. These methods compare phrase-structure bracketings produced by the parser with bracketings in the annotated corpus, or \"treebank\". Parser evaluation using short textual entailments has the following advantages compared to treebank based evaluation.", |
| "cite_spans": [ |
| { |
| "start": 56, |
| "end": 76, |
| "text": "(Black et al., 1991)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Consistency: Recognizing syntactic entailments is a more natural task for people than treebank annotation.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Focusing on a natural human competence makes it practical to collect high quality evaluation data from untrained annotators. The PETE dataset was annotated by untrained Amazon Mechanical Turk workers at an insignificant cost and each annotation is based on the unanimous agreement of at least three workers. In contrast, of the 36306 constituent strings that appear multiple times in the Penn Treebank (Marcus et al., 1994) , 5646 (15%) have multiple conflicting annotations. If indicative of the general level of inconsistency, 15% is a very high number given that the state of the art parsers claim f-scores above 90% (Charniak and Johnson, 2005) .", |
| "cite_spans": [ |
| { |
| "start": 402, |
| "end": 423, |
| "text": "(Marcus et al., 1994)", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 620, |
| "end": 648, |
| "text": "(Charniak and Johnson, 2005)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Relevance: PETE automatically focuses attention on semantically relevant phenomena rather than differences in annotation style or linguistic convention. Whether a phrase is tagged ADJP vs ADVP rarely affects semantic interpretation. Attaching the wrong subject to a verb or the wrong prepositional phrase to a noun changes the meaning of the sentence. Standard treebank based evaluation metrics do not distinguish between semantically relevant and irrelevant errors (Bonnema et al., 1997) . In PETE semantically relevant differences lead to different entailments, semantically irrelevant differences do not.", |
| "cite_spans": [ |
| { |
| "start": 466, |
| "end": 488, |
| "text": "(Bonnema et al., 1997)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Framework independence: Entailment recognition is a formalism independent task. A common evaluation method for parsers that do not use the Penn Treebank formalism is to automatically convert the Penn Treebank to the appropriate formalism and to perform treebank based evaluation (Nivre et al., 2007a; Hockenmaier and Steedman, 2007) . The inevitable conversion errors compound the already mentioned problems of treebank based evaluation. In addition, manually designed treebanks do not naturally lend themselves to unsupervised parser evaluation. Unlike treebank based evaluation, PETE can compare phrase structure parsers, dependency parsers, unsupervised parsers and other approaches on an equal footing.", |
| "cite_spans": [ |
| { |
| "start": 279, |
| "end": 300, |
| "text": "(Nivre et al., 2007a;", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 301, |
| "end": 332, |
| "text": "Hockenmaier and Steedman, 2007)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "PETE was inspired by earlier work on representations of grammatical dependency, proposed for ease of use by end users and suitable for parser evaluation. These include the grammatical relations (GR) by (Carroll et al., 1999) , the PARC representation (King et al., 2003) , and Stanford typed dependencies (SD) (De Marneffe et al., 2006 ) (See (Bos and others, 2008) for other proposals). Each use a set of binary relations between words in a sentence as the primary unit of representation. They share some common motivations: usability by people who are not (computational) linguists and suitability for relation extraction applications.", |
| "cite_spans": [ |
| { |
| "start": 202, |
| "end": 224, |
| "text": "(Carroll et al., 1999)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 251, |
| "end": 270, |
| "text": "(King et al., 2003)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 310, |
| "end": 335, |
| "text": "(De Marneffe et al., 2006", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 343, |
| "end": 365, |
| "text": "(Bos and others, 2008)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Here is an example sentence and its SD representation (De Marneffe and Manning, 2008) :", |
| "cite_spans": [ |
| { |
| "start": 54, |
| "end": 85, |
| "text": "(De Marneffe and Manning, 2008)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Bell, based in Los Angeles, makes and distributes electronic, computer and building products.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "nsubj(makes-8, Bell-1) nsubj(distributes-10, Bell-1) partmod(Bell-1, based-3) nn(Angeles-6, Los-5) prep-in(based-3, Angeles-6) conj-and(makes-8, distributes-10) amod(products-16, electronic-11) conj-and(electronic-11, computer-13) amod(products-16, computer-13) conj-and(electronic-11, building-15) amod(products-16, building-15) dobj(makes-8, products-16) PETE goes one step further by translating most of these dependencies into natural language entailments.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Bell makes something. Bell distributes something. Someone is based in Los Angeles. Someone makes products. PETE has some advantages over representations based on grammatical relations. For example SD defines 55 relations organized in a hierarchy, and it may be non-trivial for a non-linguist to understand the difference between ccomp (clausal complement with internal subject) and xcomp (clausal complement with external subject) or between nsubj (nominal subject) and xsubj (controlling subject). In fact it could be argued that proposals like SD replace one artificial annotation formalism with another and no two such proposals agree on the ideal set of binary relations to use. In contrast, untrained annotators have no difficulty unanimously agreeing on the validity of most PETE type entailments.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "However there are also significant challenges associated with an evaluation scheme like PETE. It is not always clear how to convert certain relations into grammatical hypothesis sentences without including most of the original sentence in the hypothesis. Including too much of the sentence in the hypothesis would increase the chances of getting the right answer with the wrong parse. Grammatical hypothesis sentences are especially difficult to construct when a (negative) entailment is based on a bad parse of the sentence. Introducing dummy words like \"someone\" or \"something\" alleviates part of the problem but does not help in the case of clausal complements. In summary, PETE makes the annotation phase more practical and consistent but shifts the difficulty to the entailment creation phase. PETE gets closer to an extrinsic evaluation by focusing on semantically relevant, application oriented differences that can be expressed in natural language sentences. This makes the evaluation procedure indirect: a parser developer has to write an extension that can handle entailment questions. However, given the simplicity of the entailments, the complexity of such an extension is comparable to one that extracts grammatical relations.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The balance of what is being evaluated is also important. A treebank based evaluation scheme may mix semantically relevant and irrelevant mistakes, but at least it covers every sentence at a uniform level of detail. In this evaluation, we focused on sentences and relations where state of the art parsers disagree. We hope this methodology will uncover weaknesses that the next generation systems can focus on.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The remaining sections will go into more detail about these challenges and the solutions we have chosen to implement. Section 2 explains the method followed to create the PETE dataset. Sec-tion 3 evaluates the baseline systems the task organizers created by implementing simple entailment extensions for several state of the art parsers. Section 4 presents the participating systems, their methods and results. Section 5 summarizes our contribution.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "To generate the entailments for the PETE task we followed the following three steps:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dataset", |
| "sec_num": "2" |
| }, |
| { |
| "text": "1. Identify syntactic dependencies that are challenging to state of the art parsers.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dataset", |
| "sec_num": "2" |
| }, |
| { |
| "text": "2. Construct short entailment sentences that paraphrase those dependencies.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dataset", |
| "sec_num": "2" |
| }, |
| { |
| "text": "3. Identify the subset of the entailments with high inter-annotator agreement.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dataset", |
| "sec_num": "2" |
| }, |
| { |
| "text": "To identify syntactic dependencies that are challenging for current state of the art parsers, we used example sentences from the following sources:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Identifying Challenging Dependencies", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "\u2022 The \"Unbounded Dependency Corpus\" (Rimell et al., 2009 ). An unbounded dependency construction contains a word or phrase which appears to have been moved, while being interpreted in the position of the resulting \"gap\". An unlimited number of clause boundaries may intervene between the moved element and the gap (hence \"unbounded\").", |
| "cite_spans": [ |
| { |
| "start": 36, |
| "end": 56, |
| "text": "(Rimell et al., 2009", |
| "ref_id": "BIBREF20" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Identifying Challenging Dependencies", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "\u2022 A list of sentences from the Penn Treebank on which the Charniak parser (Charniak and Johnson, 2005) performs poorly 1 .", |
| "cite_spans": [ |
| { |
| "start": 74, |
| "end": 102, |
| "text": "(Charniak and Johnson, 2005)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Identifying Challenging Dependencies", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "\u2022 The Brown section of the Penn Treebank.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Identifying Challenging Dependencies", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "We tested a number of parsers (both phrase structure and dependency) on these sentences and identified the differences in their output. We took sentences where at least one of the parsers gave a different answer than the others or the gold parse. Some of these differences reflected linguistic convention rather than semantic disagreement (e.g. representation of coordination) and some did not represent meaningful differences that can be expressed with entailments (e.g. labeling a phrase ADJP vs ADVP). The remaining differences typically reflected genuine semantic disagreements 1 http://www.cs.brown.edu/\u02dcec/papers/badPars.txt.gz that would effect downstream applications. These were chosen to turn into entailments in the next step.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Identifying Challenging Dependencies", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "We tried to make the entailments as targeted as possible by building them around two content words that are syntactically related. When the two content words were not sufficient to construct a grammatical sentence we used one of the following techniques:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Constructing Entailments", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "\u2022 Complete the mandatory elements using the words \"somebody\" or \"something\". (e.g. To test the subject-verb dependency in \"John kissed Mary.\" we construct the entailment \"John kissed somebody.\")", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Constructing Entailments", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "\u2022 Make a passive sentence to avoid using a spurious subject. (e.g. To test the verb-object dependency in \"John kissed Mary.\" we construct the entailment \"Mary was kissed.\")", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Constructing Entailments", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "\u2022 Make a copular sentence or use existential \"there\" to express noun modification. (e.g. To test the noun-modifier dependency in \"The big red boat sank.\" we construct the entailment \"The boat was big.\" or \"There was a big boat.\")", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Constructing Entailments", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "To identify the entailments that are clear to human judgement we used the following procedure:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Filtering Entailments", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "1. Each entailment was tagged by 5 untrained annotators from the Amazon Mechanical Turk crowdsourcing service.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Filtering Entailments", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "2. The results from the annotators whose agreement with the gold parse fell below 70% were eliminated.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Filtering Entailments", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "3. The entailments for which there was unanimous agreement of at least 3 annotators were kept.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Filtering Entailments", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "The instructions for the annotators were brief and targeted people with no linguistic background:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Filtering Entailments", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "Computers try to understand long sentences by dividing them into a set of short facts. You will help judge whether the computer extracted the right facts from a given set of 25 English sentences. Each of the following examples consists of a sentence (T), and a short statement (H) derived from this sentence by a computer. Please read both of them carefully and choose \"Yes\" if the meaning of (H) can be inferred from the meaning of (T). Here is an example: (T) Any lingering suspicion that this was a trick Al Budd had thought up was dispelled. (H) The suspicion was dispelled. Answer: YES (H) The suspicion was a trick. Answer: NO You can choose the third option \"Not sure\" when the (H) statement is unrelated, unclear, ungrammatical or confusing in any other manner.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Filtering Entailments", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "The \"Not sure\" answers were grouped with the \"No\" answers during evaluation. Approximately 50% of the original entailments were retained after the inter-annotator agreement filtering.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Filtering Entailments", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "The final dataset contained 367 entailments which were randomly divided into a 66 sentence development test and a 301 sentence test set. 52% of the entailments in the test set were positive.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dataset statistics", |
| "sec_num": "2.4" |
| }, |
| { |
| "text": "Approximately half of the final entailments were from the Unbounded Dependency Corpus, a third were from the Brown section of the Penn Treebank, and the remaining were from the Charniak sentences. ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dataset statistics", |
| "sec_num": "2.4" |
| }, |
| { |
| "text": "In order to establish baseline results for this task, we built an entailment decision system for CoNLL format dependency files and tested several publicly available parsers. The parsers used were the Berkeley Parser (Petrov and Klein, 2007) , Charniak Parser (Charniak and Johnson, 2005) , Collins Parser (Collins, 2003) , Malt Parser (Nivre et al., 2007b ), MSTParser (McDonald et al., 2005 and", |
| "cite_spans": [ |
| { |
| "start": 216, |
| "end": 240, |
| "text": "(Petrov and Klein, 2007)", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 259, |
| "end": 287, |
| "text": "(Charniak and Johnson, 2005)", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 305, |
| "end": 320, |
| "text": "(Collins, 2003)", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 335, |
| "end": 355, |
| "text": "(Nivre et al., 2007b", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 356, |
| "end": 391, |
| "text": "), MSTParser (McDonald et al., 2005", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baselines", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Stanford Parser (Klein and Manning, 2003) . Each parser was trained on sections 02-21 of the WSJ section of Penn Treebank. Outputs of phrase structure parsers were automatically annotated with function tags using Blaheta's function tagger (Blaheta and Charniak, 2000) and converted to the dependency structure with LTH Constituentto-Dependency Conversion Tool (Johansson and Nugues, 2007) .", |
| "cite_spans": [ |
| { |
| "start": 16, |
| "end": 41, |
| "text": "(Klein and Manning, 2003)", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 239, |
| "end": 267, |
| "text": "(Blaheta and Charniak, 2000)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 360, |
| "end": 388, |
| "text": "(Johansson and Nugues, 2007)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baselines", |
| "sec_num": "3" |
| }, |
| { |
| "text": "To decide the entailments both the test and hypothesis sentences were parsed. All the content words in the hypothesis sentence were determined by using part-of-speech tags and dependency relations. After applying some heuristics such as active-passive conversion, the extracted dependency path between the content words was searched in the dependency graph of the test sentence. In this search process, same relation types for the direct relations between the content word pairs and isomorphic subgraphs in the test and hypothesis sentences were required for the \"YES\" answer. Table 2 lists the baseline results achieved. There are significant differences in the entailment accuracies of systems that have comparable unlabeled attachment scores. One potential reason for this difference is the composition of the PETE dataset which emphasizes challenging syntactic constructions that some parsers may be better at. Another reason is the complete indifference of treebank based measures like UAS to the semantic significance of various dependencies and their impact on potential applications. ", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 577, |
| "end": 584, |
| "text": "Table 2", |
| "ref_id": "TABREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Baselines", |
| "sec_num": "3" |
| }, |
| { |
| "text": "There were 20 systems from 7 teams participating in the PETE task. Table 3 : Participating systems and their scores. The system identifier consists of the participant ID, system ID, and the system name given by the participant. Accuracy gives the percentage of correct entailments. Precision, Recall and F1 are calculated for positive entailments.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 67, |
| "end": 74, |
| "text": "Table 3", |
| "ref_id": "TABREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Systems", |
| "sec_num": "4" |
| }, |
| { |
| "text": "tems performed above the \"always yes\" baseline of 51.83%. Most systems started the entailment decision process by extracting syntactic dependencies, grammatical relations, or predicates by parsing the text and hypothesis sentences. Several submissions, including the top two scoring systems used the C&C Parser (Clark and Curran, 2007) which is based on Combinatory Categorical Grammar (CCG) formalism. Others used dependency structures produced by Malt Parser (Nivre et al., 2007b ), MSTParser (McDonald et al., 2005 and Stanford Parser (Klein and Manning, 2003) .", |
| "cite_spans": [ |
| { |
| "start": 311, |
| "end": 335, |
| "text": "(Clark and Curran, 2007)", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 461, |
| "end": 481, |
| "text": "(Nivre et al., 2007b", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 482, |
| "end": 517, |
| "text": "), MSTParser (McDonald et al., 2005", |
| "ref_id": null |
| }, |
| { |
| "start": 538, |
| "end": 563, |
| "text": "(Klein and Manning, 2003)", |
| "ref_id": "BIBREF14" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Systems", |
| "sec_num": "4" |
| }, |
| { |
| "text": "After the parsing step, the decision for the entailment was based on the comparison of relations, predicates, or dependency paths between the text and the hypothesis. Most systems relied on heuristic methods of comparison. A notable exception is the MARS-3 system which used an SVM-based classifier to decide on the entailment using dependency path features. Table 4 lists the frequency of various grammatical relations in the instances where the top system made mistakes. A comparison with Table 1 shows the direct objects and reduced relative clauses to be the frequent causes of error.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 359, |
| "end": 366, |
| "text": "Table 4", |
| "ref_id": null |
| }, |
| { |
| "start": 491, |
| "end": 498, |
| "text": "Table 1", |
| "ref_id": "TABREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Systems", |
| "sec_num": "4" |
| }, |
| { |
| "text": "We introduced PETE, a new method for parser evaluation using textual entailments. By basing the entailments on dependencies that current state GR Entailments Direct object 51% Reduced relative clause 36% Nominal subject 20% Object of preposition 7% Passive nominal subject 7% Table 4 : Frequency of grammatical relations in entailment instances that got wrong answers from the Cambridge system. of the art parsers disagree on, we hoped to create a dataset that would focus attention on the long tail of parsing problems that do not get sufficient attention using common evaluation metrics. By further restricting ourselves to differences that can be expressed by natural language entailments, we hoped to focus on semantically relevant decisions rather than accidents of convention which get mixed up in common evaluation metrics. We chose to rely on untrained annotators on a natural inference task rather than trained annotators on an artificial tagging task because we believe (i) many subfields of computational linguistics are struggling to make progress because of the noise in artificially tagged data, and (ii) systems should try to model natural human linguistic competence rather than their dubious competence in artificial tagging tasks. Our hope is datasets like PETE will be used not only for evaluation but also for training and fine-tuning of systems in the future. Further work is needed to automate the entailment generation process and to balance the composition of syntactic phenomena covered in a PETE dataset.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 276, |
| "end": 283, |
| "text": "Table 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Contributions", |
| "sec_num": "5" |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "We would like to thank Laura Rimell, Stephan Oepen and Anna Mac for their careful analysis and valuable suggestions.\u00d6nder Eker contributed to the early development of the PETE task.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgments", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "A Procedure for Quantitatively Comparing the Syntactic Coverage of English Grammars", |
| "authors": [ |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Black", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Abney", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Flickenger", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Gdaniec", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Grishman", |
| "suffix": "" |
| }, |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Harrison", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Hindle", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Ingria", |
| "suffix": "" |
| }, |
| { |
| "first": "F", |
| "middle": [], |
| "last": "Jelinek", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Klavans", |
| "suffix": "" |
| } |
| ], |
| "year": 1991, |
| "venue": "Speech and natural language: proceedings of a workshop", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "E. Black, S. Abney, D. Flickenger, C. Gdaniec, R. Gr- ishman, P. Harrison, D. Hindle, R. Ingria, F. Jelinek, J. Klavans, et al. 1991. A Procedure for Quanti- tatively Comparing the Syntactic Coverage of En- glish Grammars. In Speech and natural language: proceedings of a workshop, held at Pacific Grove, California, February 19-22, 1991, page 306. Mor- gan Kaufmann Pub.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Assigning function tags to parsed text", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Blaheta", |
| "suffix": "" |
| }, |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Charniak", |
| "suffix": "" |
| } |
| ], |
| "year": 2000, |
| "venue": "Proceedings of the 1st North American chapter of the Association for Computational Linguistics conference", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "D. Blaheta and E. Charniak. 2000. Assigning func- tion tags to parsed text. In Proceedings of the 1st North American chapter of the Association for Com- putational Linguistics conference, page 240. Mor- gan Kaufmann Publishers Inc.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "A DOP model for semantic interpretation", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Bonnema", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Bod", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Scha", |
| "suffix": "" |
| } |
| ], |
| "year": 1997, |
| "venue": "Proceedings of the eighth conference on European chapter of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "159--167", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "R. Bonnema, R. Bod, and R. Scha. 1997. A DOP model for semantic interpretation. In Proceedings of the eighth conference on European chapter of the Association for Computational Linguistics, pages 159-167. Association for Computational Linguis- tics.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Proceedings of the Workshop on Cross-Framework and Cross-Domain Parser Evaluation. In connection with the 22nd International Conference on Computational Linguistics", |
| "authors": [ |
| { |
| "first": "Johan", |
| "middle": [], |
| "last": "Bos", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Johan Bos et al., editors. 2008. Proceedings of the Workshop on Cross-Framework and Cross-Domain Parser Evaluation. In connection with the 22nd In- ternational Conference on Computational Linguis- tics.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Corpus annotation for parser evaluation", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Carroll", |
| "suffix": "" |
| }, |
| { |
| "first": "G", |
| "middle": [], |
| "last": "Minnen", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Briscoe", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "Proceedings of the EACL workshop on Linguistically Interpreted Corpora (LINC)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "J. Carroll, G. Minnen, and T. Briscoe. 1999. Cor- pus annotation for parser evaluation. In Proceedings of the EACL workshop on Linguistically Interpreted Corpora (LINC).", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Coarse-to-fine nbest parsing and MaxEnt discriminative reranking", |
| "authors": [ |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Charniak", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Johnson", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Proceedings of the 43rd Annual Meeting on Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "E. Charniak and M. Johnson. 2005. Coarse-to-fine n- best parsing and MaxEnt discriminative reranking. In Proceedings of the 43rd Annual Meeting on As- sociation for Computational Linguistics, page 180. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Wide-coverage efficient statistical parsing with CCG and log-linear models", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Clark", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [ |
| "R" |
| ], |
| "last": "Curran", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Computational Linguistics", |
| "volume": "33", |
| "issue": "4", |
| "pages": "493--552", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "S. Clark and J.R. Curran. 2007. Wide-coverage ef- ficient statistical parsing with CCG and log-linear models. Computational Linguistics, 33(4):493-552.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Head-driven statistical models for natural language parsing", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Collins", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "", |
| "volume": "29", |
| "issue": "", |
| "pages": "589--637", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "M. Collins. 2003. Head-driven statistical models for natural language parsing. Computational linguis- tics, 29(4):589-637.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Recognizing textual entailment: Rational, evaluation and approaches", |
| "authors": [ |
| { |
| "first": "I", |
| "middle": [], |
| "last": "Dagan", |
| "suffix": "" |
| }, |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Dolan", |
| "suffix": "" |
| }, |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Magnini", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Roth", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Natural Language Engineering", |
| "volume": "15", |
| "issue": "04", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "I. Dagan, B. Dolan, B. Magnini, and D. Roth. 2009. Recognizing textual entailment: Rational, evalua- tion and approaches. Natural Language Engineer- ing, 15(04).", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Generating typed dependency parses from phrase structure parses", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [ |
| "C" |
| ], |
| "last": "De Marneffe", |
| "suffix": "" |
| }, |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Maccartney", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [ |
| "D" |
| ], |
| "last": "Manning", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "LREC", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "M.C. De Marneffe, B. MacCartney, and C.D. Manning. 2006. Generating typed dependency parses from phrase structure parses. In LREC 2006.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "CCGbank: a corpus of CCG derivations and dependency structures extracted from the Penn Treebank", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Hockenmaier", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Steedman", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Computational Linguistics", |
| "volume": "33", |
| "issue": "3", |
| "pages": "355--396", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "J. Hockenmaier and M. Steedman. 2007. CCGbank: a corpus of CCG derivations and dependency struc- tures extracted from the Penn Treebank. Computa- tional Linguistics, 33(3):355-396.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Extended constituent-to-dependency conversion for English", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Johansson", |
| "suffix": "" |
| }, |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Nugues", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proc. of the 16th Nordic Conference on Computational Linguistics (NODALIDA)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "R. Johansson and P. Nugues. 2007. Extended constituent-to-dependency conversion for English. In Proc. of the 16th Nordic Conference on Compu- tational Linguistics (NODALIDA).", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "The PARC 700 dependency bank", |
| "authors": [ |
| { |
| "first": "T", |
| "middle": [ |
| "H" |
| ], |
| "last": "King", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Crouch", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Riezler", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Dalrymple", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Kaplan", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Proceedings of the EACL03: 4th International Workshop on Linguistically Interpreted Corpora (LINC-03)", |
| "volume": "", |
| "issue": "", |
| "pages": "1--8", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "T.H. King, R. Crouch, S. Riezler, M. Dalrymple, and R. Kaplan. 2003. The PARC 700 dependency bank. In Proceedings of the EACL03: 4th Interna- tional Workshop on Linguistically Interpreted Cor- pora (LINC-03), pages 1-8.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Accurate unlexicalized parsing", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Klein", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [ |
| "D" |
| ], |
| "last": "Manning", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Proceedings of the 41st Annual Meeting on Association for Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "423--430", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "D. Klein and C.D. Manning. 2003. Accurate un- lexicalized parsing. In Proceedings of the 41st Annual Meeting on Association for Computational Linguistics-Volume 1, pages 423-430. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Building a large annotated corpus of English: The Penn Treebank", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [ |
| "P" |
| ], |
| "last": "Marcus", |
| "suffix": "" |
| }, |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Santorini", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [ |
| "A" |
| ], |
| "last": "Marcinkiewicz", |
| "suffix": "" |
| } |
| ], |
| "year": 1994, |
| "venue": "", |
| "volume": "19", |
| "issue": "", |
| "pages": "313--330", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "M.P. Marcus, B. Santorini, and M.A. Marcinkiewicz. 1994. Building a large annotated corpus of En- glish: The Penn Treebank. Computational linguis- tics, 19(2):313-330.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Non-projective dependency parsing using spanning tree algorithms", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Mcdonald", |
| "suffix": "" |
| }, |
| { |
| "first": "F", |
| "middle": [], |
| "last": "Pereira", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Ribarov", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Hajic", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Proceedings of HLT/EMNLP", |
| "volume": "", |
| "issue": "", |
| "pages": "523--530", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "R. McDonald, F. Pereira, K. Ribarov, and J. Ha- jic. 2005. Non-projective dependency parsing us- ing spanning tree algorithms. In Proceedings of HLT/EMNLP, pages 523-530.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "The CoNLL 2007 shared task on dependency parsing", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Nivre", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Hall", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "K\u00fcbler", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Mcdonald", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Nilsson", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Riedel", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Yuret", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proceedings of the CoNLL Shared Task Session of EMNLP-CoNLL", |
| "volume": "7", |
| "issue": "", |
| "pages": "915--932", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "J. Nivre, J. Hall, S. K\u00fcbler, R. McDonald, J. Nilsson, S. Riedel, and D. Yuret. 2007a. The CoNLL 2007 shared task on dependency parsing. In Proceed- ings of the CoNLL Shared Task Session of EMNLP- CoNLL, volume 7, pages 915-932.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Malt-Parser: A language-independent system for datadriven dependency parsing", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Nivre", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Hall", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Nilsson", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Chanev", |
| "suffix": "" |
| }, |
| { |
| "first": "G", |
| "middle": [], |
| "last": "Eryigit", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "K\u00fcbler", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Marinov", |
| "suffix": "" |
| }, |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Marsi", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Natural Language Engineering", |
| "volume": "13", |
| "issue": "02", |
| "pages": "95--135", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "J. Nivre, J. Hall, J. Nilsson, A. Chanev, G. Eryigit, S. K\u00fcbler, S. Marinov, and E. Marsi. 2007b. Malt- Parser: A language-independent system for data- driven dependency parsing. Natural Language En- gineering, 13(02):95-135.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Improved inference for unlexicalized parsing", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Petrov", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Klein", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proceedings of NAACL HLT 2007", |
| "volume": "", |
| "issue": "", |
| "pages": "404--411", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "S. Petrov and D. Klein. 2007. Improved inference for unlexicalized parsing. In Proceedings of NAACL HLT 2007, pages 404-411.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "Unbounded dependency recovery for parser evaluation", |
| "authors": [ |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Rimell", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Clark", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Steedman", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "813--821", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "L. Rimell, S. Clark, and M. Steedman. 2009. Un- bounded dependency recovery for parser evaluation. In Proceedings of the 2009 Conference on Empiri- cal Methods in Natural Language Processing, pages 813-821. Association for Computational Linguis- tics.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "TABREF0": { |
| "type_str": "table", |
| "text": "lists the most frequent grammatical relations encountered in the entailments.", |
| "num": null, |
| "content": "<table><tr><td>GR</td><td>Entailments</td></tr><tr><td>Direct object</td><td>42%</td></tr><tr><td>Nominal subject</td><td>33%</td></tr><tr><td colspan=\"2\">Reduced relative clause 21%</td></tr><tr><td>Relative clause</td><td>13%</td></tr><tr><td colspan=\"2\">Passive nominal subject 6%</td></tr><tr><td>Object of preposition</td><td>5%</td></tr><tr><td>Prepositional modifier</td><td>4%</td></tr><tr><td>Conjunct</td><td>2%</td></tr><tr><td>Adverbial modifier</td><td>2%</td></tr><tr><td>Free relative</td><td>2%</td></tr></table>", |
| "html": null |
| }, |
| "TABREF1": { |
| "type_str": "table", |
| "text": "Most frequent grammatical relations encountered in the entailments.", |
| "num": null, |
| "content": "<table/>", |
| "html": null |
| }, |
| "TABREF3": { |
| "type_str": "table", |
| "text": "Baseline systems: The second column gives the performance on the PETE test set, the third column gives the unlabeled attachment score on section 23 of the Penn Treebank.", |
| "num": null, |
| "content": "<table/>", |
| "html": null |
| }, |
| "TABREF4": { |
| "type_str": "table", |
| "text": "", |
| "num": null, |
| "content": "<table><tr><td>gives the percent-</td></tr></table>", |
| "html": null |
| } |
| } |
| } |
| } |