ACL-OCL / Base_JSON /prefixW /json /W13 /W13-0121.json
Benjamin Aw
Add updated pkl file v3
6fa4bc9
{
"paper_id": "W13-0121",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T04:57:34.719753Z"
},
"title": "Semantic Annotation of Textual Entailment",
"authors": [
{
"first": "Assaf",
"middle": [],
"last": "Toledo",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "Utrecht University",
"location": {}
},
"email": "a.toledo@uu.nl"
},
{
"first": "Sophia",
"middle": [],
"last": "Katrenko",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "Utrecht University",
"location": {}
},
"email": "sophia@katrenko.com"
},
{
"first": "Heidi",
"middle": [],
"last": "Klockmann",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "Utrecht University",
"location": {}
},
"email": "h.e.klockmann@uu.nl"
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "We introduce a new formal semantic model for annotating textual entailments, that describes restrictive, intersective and appositive modification. The model contains a formally defined interpreted lexicon, which specifies the inventory of symbols and the supported semantic operators, and an informally defined annotation scheme that instructs annotators in which way to bind words and constructions from a given pair of premise and hypothesis to the interpreted lexicon. We explore the applicability of the proposed model to the Recognizing Textual Entailment (RTE) 1-4 corpora and describe a first-stage annotation scheme based on which manual annotation work was carried out. The constructions we annotated were found to occur in 80.65% of the entailments in RTE 1-4 and were annotated with cross-annotator agreement of 68% on average. The annotated RTE corpora are publicly available for the research community. 1 Pairs of sentences in RTE 1-3 are categorized in two classes: yesor no-entailment; pairs in RTE 4-5 are categorized in three classes: entailment, contradiction and unknown. We label the judgments yes-entailment from RTE 1-3 and entailment from RTE 4-5 as valid, and the other judgments as invalid. 2 Pair 410 from the test set of RTE 2. EC stands for European Commission",
"pdf_parse": {
"paper_id": "W13-0121",
"_pdf_hash": "",
"abstract": [
{
"text": "We introduce a new formal semantic model for annotating textual entailments, that describes restrictive, intersective and appositive modification. The model contains a formally defined interpreted lexicon, which specifies the inventory of symbols and the supported semantic operators, and an informally defined annotation scheme that instructs annotators in which way to bind words and constructions from a given pair of premise and hypothesis to the interpreted lexicon. We explore the applicability of the proposed model to the Recognizing Textual Entailment (RTE) 1-4 corpora and describe a first-stage annotation scheme based on which manual annotation work was carried out. The constructions we annotated were found to occur in 80.65% of the entailments in RTE 1-4 and were annotated with cross-annotator agreement of 68% on average. The annotated RTE corpora are publicly available for the research community. 1 Pairs of sentences in RTE 1-3 are categorized in two classes: yesor no-entailment; pairs in RTE 4-5 are categorized in three classes: entailment, contradiction and unknown. We label the judgments yes-entailment from RTE 1-3 and entailment from RTE 4-5 as valid, and the other judgments as invalid. 2 Pair 410 from the test set of RTE 2. EC stands for European Commission",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "The RTE challenges (Dagan et al., 2006) aim to automatically determine whether an entailment relation obtains between a naturally occurring text sentence (T) and a hypothesis sentence (H). The RTE corpus (Bar Haim et al., 2006; Giampiccolo et al., 2007 Giampiccolo et al., , 2008 Bentivogli et al., 2009) , which is currently the only available resource of textual entailments, marks entailment candidates as valid/invalid. 1 Example 1",
"cite_spans": [
{
"start": 19,
"end": 39,
"text": "(Dagan et al., 2006)",
"ref_id": "BIBREF6"
},
{
"start": 204,
"end": 227,
"text": "(Bar Haim et al., 2006;",
"ref_id": "BIBREF0"
},
{
"start": 228,
"end": 252,
"text": "Giampiccolo et al., 2007",
"ref_id": "BIBREF9"
},
{
"start": 253,
"end": 279,
"text": "Giampiccolo et al., , 2008",
"ref_id": "BIBREF8"
},
{
"start": 280,
"end": 304,
"text": "Bentivogli et al., 2009)",
"ref_id": "BIBREF1"
},
{
"start": 424,
"end": 425,
"text": "1",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "2"
},
{
"text": "\u2022 T: The head of the Italian opposition, Romano Prodi, was the last president of the EC.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "2"
},
{
"text": "\u2022 H: Romano Prodi is a former president of the EC. 2",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "2"
},
{
"text": "\u2022 Entailment: Valid",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "2"
},
{
"text": "This categorization contains no indication of the linguistic processes that underlie entailment. In the lack of a gold standard of inferential phenomena, entailment systems can be compared based on their performance, but their inferential processes are not directly accessible for analysis. The goal of this work is to elucidate some central inferential processes underlying entailments in the RTE corpus. By doing that, we aim to advance the possibility of creating a benchmark for modeling entailment recognition. We presume that this goal is to be achieved incrementally by modeling increasingly complex semantic phenomena. To this end, we employ a standard model-theoretic approach to entailment in order to combine gold standard annotations with a computational framework. The model contains a formally defined interpreted lexicon, which specifies the inventory of symbols and semantic operators that are supported, and an informally defined annotation scheme that instructs annotators how to bind words and constructions from a given T-H pair to entries in the interpreted lexicon. Our choice to focus on the semantic phenomena of restrictive, intersective and appositive modification is driven by their predominance in the RTE datasets, the ability to annotate them with high consistency and the possibility to capture their various syntactic expressions by a limited set of concepts.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "2"
},
{
"text": "However, currently we are only at the first stages of implementing the theoretical semantic model using an annotation platform combined with a theorem prover. In the course of the development of this model we adopted a narrower annotation scheme by which modification phenomena were annotated in all valid entailment pairs from RTE 1-4 without accounting for the way in which the annotated phenomena contribute to the inference being made. This work allowed us to perform data analysis and to further learn about the phenomena of interest as part of the development of the semantic model.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "2"
},
{
"text": "The structure of this paper is as follows. Section 3 reviews some related methods used in Bos et al. (2004) and MacCartney and Manning (2007) . In Section 4 we introduce the formal semantic model on which we rely and use it for analyzing some illustrative textual entailments. Section 5 points out a challenge in applying this model to parts of the RTE data and describes our first-stage annotation scheme. We elaborate on the methods employed in applying this scheme to the datasets of RTE 1-4, and present some quantitative data on the targeted phenomena and inter-annotator agreement. Section 6 concludes.",
"cite_spans": [
{
"start": 90,
"end": 107,
"text": "Bos et al. (2004)",
"ref_id": "BIBREF2"
},
{
"start": 112,
"end": 141,
"text": "MacCartney and Manning (2007)",
"ref_id": "BIBREF11"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "2"
},
{
"text": "Bos and Markert (2005) utilizes a CCG parser (Bos et al., 2004) to represent the text and hypothesis in discourse representation structures (DRSs, Kamp and Reyle 1993) that encapsulate information on argument structure, polarity, etc. The DRSs of the text and hypothesis are then translated into formulae in first order logic, and a theorem prover is used in order to search whether there is a logical proof from the text formula to the hypothesis formula. The system reached a relatively high precision score of 76% in recognizing the positive cases in RTE 2 but suffered from a very low recall of 5.8%.",
"cite_spans": [
{
"start": 45,
"end": 63,
"text": "(Bos et al., 2004)",
"ref_id": "BIBREF2"
},
{
"start": 147,
"end": 167,
"text": "Kamp and Reyle 1993)",
"ref_id": "BIBREF10"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "3"
},
{
"text": "MacCartney and Manning (2007)'s system recognizes monotonic relations (or lack thereof) between aligned lexical items in the text and hypothesis and employs a model of compositional semantics to calculate a sentence-level entailment prediction. The recognition of monotonic relations is done using an adapted version of Sanchez Valencia's Natural Logic (Valencia, 1991) , the alignment between the text and hypothesis is based on a cost function that extends Levenshtein string-edit algorithm, and the entailment is classified by a decision tree classifier, trained on a small data set of 69 handmade problems. The system was tested on RTE 3 and achieved relatively high precision scores of 76.39% and 68.06% on the positive cases in the development and test sets respectively. This system also suffers from low recall scores of 26.70% and 31.71% respectively.",
"cite_spans": [
{
"start": 353,
"end": 369,
"text": "(Valencia, 1991)",
"ref_id": "BIBREF15"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "3"
},
{
"text": "The model we propose in this work diverges from these systems in two respects: (a) its first goal is to develop gold standard semantic annotations based on a general formal semantic model; (b) it does not aim to represent phenomena that are not accounted for in this model. For example, consider the following inference, which is based on causal reasoning: Khan sold nuclear plans \u21d2 Khan possessed nuclear plans. 3 Causal reasoning and lexical relations are not part of the semantic phenomena addressed in this paper, and a pattern in the form of X sold Y \u21d2 X possessed Y should be defined ad-hoc by annotators to align the instances of the verbs sell and possess. This approach allows us to concentrate on the logical aspects of textual entailment, while phenomena involving lexical semantics and world knowledge are handled by a shallow analysis. 4 To model entailment in natural language, we assume that entailment describes a preorder on natural language sentences. Thus, we assume that any sentence trivially entails itself (reflexivity); and given two entailments T 1 \u21d2 H 1 and T 2 \u21d2 H 2 where H 1 and T 2 are identical sentences, we assume T 1 \u21d2 H 2 (transitivity). A computational theory of entailment should describe an approximation of this preorder on natural language sentences. We use a standard model-theoretical extensional semantics, based on the simple partial order on the domain of truth-values. Each model M assigns sentences a truth-value in the set {0, 1}. Such a Tarskian theory of entailment is considered adequate if the intuitive entailment preorder on sentences can be described as the pairs of sentences T and H whose truth-values",
"cite_spans": [
{
"start": 849,
"end": 850,
"text": "4",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "3"
},
{
"text": "[[T]] M and [[H]] M satisfy [[T]] M \u2264 [[H]] M for all models M .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "3"
},
{
"text": "In this section we give the essentials of this modeltheoretic approach to entailment that are relevant to the annotated phenomena and illustrate it using a small interpreted lexicon, simplifying the analysis of some representative examples from the RTE.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "3"
},
{
"text": "The interpreted lexicon presented in Table 1 illustrates our treatment of major lexical categories over types e, t and their functional compounds. Our aim is to allow binding of words and expressions in entailment data to the lexicon. Each word is stated in its literal form, the type assigned to it, and its denotation in intended models. Denotations that are assumed to be arbitrary in intended models are given in boldface. For example, the intransitive use of the verb sit is assigned the type et and its denotation sit is an arbitrary function of this type. By contrast, other lexical items have their denotations restricted by the intended models. For example, the definite article the is assigned the type (et)e and its denotation is fixed as the iota operator. The functions that we use for defining denotations are specified in Figure 1 . Several items in the lexicon are assigned more than one type and/or more than one denotation due to ambiguity in natural language. The following list explains some of the main items in the lexicon:",
"cite_spans": [],
"ref_spans": [
{
"start": 37,
"end": 44,
"text": "Table 1",
"ref_id": "TABREF1"
},
{
"start": 837,
"end": 846,
"text": "Figure 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "An interpreted lexicon",
"sec_num": "4.1"
},
{
"text": "\u2022 The coordinator and, when appearing as predicate conjunction, is analyzed as a function -AND, mapping any two et predicates A and B to a predicate that sends every entity e to the truth-value of the conjunction A(x) \u2227 B(x). \u2022 The copular is and the article a in copular sentences (e.g. Dan is a man / Dan is short) are analyzed as identity functions IS and A of type (et)(et) respectively. In copula sentences that express an equality relation (e.g. Dan is Jan), is is analyzed by the equality function IS eq of type e(et). \u2022 The word some denotes the existential quantifier SOME, as it is used in intransitive sentences such as some man sat (transitive sentences like Jan saw some man are not treated here). \u2022 The relative pronoun who allows noun modification either by a restrictive relative clause denoted by WHO R or by an appositive clause denoted by WHO A . WHO R is expressed in sentences such as the alien who is a nun sat, in which the pronoun creates a complex predicate, alien who is a nun. WHO A appears in sentences such as the alien, who is a nun, sat where the pronoun adds information on a given entity x. The resulting entity is x if A holds of x, and undefined otherwise. \u2022 The adjectives short and Dutch, when appearing as modifiers, restrict the denotation of the noun they attach to: a short/Dutch man is a man. Dutch is furthermore intersective: a Dutch man is invariably Dutch. The predicate Dutch is defined as an arbitrary constant dutch of type et. The modifier is derived by a function I m identical to AND. The restrictive modifier short is defined by the function R m and a constant short of type (et)(et). The predicative denotation of short is defined using the function P r as the set of \"short things\" -by applying the constant to all entities. See Pratt-Hartmann and Moss (2009) for a wider coverage of some of the same semantic ground that goes further in dealing with comparative constructions and transitive verbs.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "An interpreted lexicon",
"sec_num": "4.1"
},
{
"text": "Some central logical semantic aspects of entailments from the RTE can be formally analyzed using the lexicon in Table 1 . We analyze entailments by binding expressions in the RTE data to structurally equivalent expressions containing items in the interpreted lexicon. This analysis is three-fold: 1. Phenomena Simplification: we simplify the text and hypothesis to exclude inferential phenomena that we do not handle in the scope of this work. For instance, in Example 2, the inference Google operates on the web \u21d2 Google is on the web is based on lexical knowledge, which we do not address here, and therefore it is handled as part of the simplification step. 2. Binding to Lexicon: we bind the constructions in the data to parallel constructions in the interpreted lexicon that share the same structure and semantic properties. This step produces a text T Lexicon and a hypothesis H Lexicon as new structurally equivalent versions of the simplified text and hypothesis. The parse trees are assumed in a way that allows to apply the interpreted lexicon. 3. Proof of Entailment: using predicate calculus and lambda calculus reductions, we establish a logical proof between T Lexicon and H Lexicon . 5 Example 2",
"cite_spans": [],
"ref_spans": [
{
"start": 109,
"end": 119,
"text": "in Table 1",
"ref_id": "TABREF1"
}
],
"eq_spans": [],
"section": "Analyzing entailments using the interpreted lexicon",
"sec_num": "4.2"
},
{
"text": "\u2022 Data: -T: The largest search engine on the web, Google receives over 200 million queries each day through its various services.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Analyzing entailments using the interpreted lexicon",
"sec_num": "4.2"
},
{
"text": "-H: Google operates on the web. 6 1. Phenomena Simplification:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Analyzing entailments using the interpreted lexicon",
"sec_num": "4.2"
},
{
"text": "In the text: adding an overt appositive WH pronoun to match the interpreted lexicon:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Analyzing entailments using the interpreted lexicon",
"sec_num": "4.2"
},
{
"text": "\u2022 T Original : The largest search engine on the web, Google receives...",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Analyzing entailments using the interpreted lexicon",
"sec_num": "4.2"
},
{
"text": "\u2022 T Simple : The largest search engine on the web, which is Google, receives... In the hypothesis: reducing the meaning of 'X operates on Y' to 'X is on Y':",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Analyzing entailments using the interpreted lexicon",
"sec_num": "4.2"
},
{
"text": "\u2022 ",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Analyzing entailments using the interpreted lexicon",
"sec_num": "4.2"
},
{
"text": "= (see(dan))((WHO A (IS eq (jan)))(\u03b9((R m (short)) ((I m (dutch))(man))))) analysis = (see(dan))(\u03b9(\u03bby.y = (\u03b9((R m (short))((I m (dutch))(man)))) \u2227jan = (\u03b9((R m (short))((I m (dutch))(man)))))) def. of WHO A + IS eq , func application By definition of \u03b9: jan = \u03b9((R m (short))((I m (dutch))(man))) \u21d2 jan = \u03b9(\u03bby e .(short(\u03bbx e .man(x) \u2227 dutch(x)))(y)\u2227 man(y) \u2227 dutch(y)) def. of R m , I m , \u2227 + func. application By definition of \u03b9: (short(\u03bbx e .man(x) \u2227 dutch(x)))(jan) \u2227 man(jan) \u2227 dutch(jan) \u2264 dutch(jan) = (IS(dutch))(jan) = [[Jan [is Dutch] ]] M",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Analyzing entailments using the interpreted lexicon",
"sec_num": "4.2"
},
{
"text": "def. of \u2227, IS + analysis A crucial step in this analysis is our assumption that on the web is an intersective modifier of search engine. This allows the subsumption of search engine on the web by on the web. In the interpreted lexicon we describe this behavior using the intersective denotation of the modifier Dutch. Let us investigate further the implications of this annotation in the following hypothetical example. Example 3",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Analyzing entailments using the interpreted lexicon",
"sec_num": "4.2"
},
{
"text": "1. Pair 1: T 1 : Jan is a short Dutch man \u21d2 H 1 : Jan is a short man no entailment 2. Pair 2: T 2 : Jan is a black Dutch man \u21d2 H 2 : Jan is a black man entailment From a purely textual/syntactic point of view, these two T-H pairs are indistinguishable. The lexical overlap between the text and hypothesis in both pairs is 100%. This does not allow entailment systems to rely on textual measurements to identify that the pairs need to be classified differently. Such a perfect score of overlap may lead to a false positive classification in Pair 1 or conversely, to a false negative in Pair 2. Also syntactically, both short and black serve as adjectives attached to a noun phrase -Dutch man. There is nothing in this syntactic configuration to suggest that omitting Dutch in Pair 1 might result in a different entailment classification than omitting it in Pair 2. However, from a semantic point of view, based on annotations of abstract relations between predicates and their modifiers, we can correctly analyze both the non-validity of the entailment in Pair 1 and the validity of the entailment in Pair 2.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Analyzing entailments using the interpreted lexicon",
"sec_num": "4.2"
},
{
"text": "\u2022 Analysis of Pair 1 To validate that there is no entailment between a text and a hypothesis means to show that there is an intended model M = E, I in which there is no \u2264 relation between their denotations.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Analyzing entailments using the interpreted lexicon",
"sec_num": "4.2"
},
{
"text": "Let M be an intended model that satisfies the following: ",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Analyzing entailments using the interpreted lexicon",
"sec_num": "4.2"
},
{
"text": "-man et characterizes {dan,",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Analyzing entailments using the interpreted lexicon",
"sec_num": "4.2"
},
{
"text": "[[ Jan [is [a [short man]]] ]] M = (IS(A((R m (short))(man))))(jan) analysis = ((R m (short))(man))(jan) def. of A, IS = (((\u03bbM (et)(et) .\u03bbA et .\u03bby e .M (A)(y) \u2227A(y))(short))(man))(jan) def. of R m = 0 \u2227 1 = 0",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Analyzing entailments using the interpreted lexicon",
"sec_num": "4.2"
},
{
"text": "func. application + denotations in M Intuitively, Jan can be a man who is considered to be short in the population of Dutch men, hence (short(\u03bbx e .man(x)\u2227dutch(x))) (jan) = 1, but not in the population of all men, hence (short (man))(jan) = 0. This is a consequence of having short denoting a non-intersective modifier: the set denoted by short(\u03bbx e .man(x) \u2227 dutch(x)) is not necessarily a subset of short(man).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Analyzing entailments using the interpreted lexicon",
"sec_num": "4.2"
},
{
"text": "Let us assume parse trees as follows: In the first stages of our attempt to implement the theoretical model described above, we faced a practical problem concerning the binding of expressions in the RTE data to structurally equivalent expressions in the interpreted lexicon: we currently lack an annotation scheme and a user interface that allows annotators to consistently and effectively annotate RTE data. The root of this problem lies in the intricate ways in which the semantic phenomena that we are concerned with are combined with other phenomena or with each other. Simplifying RTE material to an extent that allows binding it to the lexicon as in the above example is often not straightforward. Consider the following example: Example 4",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "\u2022 Analysis of Pair 2",
"sec_num": null
},
{
"text": "-Text: Jan [",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "\u2022 Analysis of Pair 2",
"sec_num": null
},
{
"text": "\u2022 T: Comdex -once among the world's largest trade shows, the launching pad for new computer and software products, and a Las Vegas fixture for 20 years -has been canceled for this year.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "\u2022 Analysis of Pair 2",
"sec_num": null
},
{
"text": "\u2022 H: Las Vegas hosted the Comdex trade show for 20 years. 9 Validating the entailment in this pair requires a lexical alignment between an expression in the text and the word hosted in the hypothesis. However, there is no expression in the text to establish this alignment. In the text, the noun Comdex is in an appositive relation with three conjoined propositions: (i) once among the world's largest trade shows; (ii) the launching pad for new computer and software products; and (iii) a Las Vegas fixture for 20 years. The third element contains a locative restrictive modification in which Las Vegas modifies fixture. The apposition licenses the inference that Comdex IS a Las Vegas fixture and serves as a prerequisite for the alignment: Comdex is a Las Vegas fixture \u21d2 Las Vegas hosted Comdex that simplifies the lexical inference. This alignment is also required for validating the modification by the temporal prepositional phrase for 20 years which in the text modifies a noun, fixture, and in the hypothesis modifies a verb, host -apparently two unrelated lexical items. This example illustrates the difficulty in separating lexical inferences from the semantic relations that underlie the constructions they appear in. In this sense, the manual annotation process that we exemplified in Section 4, in which the stage of Phenomena Simplification takes place before the semantic machinery applies, is challenging and requires further investigation with RTE data in order to see what part of the RTE can be annotated using this paradigm, and what elements are needed in order to extend its coverage.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "\u2022 Analysis of Pair 2",
"sec_num": null
},
{
"text": "Due to this challenge, and in order to enhance our understanding of the phenomena in the RTE corpora, we adopted a narrower annotation scheme that was carried out on RTE 1-4, named SemAnTE 1.0 -Semantic Annotation of Textual Entailment. 10 In this annotation work we focused on valid entailments involving restrictive, intersective and appositive modification that contribute to the recognition of the entailment. 11 In this approach, a construction is annotated if its semantics are required for validating the entailment, but no account is made of the compositional method in which the meaning of the full sentence is obtained. Annotations were marked in 80.65% of the entailments in the RTE 1-4 corpora and reached cross-annotator agreement of 68% on average in four consistency checks. The internal structure of the annotated XML files and a use-case of the annotations for evaluating an entailment component in the BIUTEE recognizer (Stern and Dagan, 2011) are presented in Toledo et al. (2012) . See Garoufi (2007) for other relevant work on semantic analysis and annotation of textual entailment done on RTE 2.",
"cite_spans": [
{
"start": 938,
"end": 961,
"text": "(Stern and Dagan, 2011)",
"ref_id": "BIBREF13"
},
{
"start": 979,
"end": 999,
"text": "Toledo et al. (2012)",
"ref_id": "BIBREF14"
},
{
"start": 1006,
"end": 1020,
"text": "Garoufi (2007)",
"ref_id": "BIBREF7"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "\u2022 Analysis of Pair 2",
"sec_num": null
},
{
"text": "Our annotations mark inferences by aligning strings in the text and the hypothesis. This is done by pairing each annotation in the text with a corresponding annotation in the hypothesis that marks the output of the inferential process of the phenomenon in question. In the rest of this section we illustrate the phenomena and underline the annotated part in the text with its correspondence in the hypothesis.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Phenomena Annotated",
"sec_num": "5.1"
},
{
"text": "\u2022 T: A Cuban M odif ier American M odif iee who is accused of espionage pleads innocent.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Restrictive modification (RMOD)",
"sec_num": "5.2"
},
{
"text": "\u2022 H: American accused of espionage. In this case, Cuban modifies American and restricts the set of Americans to Cuban Americans. This instance of RMOD validates the inference from Cuban American to American which is required for establishing the entailment. The intersective nature of the process is not exploited in the actual inference, since the hypothesis does not report that the accused person is Cuban. Thus, only the restrictive property of the modifier Cuban is here relevant for the validity of the entailment. More syntactic configurations:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Restrictive modification (RMOD)",
"sec_num": "5.2"
},
{
"text": "\u2022 A verb phrase restricted by a prepositional phrase: -T: The watchdog International Atomic Energy Agency meets in Vienna M odif iee on September 19 M odif ier . -H: The International Atomic Energy Agency holds a meeting in Vienna.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Restrictive modification (RMOD)",
"sec_num": "5.2"
},
{
"text": "\u2022 A noun phrase restricted by a prepositional phrase:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Restrictive modification (RMOD)",
"sec_num": "5.2"
},
{
"text": "-T: U.S. officials have been warning for weeks of possible terror attacks M odif iee against U.S. interests M odif ier . -H: The United States has warned a number of times of possible terrorist attacks.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Restrictive modification (RMOD)",
"sec_num": "5.2"
},
{
"text": "\u2022 T: Nixon was impeached and became the first president ever to resign on August 9th 1974.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Intersective Modification (CONJ)",
"sec_num": "5.3"
},
{
"text": "\u2022 H: Nixon was the first president ever to resign. This conjunction intersects the two verb phrases was impeached and became the first president ever to resign. The entailment relies on a subsumption of the full construction to the second conjunct. In addition to canonical conjunctive constructions, CONJ appears also in Restrictive Relative Clauses whereby the relative clause is interpreted intersectively with the noun being modified:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Intersective Modification (CONJ)",
"sec_num": "5.3"
},
{
"text": "\u2022 T: Iran will soon release eight British servicemen detained along with three vessels.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Intersective Modification (CONJ)",
"sec_num": "5.3"
},
{
"text": "\u2022 H: British servicemen detained.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Intersective Modification (CONJ)",
"sec_num": "5.3"
},
{
"text": "\u2022 Appositive subsumption (left part):",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Appositive modification (APP)",
"sec_num": "5.4"
},
{
"text": "-T: Mr. Conway, Iamgold's chief executive officer, said the vote would be close.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Appositive modification (APP)",
"sec_num": "5.4"
},
{
"text": "-H: Mr. Conway said the vote would be close.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Appositive modification (APP)",
"sec_num": "5.4"
},
{
"text": "\u2022 Identification of the two parts of the apposition as referring to one another: -T: The incident in Mogadishu, the Somali capital, came as U.S. forces began the final phase of their promised March 31 pullout. -H: The capital of Somalia is Mogadishu.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Appositive modification (APP)",
"sec_num": "5.4"
},
{
"text": "In addition to appositions, APP is annotated in several more syntactic constructions:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Appositive modification (APP)",
"sec_num": "5.4"
},
{
"text": "\u2022 Non-Restrictive Relative Clauses:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Appositive modification (APP)",
"sec_num": "5.4"
},
{
"text": "-T: A senior coalition official in Iraq said the body, which was found by U.S. military police west of Baghdad, appeared to have been thrown from a vehicle. -H: A body has been found by U. S. military police.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Appositive modification (APP)",
"sec_num": "5.4"
},
{
"text": "\u2022 Title Constructions:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Appositive modification (APP)",
"sec_num": "5.4"
},
{
"text": "-T: Prime Minister Silvio Berlusconi was elected March 28 with a mandate to reform Italy's business regulations and pull the economy out of recession. -H: The Prime Minister is Silvio Berlusconi.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Appositive modification (APP)",
"sec_num": "5.4"
},
{
"text": "Given a pair from the RTE in which the entailment relation obtains between the text and hypothesis, the task for the annotators is defined as follows: 1. Read the data, verify the entailment and describe informally why the entailment holds.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Marking Annotations",
"sec_num": "5.5"
},
{
"text": "2. Annotate all instances of RMOD, APP and CONJ that play a role in the inferential process.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Marking Annotations",
"sec_num": "5.5"
},
{
"text": "The annotated corpus is based on the scheme described above, applied to the datasets of RTE 1-4 (Dagan et al., 2006; Bar Haim et al., 2006; Giampiccolo et al., 2007 Giampiccolo et al., , 2008 . We report annotation statistics in Table 2 and consistency measurements in Table 3 . In each consistency check we picked 50-70 entailment pairs that both annotators worked on independently, and compared the phenomena that were annotated.",
"cite_spans": [
{
"start": 96,
"end": 116,
"text": "(Dagan et al., 2006;",
"ref_id": "BIBREF6"
},
{
"start": 117,
"end": 139,
"text": "Bar Haim et al., 2006;",
"ref_id": "BIBREF0"
},
{
"start": 140,
"end": 164,
"text": "Giampiccolo et al., 2007",
"ref_id": "BIBREF9"
},
{
"start": 165,
"end": 191,
"text": "Giampiccolo et al., , 2008",
"ref_id": "BIBREF8"
}
],
"ref_spans": [
{
"start": 229,
"end": 236,
"text": "Table 2",
"ref_id": "TABREF5"
},
{
"start": 269,
"end": 276,
"text": "Table 3",
"ref_id": "TABREF6"
}
],
"eq_spans": [],
"section": "Annotation Statistics and Consistency",
"sec_num": "5.6"
},
{
"text": "We used GATE Developer (Cunningham et al., 2011) to annotate the original RTE XML files. The work was performed in two steps using GATE annotation schemes that correspond to RMOD, APP and CONJ: (1) marking the relevant string in the text using one of GATE's schemes (e.g. a scheme of appositive modification), and (2) -marking a string in the hypothesis that corresponds to the output of the inferential process. The annotation in the hypothesis was done using a dedicated reference to scheme.",
"cite_spans": [
{
"start": 23,
"end": 48,
"text": "(Cunningham et al., 2011)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Annotation Platform",
"sec_num": "5.7"
},
{
"text": "Consider the following pair from RTE 2: Example 5",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Connection to the interpreted lexicon approach",
"sec_num": "5.8"
},
{
"text": "\u2022 T: The anti-terrorist court found two men guilty of murdering Shapour Bakhtiar and his secretary Sorush Katibeh, who were found with their throats cut in August 1991.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Connection to the interpreted lexicon approach",
"sec_num": "5.8"
},
{
"text": "\u2022 H: Shapour Bakhtiar died in 1991. Several entailment patterns in this example can be explained by appealing to the semantics of APP, CONJ and RMOD, as follows:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Connection to the interpreted lexicon approach",
"sec_num": "5.8"
},
{
"text": "\u2022 APP: The appositive modification in Shapour Bakhtiar and his secretary Sorush Katibeh, who were found with their throats cut in August 1991 licenses the inference that Shapour Bakhtiar and his secretary Sorush Katibeh were found with their throats cut in August 1991.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Connection to the interpreted lexicon approach",
"sec_num": "5.8"
},
{
"text": "\u2022 RMOD: The restrictive modification in August 1991 licenses a subsumption to 1991. \u2022 CONJ: The conjunction in Shapour Bakhtiar and his secretary Sorush Katibeh licenses a subsumption of this expression to Shapour Bakhtiar. By combining these three patterns, we can infer that Shapour Bakhtiar was found with his throat cut in 1991. However, additional world knowledge is required to infer that found with his throat cut entails died. In our current annotation scheme this inference cannot be handled since lexical alignment of unmodeled phenomena is not supported . This motivates a more robust approach as proposed in Section 4.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Connection to the interpreted lexicon approach",
"sec_num": "5.8"
},
{
"text": "The goal of this research is to establish a model-theoretic benchmark explaining entailment data. We have presented a model that utilizes standard semantic principles and illustrated the way it accounts for textual entailment from the RTE corpora. The model centers upon an interpreted lexicon that comprises words and operators. These elements are used to represent a fragment of English to which premises and hypotheses may be bound.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusions",
"sec_num": "6"
},
{
"text": "We focus on the annotation of semantic phenomena which are predominant in the RTE corpora and can be annotated with high consistency, but which may have several syntactic expressions and therefore allow us to generalize regarding abstract entailment patterns. Non-modeled phenomena that exist in the data are simplified in a preparatory step but cases in which such phenomena are deeply intertwined with the semantic phenomena that we model pose a challenge for the formalization of an annotation scheme.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusions",
"sec_num": "6"
},
{
"text": "At a first stage, we carried out a restricted annotation scheme by which instances of restrictive, intersective, and appositive modification are marked in entailment pairs with no account for the full inferential process between the premise and the hypothesis. These phenomena were found in 80.65% of the entailments in RTE 1-4 and were marked with cross-annotator agreement of 68% on average.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusions",
"sec_num": "6"
},
{
"text": "We are currently investigating different directions in the formulation of an extensive annotation scheme coincident with the model we described and are aiming to develop a corresponding annotation platform. This platform would allow annotators to bind constructions manifesting supported semantic phenomena to representations in the interpreted lexicon as well as to simplify lexical/syntactic phenomena of the kind illustrated in Examples 2 and 4 by textual alignment. In the next stages of this project, we plan to use an external theorem prover to automatically validate the entailment relation (or lack thereof).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusions",
"sec_num": "6"
},
{
"text": "This example of causal reasoning is taken fromMacCartney and Manning (2007).4 Another related work, which approaches inference in natural language as part of a semantic paradigm, is the FraCaS test suite(Cooper et al., 1996). This suit concerns examples that mainly rely on generalized quantification, argument monotonicity, plurality, anaphora resolution, ellipsis, etc. Entailments based on these phenomena are not very common in the RTE data that are analyzed here. Further research is needed in order to integrate data like those in FraCaS into a formal annotation scheme like the one suggested in this paper.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "The only higher-order constants in the above lexicon are the (et)(et) constants attributed to non-intersective restrictive modifiers. Treating them in predicate calculus theorem provers may require some ad hoc assumptions.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "Pair 955 from the test set of RTE 4(Giampiccolo et al., 2008).7 Note that the post-nominal intersective modifier on the web is bound to a pre-nominal modifier Dutch. This is done in order to match the vocabulary of the interpreted lexicon, in which the only intersective modifier is Dutch.8 In this example, T Simple (consequently from T Original ) is structurally ambiguous between The [largest [search engine on the web]], which is Google, receives... and The [[largest search engine] on the web], which is Google, receives.... We illustrate the former analysis here. The latter analysis can be handled in a similar vein.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "Pair 214 from the development set of RTE 1.(Dagan et al., 2006) 10 The annotated files of SemAnTE are publicly available for download at http://sophia.katrenko.com/CorpusDownload/ 11 Annotators were instructed to construct a full inferential process informally and then to recognize the contribution of the phenomena we aimed to annotate. This method could be applied efficiently only to valid entailments. Invalid entailments marked as unknown exhibit an unidentified relation between the text and hypothesis, and pairs marked as contradictory rarely center upon the phenomena in question.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
}
],
"back_matter": [],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "The second pascal recognising textual entailment challenge",
"authors": [
{
"first": "R",
"middle": [],
"last": "Bar Haim",
"suffix": ""
},
{
"first": "I",
"middle": [],
"last": "Dagan",
"suffix": ""
},
{
"first": "B",
"middle": [],
"last": "Dolan",
"suffix": ""
},
{
"first": "L",
"middle": [],
"last": "Ferro",
"suffix": ""
},
{
"first": "D",
"middle": [],
"last": "Giampiccolo",
"suffix": ""
},
{
"first": "B",
"middle": [],
"last": "Magnini",
"suffix": ""
},
{
"first": "I",
"middle": [],
"last": "Szpektor",
"suffix": ""
}
],
"year": 2006,
"venue": "Proceedings of the Second PASCAL Challenges Workshop on Recognising Textual Entailment",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Bar Haim, R., I. Dagan, B. Dolan, L. Ferro, D. Giampiccolo, B. Magnini, and I. Szpektor (2006). The second pascal recognising textual entailment challenge. In In Proceedings of the Second PASCAL Challenges Workshop on Recognising Textual Entailment.",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "The fifth pascal recognizing textual entailment challenge",
"authors": [
{
"first": "L",
"middle": [],
"last": "Bentivogli",
"suffix": ""
},
{
"first": "I",
"middle": [],
"last": "Dagan",
"suffix": ""
},
{
"first": "H",
"middle": [
"T"
],
"last": "Dang",
"suffix": ""
},
{
"first": "D",
"middle": [],
"last": "Giampiccolo",
"suffix": ""
},
{
"first": "B",
"middle": [],
"last": "Magnini",
"suffix": ""
}
],
"year": 2009,
"venue": "Proceedings of TAC 9",
"volume": "",
"issue": "",
"pages": "14--24",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Bentivogli, L., I. Dagan, H. T. Dang, D. Giampiccolo, and B. Magnini (2009). The fifth pascal recogniz- ing textual entailment challenge. Proceedings of TAC 9, 14-24.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "Wide-coverage semantic representations from a CCG parser",
"authors": [
{
"first": "J",
"middle": [],
"last": "Bos",
"suffix": ""
},
{
"first": "S",
"middle": [],
"last": "Clark",
"suffix": ""
},
{
"first": "M",
"middle": [],
"last": "Steedman",
"suffix": ""
},
{
"first": "J",
"middle": [
"R"
],
"last": "Curran",
"suffix": ""
},
{
"first": "J",
"middle": [],
"last": "Hockenmaier",
"suffix": ""
}
],
"year": 2004,
"venue": "Proceedings of the 20th international conference on Computational Linguistics",
"volume": "",
"issue": "",
"pages": "12--40",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Bos, J., S. Clark, M. Steedman, J. R. Curran, and J. Hockenmaier (2004). Wide-coverage semantic repre- sentations from a CCG parser. In Proceedings of the 20th international conference on Computational Linguistics, pp. 12-40.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "Recognising textual entailment with logical inference",
"authors": [
{
"first": "J",
"middle": [],
"last": "Bos",
"suffix": ""
},
{
"first": "K",
"middle": [],
"last": "Markert",
"suffix": ""
}
],
"year": 2005,
"venue": "Proceedings of the conference on Human Language Technology and Empirical Methods in Natural Language Processing",
"volume": "",
"issue": "",
"pages": "628--635",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Bos, J. and K. Markert (2005). Recognising textual entailment with logical inference. In Proceedings of the conference on Human Language Technology and Empirical Methods in Natural Language Pro- cessing, pp. 628-635.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "Using the Framework",
"authors": [
{
"first": "R",
"middle": [],
"last": "Cooper",
"suffix": ""
},
{
"first": "D",
"middle": [],
"last": "Crouch",
"suffix": ""
},
{
"first": "J",
"middle": [],
"last": "Van Eijck",
"suffix": ""
},
{
"first": "C",
"middle": [],
"last": "Fox",
"suffix": ""
},
{
"first": "J",
"middle": [],
"last": "Van Genabith",
"suffix": ""
},
{
"first": "J",
"middle": [],
"last": "Jaspars",
"suffix": ""
},
{
"first": "H",
"middle": [],
"last": "Kamp",
"suffix": ""
},
{
"first": "D",
"middle": [],
"last": "Milward",
"suffix": ""
},
{
"first": "M",
"middle": [],
"last": "Pinkal",
"suffix": ""
},
{
"first": "M",
"middle": [],
"last": "Poesio",
"suffix": ""
},
{
"first": "S",
"middle": [],
"last": "Pulman",
"suffix": ""
},
{
"first": "T",
"middle": [],
"last": "Briscoe",
"suffix": ""
},
{
"first": "H",
"middle": [],
"last": "Maier",
"suffix": ""
},
{
"first": "K",
"middle": [],
"last": "Konrad",
"suffix": ""
}
],
"year": 1996,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Cooper, R., D. Crouch, J. Van Eijck, C. Fox, J. Van Genabith, J. Jaspars, H. Kamp, D. Milward, M. Pinkal, M. Poesio, S. Pulman, T. Briscoe, H. Maier, and K. Konrad (1996). Using the Frame- work. The Fracas Consortium.",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "The pascal recognising textual entailment challenge",
"authors": [
{
"first": "I",
"middle": [],
"last": "Dagan",
"suffix": ""
},
{
"first": "B",
"middle": [],
"last": "O. Glickman",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Magnini",
"suffix": ""
}
],
"year": 2006,
"venue": "Machine Learning Challenges. Evaluating Predictive Uncertainty, Visual Object Classification, and Recognising Tectual Entailment",
"volume": "",
"issue": "",
"pages": "177--190",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Dagan, I., O. Glickman, and B. Magnini (2006). The pascal recognising textual entailment challenge. Machine Learning Challenges. Evaluating Predictive Uncertainty, Visual Object Classification, and Recognising Tectual Entailment, 177-190.",
"links": null
},
"BIBREF7": {
"ref_id": "b7",
"title": "Towards a better understanding of applied textual entailment: Annotation and evaluation of the rte-2 dataset",
"authors": [
{
"first": "K",
"middle": [],
"last": "Garoufi",
"suffix": ""
}
],
"year": 2007,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Garoufi, K. (2007). Towards a better understanding of applied textual entailment: Annotation and eval- uation of the rte-2 dataset. Master's thesis, Saarland University.",
"links": null
},
"BIBREF8": {
"ref_id": "b8",
"title": "The fourth pascal recognising textual entailment challenge",
"authors": [
{
"first": "D",
"middle": [],
"last": "Giampiccolo",
"suffix": ""
},
{
"first": "H",
"middle": [
"T"
],
"last": "Dang",
"suffix": ""
},
{
"first": "B",
"middle": [],
"last": "Magnini",
"suffix": ""
},
{
"first": "I",
"middle": [],
"last": "Dagan",
"suffix": ""
},
{
"first": "E",
"middle": [],
"last": "Cabrio",
"suffix": ""
}
],
"year": 2008,
"venue": "TAC 2008 Proceedings",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Giampiccolo, D., H. T. Dang, B. Magnini, I. Dagan, and E. Cabrio (2008). The fourth pascal recognising textual entailment challenge. In In TAC 2008 Proceedings.",
"links": null
},
"BIBREF9": {
"ref_id": "b9",
"title": "The third pascal recognizing textual entailment challenge",
"authors": [
{
"first": "D",
"middle": [],
"last": "Giampiccolo",
"suffix": ""
},
{
"first": "B",
"middle": [],
"last": "Magnini",
"suffix": ""
},
{
"first": "I",
"middle": [],
"last": "Dagan",
"suffix": ""
},
{
"first": "B",
"middle": [],
"last": "Dolan",
"suffix": ""
}
],
"year": 2007,
"venue": "Proceedings of the ACL-PASCAL Workshop on Textual Entailment and Paraphrasing, RTE '07",
"volume": "",
"issue": "",
"pages": "1--9",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Giampiccolo, D., B. Magnini, I. Dagan, and B. Dolan (2007). The third pascal recognizing textual entailment challenge. In In Proceedings of the ACL-PASCAL Workshop on Textual Entailment and Paraphrasing, RTE '07, Stroudsburg, PA, USA, pp. 1-9. Association for Computational Linguistics.",
"links": null
},
"BIBREF10": {
"ref_id": "b10",
"title": "From discourse to logic: Introduction to model-theoretic semantics of natural language, formal logic and discourse representation theory",
"authors": [
{
"first": "H",
"middle": [],
"last": "Kamp",
"suffix": ""
},
{
"first": "U",
"middle": [],
"last": "Reyle",
"suffix": ""
}
],
"year": 1993,
"venue": "",
"volume": "42",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Kamp, H. and U. Reyle (1993). From discourse to logic: Introduction to model-theoretic semantics of natural language, formal logic and discourse representation theory, Volume 42. Kluwer Academic Dordrecht,, The Netherlands.",
"links": null
},
"BIBREF11": {
"ref_id": "b11",
"title": "Natural logic for textual inference",
"authors": [
{
"first": "B",
"middle": [],
"last": "Maccartney",
"suffix": ""
},
{
"first": "C",
"middle": [
"D"
],
"last": "Manning",
"suffix": ""
}
],
"year": 2007,
"venue": "Proceedings of the ACL-PASCAL Workshop on Textual Entailment and Paraphrasing",
"volume": "",
"issue": "",
"pages": "193--200",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "MacCartney, B. and C. D. Manning (2007). Natural logic for textual inference. In Proceedings of the ACL-PASCAL Workshop on Textual Entailment and Paraphrasing, pp. 193-200.",
"links": null
},
"BIBREF12": {
"ref_id": "b12",
"title": "Logics for the relational syllogistic",
"authors": [
{
"first": "I",
"middle": [],
"last": "Pratt-Hartmann",
"suffix": ""
},
{
"first": "L",
"middle": [
"S"
],
"last": "Moss",
"suffix": ""
}
],
"year": 2009,
"venue": "Review of Symbolic Logic",
"volume": "",
"issue": "",
"pages": "647--683",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Pratt-Hartmann, I. and L. S. Moss (2009). Logics for the relational syllogistic. Review of Symbolic Logic, 647-683.",
"links": null
},
"BIBREF13": {
"ref_id": "b13",
"title": "A confidence model for syntactically-motivated entailment proofs",
"authors": [
{
"first": "A",
"middle": [],
"last": "Stern",
"suffix": ""
},
{
"first": "I",
"middle": [],
"last": "Dagan",
"suffix": ""
}
],
"year": 2011,
"venue": "Proceedings of RANLP",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Stern, A. and I. Dagan (2011). A confidence model for syntactically-motivated entailment proofs. In Proceedings of RANLP 2011.",
"links": null
},
"BIBREF14": {
"ref_id": "b14",
"title": "Semantic annotation for textual entailment recognition",
"authors": [
{
"first": "A",
"middle": [],
"last": "Toledo",
"suffix": ""
},
{
"first": "S",
"middle": [],
"last": "Katrenko",
"suffix": ""
},
{
"first": "S",
"middle": [],
"last": "Alexandropoulou",
"suffix": ""
},
{
"first": "H",
"middle": [],
"last": "Klockmann",
"suffix": ""
},
{
"first": "A",
"middle": [],
"last": "Stern",
"suffix": ""
},
{
"first": "I",
"middle": [],
"last": "Dagan",
"suffix": ""
},
{
"first": "Y",
"middle": [],
"last": "Winter",
"suffix": ""
}
],
"year": 2012,
"venue": "Proceedings of the Eleventh Mexican International Conference on Artificial Intelligence (MICAI)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Toledo, A., S. Katrenko, S. Alexandropoulou, H. Klockmann, A. Stern, I. Dagan, and Y. Winter (2012). Semantic annotation for textual entailment recognition. In Proceedings of the Eleventh Mexican In- ternational Conference on Artificial Intelligence (MICAI).",
"links": null
},
"BIBREF15": {
"ref_id": "b15",
"title": "Studies on natural logic and categorial grammar",
"authors": [
{
"first": "V",
"middle": [
"S"
],
"last": "Valencia",
"suffix": ""
}
],
"year": 1991,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Valencia, V. S. (1991). Studies on natural logic and categorial grammar. Ph. D. thesis, University of Amsterdam.",
"links": null
}
},
"ref_entries": {
"FIGREF0": {
"type_str": "figure",
"num": null,
"uris": null,
"text": "if A = (\u03bbx e .x = a) undefined otherwise (iota operator) SOME = \u03bbA et .\u03bbB et .\u2203x e .A(x) \u2227 B(x) WHO R = AND = \u03bbA et .\u03bbB et .\u03bbx e .B(x) \u2227 A(x) WHO A = \u03bbA et .\u03bbx e .\u03b9(\u03bby.y = x \u2227 A(x)) P r = \u03bbM (et)(et) .\u03bbx e .M (\u03bby e .1)(x) deriving a predicate from a general modifier I m = AND = \u03bbA et .\u03bbB et .\u03bbx e .B(x) \u2227 A(x) deriving an intersective modifier R m = \u03bbM (et)(et) .\u03bbA et .\u03bbx e .M (A)(x) \u2227 A(x) deriving a restrictive modifierFigure 1: Functions used in the interpreted lexicon"
},
"FIGREF1": {
"type_str": "figure",
"num": null,
"uris": null,
"text": "(short))((I m (dutch))(man)))))(jan)analysis = ((R m (short))((I m (dutch))(man)))(jan) def. of A, IS = (((\u03bbM (et)(et) .\u03bbA et .\u03bby e .M (A)(y) \u2227 A(y))(short)) (((\u03bbA et .\u03bbB et .\u03bbx e .B(x) \u2227 A(x)) (dutch))(man)))(jan)def. of I m , R m = 1 \u2227 1 \u2227 1 = 1 func. application + denotations in M -Hypothesis:"
},
"FIGREF2": {
"type_str": "figure",
"num": null,
"uris": null,
"text": "(black))((I m (dutch))(man)))))(jan) analysis = (((\u03bbA et .\u03bbB et .\u03bby e .B(y) \u2227 A(y))(black))(((\u03bbA et .\u03bbB et .\u03bbx e .B(x) \u2227 A(x)) (dutch))(man)))(jan) def. of A, IS, I m = dutch(jan) \u2227 (man(jan) \u2227 black(jan)) func. application \u2264 man(jan) \u2227 black(jan) def. of \u2227 = (IS(A((I m (black))(man))))(jan) = [[ Jan [is [a [black man]]] ]] M beta reduction + def. of I m , A, IS + analysis In this case we rely on the intersectivity of black, which in conjunction with the intersectivity of Dutch licenses the inference that the set characterized by the et function [[ black [Dutch man] ]] M equals to the set characterized by [[ Dutch [black man] ]] M , which is a subset of the set characterized by [[ black man ]] M ."
},
"TABREF1": {
"num": null,
"html": null,
"text": "An Interpreted Lexicon",
"type_str": "table",
"content": "<table/>"
},
"TABREF2": {
"num": null,
"html": null,
"text": "Simple : [The largest search engine on the web, which is Google,] receives... \u2022 T Lexicon : [The short Dutch man, who is Jan,] saw Dan Hypothesis: \u2022 H Simple : Google [is [on the web]] \u2022 H Lexicon : Jan [is Dutch] 3. Proof of entailment T Lexicon \u21d2 H Lexicon : Let M be an intended model,",
"type_str": "table",
"content": "<table><tr><td>H Original : Google operates on the web</td></tr><tr><td>\u2022 H Simple : Google is on the web</td></tr><tr><td>2. Binding to Lexicon:</td></tr><tr><td>Text 7,8 :</td></tr><tr><td>\u2022 T</td></tr></table>"
},
"TABREF3": {
"num": null,
"html": null,
"text": "jan, vim} dutch et characterizes {jan, vim} short(man) et characterizes {dan} short(\u03bby e .man(y) \u2227 dutch(y)) et characterizes {jan} Let us assume parse trees as follows:",
"type_str": "table",
"content": "<table><tr><td>-Text: Jan [is [a [short [Dutch man]]]]</td></tr><tr><td>-Hypothesis: Jan [is [a [short man]]]</td></tr><tr><td>Consider the denotations of the text and hypothesis in the model M:</td></tr><tr><td>-Text:</td></tr><tr><td>[[ Jan [is [a [short [Dutch man]]]] ]] M</td></tr><tr><td>= (IS(A((R m</td></tr></table>"
},
"TABREF4": {
"num": null,
"html": null,
"text": "In analyzing this pair we can show a proof of entailment. Let M be an intended model,",
"type_str": "table",
"content": "<table><tr><td>[[ Jan [is [a [black [Dutch man]]]] ]] M</td></tr><tr><td>= (IS(A((I m</td></tr></table>"
},
"TABREF5": {
"num": null,
"html": null,
"text": "Counters of annotations in RTE 1-4 separated into development and test sets. A # indicates the number of annotations, P # indicates the number of entailment pairs containing an annotation and P % indicates the portion of annotated pairs relative to the total amount of entailment pairs.",
"type_str": "table",
"content": "<table><tr><td/><td/><td>(a) RTE 1</td><td/><td/><td>(b) RTE 2</td></tr><tr><td/><td/><td>Dev set</td><td>Test set</td><td/><td>Dev set</td><td>Test set</td></tr><tr><td>Ann.</td><td colspan=\"3\">A # P # P % A # P # P %</td><td>Ann.</td><td>A # P # P % A # P # P %</td></tr><tr><td>APP</td><td>97</td><td colspan=\"2\">87 31 161 134 34</td><td>APP</td><td>179 149 37 155 135 34</td></tr><tr><td>CONJ</td><td>90</td><td colspan=\"2\">79 28 126 112 28</td><td colspan=\"2\">CONJ 141 119 30 161 144 36</td></tr><tr><td colspan=\"4\">RMOD 180 124 44 243 167 42</td><td colspan=\"2\">RMOD 314 205 51 394 236 59</td></tr><tr><td>Any</td><td colspan=\"3\">367 210 74 530 297 74</td><td>Any</td><td>634 318 80 710 350 88</td></tr><tr><td/><td/><td>(c) RTE 3</td><td/><td/><td>(d) RTE 4</td></tr><tr><td/><td/><td>Dev set</td><td>Test set</td><td/><td>Test set</td></tr><tr><td>Ann.</td><td colspan=\"3\">A # P # P % A # P # P %</td><td/><td>Ann.</td><td>A # P # P %</td></tr><tr><td>APP</td><td colspan=\"3\">188 150 38 166 136 34</td><td/><td>APP</td><td>259 200 40</td></tr><tr><td colspan=\"4\">CONJ 176 138 35 162 134 34</td><td/><td>CONJ 192 164 33</td></tr><tr><td colspan=\"4\">RMOD 300 201 50 307 193 48</td><td/><td>RMOD 429 271 54</td></tr><tr><td>Any</td><td colspan=\"3\">664 329 82 635 328 82</td><td/><td>Any</td><td>880 413 83</td></tr></table>"
},
"TABREF6": {
"num": null,
"html": null,
"text": "Results of Four Consistency Checks. Each check examined 50-70 annotated pairs from RTE 1-4. In these four checks 66%, 74.11%, 66.67% and 64.66% of the annotations were identical, respectively. On average, 68.03% of the annotations we checked were identical. The rubric Incorrect Ann. presents cases of annotations done with an incorrect scheme or with an incorrect scope. Ambig.-Struct. are cases of structural or modifier-attachment ambiguity in the text that led to divergent annotations. Ambig.-Infer. are cases of divergent annotations stemming from several possible analyses of the inference. Ambig.-Scheme refers to instances of divergent annotations due to unclarity or limited specification in the annotation scheme. The last two measures are reported only for the second, third and forth checks.",
"type_str": "table",
"content": "<table><tr><td>Measure</td><td colspan=\"2\">RTE 1 RTE 1+2</td><td>RTE 3</td><td>RTE 4</td></tr><tr><td colspan=\"5\">Data Source(s) Dev set Test sets Dev+Test sets Test set</td></tr><tr><td>Entailment Pairs</td><td>50</td><td>70</td><td>70</td><td>70</td></tr><tr><td>Total Ann.</td><td>93</td><td>112</td><td>99</td><td>133</td></tr><tr><td>Identical Ann.</td><td>62</td><td>83</td><td>66</td><td>86</td></tr><tr><td>Missing Ann.</td><td>2</td><td>7</td><td>7</td><td>10</td></tr><tr><td>Incorrect Ann.</td><td>10</td><td>1</td><td>2</td><td>2</td></tr><tr><td>Ambig.-Struct.</td><td>9</td><td>16</td><td>20</td><td>15</td></tr><tr><td>Ambig.-Infer.</td><td>N/A</td><td>8</td><td>13</td><td>12</td></tr><tr><td>Ambig.-Scheme</td><td>N/A</td><td>0</td><td>9</td><td>7</td></tr><tr><td colspan=\"2\">Consistency (%) 66.67</td><td>74.11</td><td>66.67</td><td>64.66</td></tr></table>"
}
}
}
}