| { |
| "paper_id": "D11-1027", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T16:33:07.247281Z" |
| }, |
| "title": "Minimally Supervised Event Causality Identification", |
| "authors": [ |
| { |
| "first": "Xuan", |
| "middle": [], |
| "last": "Quang", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "quangdo2@illinois.edu" |
| }, |
| { |
| "first": "Yee", |
| "middle": [ |
| "Seng" |
| ], |
| "last": "Do", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| }, |
| { |
| "first": "Dan", |
| "middle": [], |
| "last": "Chan", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "chanys@illinois.edu" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Roth", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "This paper develops a minimally supervised approach, based on focused distributional similarity methods and discourse connectives, for identifying of causality relations between events in context. While it has been shown that distributional similarity can help identifying causality, we observe that discourse connectives and the particular discourse relation they evoke in context provide additional information towards determining causality between events. We show that combining discourse relation predictions and distributional similarity methods in a global inference procedure provides additional improvements towards determining event causality.", |
| "pdf_parse": { |
| "paper_id": "D11-1027", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "This paper develops a minimally supervised approach, based on focused distributional similarity methods and discourse connectives, for identifying of causality relations between events in context. While it has been shown that distributional similarity can help identifying causality, we observe that discourse connectives and the particular discourse relation they evoke in context provide additional information towards determining causality between events. We show that combining discourse relation predictions and distributional similarity methods in a global inference procedure provides additional improvements towards determining event causality.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "An important part of text understanding arises from understanding the semantics of events described in the narrative, such as identifying the events that are mentioned and how they are related semantically. For instance, when given a sentence \"The police arrested him because he killed someone.\", humans understand that there are two events, triggered by the words \"arrested\" and \"killed\", and that there is a causality relationship between these two events. Besides being an important component of discourse understanding, automatically identifying causal relations between events is important for various natural language processing (NLP) applications such as question answering, etc. In this work, we automatically detect and extract causal relations between events in text.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Despite its importance, prior work on event causality extraction in context in the NLP literature is relatively sparse. In (Girju, 2003) , the author used noun-verb-noun lexico-syntactic patterns to learn that \"mosquitoes cause malaria\", where the cause and effect mentions are nominals and not necessarily event evoking words. In (Sun et al., 2007) , the authors focused on detecting causality between search query pairs in temporal query logs. (Beamer and Girju, 2009) tried to detect causal relations between verbs in a corpus of screen plays, but limited themselves to consecutive, or adjacent verb pairs. In (Riaz and Girju, 2010) , the authors first cluster sentences into topic-specific scenarios, and then focus on building a dataset of causal text spans, where each span is headed by a verb. Thus, their focus was not on identifying causal relations between events in a given text document.", |
| "cite_spans": [ |
| { |
| "start": 123, |
| "end": 136, |
| "text": "(Girju, 2003)", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 331, |
| "end": 349, |
| "text": "(Sun et al., 2007)", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 446, |
| "end": 470, |
| "text": "(Beamer and Girju, 2009)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 613, |
| "end": 635, |
| "text": "(Riaz and Girju, 2010)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this paper, given a text document, we first identify events and their associated arguments. We then identify causality or relatedness relations between event pairs. To do this, we develop a minimally supervised approach using focused distributional similarity methods, such as co-occurrence counts of events collected automatically from an unannotated corpus, to measure and predict existence of causality relations between event pairs. Then, we build on the observation that discourse connectives and the particular discourse relation they evoke in context provide additional information towards determining causality between events. For instance, in the example sentence provided at the beginning of this section, the words \"arrested\" and \"killed\" probably have a relatively high apriori likelihood of being ca-sually related. However, knowing that the connective \"because\" evokes a contingency discourse relation between the text spans \"The police arrested him\" and \"he killed someone\" provides further evidence towards predicting causality. The contributions of this paper are summarized below:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 Our focus is on identifying causality between event pairs in context. Since events are often triggered by either verbs (e.g. \"attack\") or nouns (e.g. \"explosion\"), we allow for detection of causality between verb-verb, verb-noun, and noun-noun triggered event pairs. To the best of our knowledge, this formulation of the task is novel.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 We developed a minimally supervised approach for the task using focused distributional similarity methods that are automatically collected from an unannotated corpus. We show that our approach achieves better performance than two approaches: one based on a frequently used metric that measures association, and another based on the effect-control-dependency (ECD) metric described in a prior work (Riaz and Girju, 2010) .", |
| "cite_spans": [ |
| { |
| "start": 399, |
| "end": 421, |
| "text": "(Riaz and Girju, 2010)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 We leverage on the interactions between event causality prediction and discourse relations prediction. We combine these knowledge sources through a global inference procedure, which we formalize via an Integer Linear Programming (ILP) framework as a constraint optimization problem (Roth and Yih, 2004) . This allows us to easily define appropriate constraints to ensure that the causality and discourse predictions are coherent with each other, thereby improving the performance of causality identification.", |
| "cite_spans": [ |
| { |
| "start": 284, |
| "end": 304, |
| "text": "(Roth and Yih, 2004)", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this work, we define an event as an action or occurrence that happens with associated participants or arguments. Formally, we define an event e as: p(a 1 , a 2 , . . . , a n ), where the predicate p is the word that triggers the presence of e in text, and a 1 , a 2 , . . . , a n are the arguments associated with e. Examples of predicates could be verbs such as \"attacked\", \"employs\", nouns such as \"explosion\", \"protest\", etc., and examples of the arguments of \"attacked\" could be its subject and object nouns.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Event Causality", |
| "sec_num": "2" |
| }, |
| { |
| "text": "To measure the causality association between a pair of events e i and e j (in general, e i and e j could be extracted from the same or different documents), we should use information gathered about their predicates and arguments. A simple approach would be to directly calculate the pointwise mutual information (PMI) 1 between p i (a i 1 , a i 2 , . . . , a i n ) and p j (a j 1 , a j 2 , . . . , a j m ). However, this leads to very sparse counts as the predicate p i with its list of arguments a i 1 , . . . , a i n would rarely co-occur (within some reasonable context distance) with predicate p j and its entire list of arguments a j 1 , . . . , a j m . Hence, in this work, we measure causality association using three separate components and focused distributional similarity methods collected about event pairs as described in the rest of this section.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Event Causality", |
| "sec_num": "2" |
| }, |
| { |
| "text": "We measure the causality or cause-effect association (CEA) between two events e i and e j using the following equation:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Cause-Effect Association", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "CEA(e i , e j ) = s pp (e i , e j ) + s pa (e i , e j ) + s aa (e i , e j ) (1)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Cause-Effect Association", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "where s pp measures the association between event predicates, s pa measures the association between the predicate of an event and the arguments of the other event, and s aa measures the association between event arguments. In our work, we regard each event e as being triggered and rooted at a predicate p.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Cause-Effect Association", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "We define s pp as follows:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Predicate-Predicate Association", |
| "sec_num": "2.1.1" |
| }, |
| { |
| "text": "s pp (e i , e j ) = P M I(p i , p j ) \u00d7 max(u i , u j ) \u00d7IDF (p i , p j ) \u00d7 Dist(p i , p j ) (2)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Predicate-Predicate Association", |
| "sec_num": "2.1.1" |
| }, |
| { |
| "text": "which takes into account the PMI between predicates p i and p j of events e i and e j respectively, as well as various other pieces of information. In Suppes' Probabilistic theory of Casuality (Suppes, 1970) , he highlighted that event e is a possible cause of event e , if e happens more frequently with e than by itself, i.e. P (e |e) > P (e ). This can be easily rewritten as P (e,e ) P (e)P (e ) > 1, similar to the definition of PMI: P M I(e, e ) = log P (e, e ) P (e)P (e ) which is only positive when P (e,e ) P (e)P (e ) > 1. Next, we build on the intuition that event predicates appearing in a large number of documents are probably not important or discriminative. Thus, we penalize these predicates when calculating s pp by adopting the inverse document frequency (idf):", |
| "cite_spans": [ |
| { |
| "start": 193, |
| "end": 207, |
| "text": "(Suppes, 1970)", |
| "ref_id": "BIBREF19" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Predicate-Predicate Association", |
| "sec_num": "2.1.1" |
| }, |
| { |
| "text": "IDF (p i , p j ) = idf (p i ) \u00d7 idf (p j ) \u00d7 idf (p i , p j ), where idf (p) = log D", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Predicate-Predicate Association", |
| "sec_num": "2.1.1" |
| }, |
| { |
| "text": "1+N , D is the total number of documents in the collection and N is the number of documents that p occurs in.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Predicate-Predicate Association", |
| "sec_num": "2.1.1" |
| }, |
| { |
| "text": "We also award event pairs that are closer together, while penalizing event pairs that are further apart in texts, by incorporating the distance measure of Leacock and Chodorow (1998), which was originally used to measure similarity between concepts:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Predicate-Predicate Association", |
| "sec_num": "2.1.1" |
| }, |
| { |
| "text": "Dist(p i , p j ) = \u2212log |sent(p i ) \u2212 sent(p j )| + 1 2 \u00d7 ws ,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Predicate-Predicate Association", |
| "sec_num": "2.1.1" |
| }, |
| { |
| "text": "where sent(p) gives the sentence number (index) in which p occurs and ws indicates the window-size (of sentences) used. If p i and p j are drawn from the same sentence, the numerator of the above fraction will return 1. In our work, we set ws to 3 and thus, if p i occurs in sentence k, the furthest sentence that p j will be drawn from, is sentence k + 2. The final component of Equation 2, max(u i , u j ), takes into account whether predicates (events) p i and p j appear most frequently with each other. u i and u j are defined as follows:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Predicate-Predicate Association", |
| "sec_num": "2.1.1" |
| }, |
| { |
| "text": "u i = P (p i , p j ) max k [P (p i , p k )] \u2212 P (p i , p j ) + u j = P (p i , p j ) max k [P (p k , p j )] \u2212 P (p i , p j ) + ,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Predicate-Predicate Association", |
| "sec_num": "2.1.1" |
| }, |
| { |
| "text": "where we set = 0.01 to avoid zeros in the denominators. u i will be maximized if there is no other predicate p k having a higher co-occurrence probability with p i , i.e. p k = p j . u j is treated similarly.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Predicate-Predicate Association", |
| "sec_num": "2.1.1" |
| }, |
| { |
| "text": "We define s pa as follows:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Predicate-Argument and Argument-Argument Association", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "s pa (e i , e j ) = 1 |A e j | a\u2208Ae j P M I(p i , a) + 1 |A e i | a\u2208Ae i P M I(p j , a),", |
| "eq_num": "(3)" |
| } |
| ], |
| "section": "Predicate-Argument and Argument-Argument Association", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "where A e i and A e j are the sets of arguments of e i and e j respectively. Finally, we define s aa as follows:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Predicate-Argument and Argument-Argument Association", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "s aa (e i , e j ) = 1 |A e i ||A e j | a\u2208Ae i a \u2208Ae j P M I(a, a )", |
| "eq_num": "(4)" |
| } |
| ], |
| "section": "Predicate-Argument and Argument-Argument Association", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "Together, s pa and s aa provide additional contexts and robustness (in addition to s pp ) for measuring the cause-effect association between events e i and e j .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Predicate-Argument and Argument-Argument Association", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "Our formulation of CEA is inspired by the ECD metric defined in (Riaz and Girju, 2010) :", |
| "cite_spans": [ |
| { |
| "start": 64, |
| "end": 86, |
| "text": "(Riaz and Girju, 2010)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Predicate-Argument and Argument-Argument Association", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "ECD(a, b) = max(v, w) \u00d7 \u2212log dis(a, b) 2 \u00d7 maxDistance ,", |
| "eq_num": "(5)" |
| } |
| ], |
| "section": "Predicate-Argument and Argument-Argument Association", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "where", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Predicate-Argument and Argument-Argument Association", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "v = P (a, b) P (b) \u2212 P (a, b) + \u00d7 P (a, b) max t [P (a, b t )] \u2212 P (a, b) + w= P (a, b) P (a) \u2212 P (a, b) + \u00d7 P (a, b) max t [P (a t , b)] \u2212 P (a, b) + ,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Predicate-Argument and Argument-Argument Association", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "where ECD(a,b) measures the causality between two events a and b (headed by verbs), and the second component in the ECD equation is similar to Dist(p i , p j ). In our experiments, we will evaluate the performance of ECD against our proposed approach. So far, our definitions in this section are generic and allow for any list of event argument types. In this work, we focus on two argument types: agent (subject) and patient (object), which are typical core arguments of any event. We describe how we extract event predicates and their associated arguments in the section below.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Predicate-Argument and Argument-Argument Association", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "We consider that events are not only triggered by verbs but also by nouns. For a verb (verbal predicate), we extract its subject and object from its associated dependency parse. On the other hand, since events are also frequently triggered by nominal predicates, it is important to identify an appropriate list of event triggering nouns. In our work, we gathered such a list using the following approach:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Verbal and Nominal Predicates", |
| "sec_num": "3" |
| }, |
| { |
| "text": "\u2022 We first gather a list of deverbal nouns from the set of most frequently occurring (in the Gigaword corpus) 3,000 verbal predicate types. For each verb type v, we go through all its Word-Net 2 senses and gather all its derivationally related nouns N v 3 .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Verbal and Nominal Predicates", |
| "sec_num": "3" |
| }, |
| { |
| "text": "\u2022 From N v , we heuristically remove nouns that are less than three characters in length. We also remove nouns whose first three characters are different from the first three characters of v. For each of the remaining nouns in N v , we measured its Levenstein (edit) distance from v and keep the noun(s) with the minimum distance. When multiple nouns have the same minimum distance from v, we keep all of them.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Verbal and Nominal Predicates", |
| "sec_num": "3" |
| }, |
| { |
| "text": "\u2022 To further prune the list of nouns, we next removed all nouns ending in \"er\", \"or\", or \"ee\", as these nouns typically refer to a person, e.g. \"writer\", \"doctor\", \"employee\". We also remove nouns that are not hyponyms (children) of the first WordNet sense of the noun \"event\" 4 .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Verbal and Nominal Predicates", |
| "sec_num": "3" |
| }, |
| { |
| "text": "\u2022 Since we are concerned with nouns denoting events, FrameNet (Ruppenhofer et al., 2010) (FN) is a good resource for mining such nouns. FN consists of frames denoting situations and events. As part of the FN resource, each FN frame consists of a list of lexical units (mainly verbs and nouns) representing the semantics of the frame. Various frame-to-frame relations are also defined (in particular the inheritance relation). Hence, we gathered all the children frames of the FN frame \"Event\". From these children frames, we then gathered all their noun lexical units (words) and add them to our list of nouns. Finally, we also add a few nouns denoting natural disaster from Wikipedia 5 .", |
| "cite_spans": [ |
| { |
| "start": 62, |
| "end": 88, |
| "text": "(Ruppenhofer et al., 2010)", |
| "ref_id": "BIBREF17" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Verbal and Nominal Predicates", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Using the above approach, we gathered a list of about 2,000 noun types. This current approach is heuristics based which we intend to improve in the future, and any such improvements should subsequently improve the performance of our causality identification approach.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Verbal and Nominal Predicates", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Event triggering deverbal nouns could have associated arguments (for instance, acting as subject, object of the deverbal noun). To extract these arguments, we followed the approach of (Gurevich et al., 2008) . Briefly, the approach uses linguistic patterns to extract subjects and objects for deverbal nouns, using information from dependency parses. For more details, we refer the reader to (Gurevich et al., 2008) .", |
| "cite_spans": [ |
| { |
| "start": 184, |
| "end": 207, |
| "text": "(Gurevich et al., 2008)", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 392, |
| "end": 415, |
| "text": "(Gurevich et al., 2008)", |
| "ref_id": "BIBREF7" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Verbal and Nominal Predicates", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Discourse connectives are important for relating different text spans, helping us to understand a piece of text in relation to its context: In the example sentence above, the discourse connective (\"because\") and the discourse relation it evokes (in this case, the Cause relation) allows readers to relate its two associated text spans, \"The police arrested him\" and \"he killed someone\". Also, notice that the verbs \"arrested\" and \"killed\", which cross the two text spans, are causally related. To aid in extracting causal relations, we leverage on the identification of discourse relations to provide additional contextual information.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discourse and Causality", |
| "sec_num": "4" |
| }, |
| { |
| "text": "To identify discourse relations, we use the Penn Discourse Treebank (PDTB) (Prasad et al., 2007) , which contains annotations of discourse relations in context. The annotations are done over the Wall Street Journal corpus and the PDTB adopts a predicate-argument view of discourse relations. A discourse connective (e.g. because) takes two text spans as its arguments. In the rest of this section, we briefly describe the discourse relations in PDTB and highlight how we might leverage them to aid in determining event causality. According to the PDTB annotation manual (Prasad et al., 2007) , the truth of both spans is independent of the established discourse relation. This means that the text spans are not causally related and thus, the existence of a Comparison relation should imply that there is no causality relation across the two text spans.", |
| "cite_spans": [ |
| { |
| "start": 75, |
| "end": 96, |
| "text": "(Prasad et al., 2007)", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 570, |
| "end": 591, |
| "text": "(Prasad et al., 2007)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discourse and Causality", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Contingency A Contingency relation between two text spans indicates that the situation described in one text span causally influences the situation in the other. An example sentence is: Existence of a Contingency relation potentially implies that there exists at least one causal event pair crossing the two text spans. The PDTB annotation manual states that while the Cause and Condition discourse relations indicate casual influence in their text spans, there is no causal influence in the text spans of the Pragmatic-cause and Pragmatic-condition relations. For instance, Pragmatic-condition indicates that one span pro-vides the context in which the description of the situation in the other span is relevant; for example: Hence, there is a need to also identify fine-grained discourse relations. Most of the Expansion fine-grained relations (except for Conjunction, which could connect arbitrary pieces of text spans) should not contain causality relations across its text spans.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discourse and Causality", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Temporal These indicate that the situations described in the text spans are related temporally. An example sentence is: Temporal precedence of the (cause) event over the (effect) event is a necessary, but not sufficient requisite for causality. Hence by itself, Temporal relations are probably not discriminative enough for determining event causality.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discourse and Causality", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Our work follows the approach and features described in the state-of-the-art Ruby-based discourse system of (Lin et al., 2010) , to build an inhouse Java-based discourse relation extraction system. Our system identifies explicit connectives in text, predict their discourse relations, as well as their associated text spans. Similar to (Lin et al., 2010) , we achieved a competitive performance of slightly over 80% F1-score in identifying fine-grained relations for explicit connectives. Our system is developed using the Learning Based Java modeling lan-guage (LBJ) (Rizzolo and Roth, 2010) and will be made available soon. Due to space constraints, we refer interested readers to (Lin et al., 2010) for details on the features, etc.", |
| "cite_spans": [ |
| { |
| "start": 108, |
| "end": 126, |
| "text": "(Lin et al., 2010)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 336, |
| "end": 354, |
| "text": "(Lin et al., 2010)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 568, |
| "end": 592, |
| "text": "(Rizzolo and Roth, 2010)", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 683, |
| "end": 701, |
| "text": "(Lin et al., 2010)", |
| "ref_id": "BIBREF10" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discourse Relation Extraction System", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "In the example sentences given thus far in this section, all the connectives were explicit, as they appear in the texts. PDTB also provides annotations for implicit connectives, which we do not use in this work. Identifying implicit connectives is a harder task and incorporating these is a possible future work.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discourse Relation Extraction System", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "To exploit the interactions between event pair causality extraction and discourse relation identification, we define appropriate constraints between them, which can be enforced through the Constrained Conditional Models framework (aka ILP for NLP) (Roth and Yih, 2007; Chang et al., 2008) . In doing this, the predictions of CEA (Section 2.1) and the discourse system are forced to cohere with each other. More importantly, this should improve the performance of using only CEA to extract causal event pairs. To the best of our knowledge, this approach for causality extraction is novel.", |
| "cite_spans": [ |
| { |
| "start": 248, |
| "end": 268, |
| "text": "(Roth and Yih, 2007;", |
| "ref_id": "BIBREF16" |
| }, |
| { |
| "start": 269, |
| "end": 288, |
| "text": "Chang et al., 2008)", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Joint Inference for Causality Extraction", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Let E denote the set of event mentions in a document. Let EP = {(e i , e j ) \u2208 E \u00d7 E | e i \u2208 E, e j \u2208 E, i < j, |sent(e i ) \u2212 sent(e j )| \u2264 2} denote the set of event mention pairs in the document, where sent(e) gives the sentence number in which event e occurs. Note that in this work, we only extract event pairs that are at most two sentences apart. Next, we define L ER = {\"causal\", \"\u00ac causal\"} to be the set of event relation labels that an event pair ep \u2208 EP can be associated with.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "CEA & Discourse: Implementation Details", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "Note that the CEA metric as defined in Section 2.1 simply gives a score without it being bounded to be between 0 and 1.0. However, to use the CEA score as part of the inference process, we require that it be bounded and thus can be used as a binary prediction, that is, predicting an event pair as causal or \u00accausal.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "CEA & Discourse: Implementation Details", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "To enable this, we use a few development documents to automatically find a threshold CEA score that separates scores indicating causal vs \u00accausal. Based on this threshold, the original CEA scores are then rescaled to fall within 0 to 1.0. More details on this are in Section 6.2.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "CEA & Discourse: Implementation Details", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "Let C denote the set of connective mentions in a document. We slightly modify our discourse system as follows. We define L DR to be the set of discourse relations. We initially add all the finegrained discourse relations listed in Table 1 to L DR . In the PDTB corpus, some connective examples are labeled with just a coarse-grained relation, without further specifying a fine-grained relation. To accommodate these examples, we add the coarsegrained relations Comparison, Expansion, and Temporal to L DR . We omit the coarse-grained Contingency relation from L DR , as we want to separate Cause and Condition from Pragmatic-cause and Pragmatic-condition. This discards very few examples as only a very small number of connective examples are simply labeled with a Contingency label without further specifying a fine-grained label. We then retrained our discourse system to predict labels in L DR .", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 231, |
| "end": 238, |
| "text": "Table 1", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "CEA & Discourse: Implementation Details", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "We now describe the constraints used to support joint inference, based on the predictions of the CEA metric and the discourse classifier. Let s c (dr) be the probability that connective c is predicated to be of discourse relation dr, based on the output of our discourse classifier. Let s ep (er) be the CEA prediction score (rescaled to range in [0,1]) that event pair ep takes on the causal or \u00accausal label er. Let x c,dr be a binary indicator variable which takes on the value 1 iff c is labeled with the discourse relation dr. Similarly, let y ep,er be a binary variable which takes on the value 1 iff ep is labeled as er. We then define our objective function as follows:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Constraints", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "max |L DR | c\u2208C dr\u2208L DR s c (dr) \u2022 x c,dr +|L ER | ep\u2208EP er\u2208L ER s ep (er) \u2022 y ep,er (6)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Constraints", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "subject to the following constraints:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Constraints", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "dr\u2208L DR x c,dr = 1 \u2200c \u2208 C", |
| "eq_num": "(7)" |
| } |
| ], |
| "section": "Constraints", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "er\u2208L ER y ep,er = 1 \u2200ep \u2208 EP (8)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Constraints", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "x c,dr \u2208 {0, 1} \u2200c \u2208 C, dr \u2208 L DR (9) y ep,er \u2208 {0, 1} \u2200ep \u2208 EP, er \u2208 L ER (10)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Constraints", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "Equation 7requires that each connective c can only be assigned one discourse relation. Equation (8) requires that each event pair ep can only be causal or \u00accausal. Equations (9) and (10) indicate that x c,dr and y ep,er are binary variables.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Constraints", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "To capture the relationship between event pair causality and discourse relations, we use the following constraints:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Constraints", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "x c,\"Cause\" \u2264 ep\u2208EPc y ep,\"causal\" (11)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Constraints", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "x c,\"Condition\" \u2264 ep\u2208EPc y ep,\"causal\" , 12where both equations are defined \u2200c \u2208 C. EP c is defined to be the set of event pairs that cross the two text spans associated with c. For instance, if the first text span of c contains two event mentions e i , e j , and there is one event mention e k in the second text span of c, then EP c = {(e i , e k ), (e j , e k )}. Finally, the logical form of Equation (11) can be written as:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Constraints", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "x c,\"Cause\" \u21d2 y ep i ,\"causal\" \u2228 . . . \u2228 y ep j ,\"causal\" , where ep i , . . . , ep j are elements in EP c . This states that if we assign the Cause discourse label to c, then at least one of ep i , . . . , ep j must be assigned as causal. The interpretation of Equation 12is similar. We use two more constraints to capture the interactions between event causality and discourse relations. First, we defined C ep as the set of connectives c enclosing each event of ep in each of its text spans, i.e.: one of the text spans of c contain one of the event in ep, while the other text span of c contain the other event in ep. Next, based on the discourse relations in Section 4.1, we propose that when an event pair ep is judged to be causal, then the connective c that encloses it should be evoking one of the discourse relations in L DRa = {\"Cause\", \"Condition\", \"Temporal\", \"Asynchronous\", \"Synchrony\", \"Conjunction\"}. We capture this using the following constraint:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Constraints", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "y ep,\"causal\" \u2264 dra\u2208L DRa x c,dra \u2200c \u2208 C ep (13)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Constraints", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "The logical form of Equation (13) can be written as: y ep,\"causal\" \u21d2 x c,\"Cause\" \u2228 x c,\"Condition\" . . . \u2228 x c,\"Conjunction\" . This states that if we assign ep as causal, then we must assign to c one of the labels in L DRa .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Constraints", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "Finally, we propose that for any connectives evoking discourse relations L DR b = {\"Comparison\", \"Concession\", \"Contrast\", \"Pragmatic-concession\", \"Pragmatic-contrast\", \"Expansion\", \"Alternative\", \"Exception\", \"Instantiation\", \"List\", \"Restatement\"}, any event pair(s) that it encloses should be \u00accausal. We capture this using the following constraint:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Constraints", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "x c,dr b \u2264 y ep,\"\u00accausal\" \u2200 dr b \u2208 L DR b , ep \u2208 EP c ,", |
| "eq_num": "(14)" |
| } |
| ], |
| "section": "Constraints", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "where the logical form of Equation 14can be written as: x c,dr b \u21d2 y ep,\"\u00accausal\" .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Constraints", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "To collect the distributional statistics for measuring CEA as defined in Equation 1, we applied partof-speech tagging, lemmatization, and dependency parsing (Marneffe et al., 2006) on about 760K documents in the English Gigaword corpus (LDC catalog number LDC2003T05).", |
| "cite_spans": [ |
| { |
| "start": 157, |
| "end": 180, |
| "text": "(Marneffe et al., 2006)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments 6.1 Experimental Settings", |
| "sec_num": "6" |
| }, |
| { |
| "text": "We are not aware of any benchmark corpus for evaluating event causality extraction in contexts. Hence, we created an evaluation corpus using the following process: Using news articles collected from CNN 7 during the first three months of 2010, we randomly selected 20 articles (documents) as evaluation data, and 5 documents as development data.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments 6.1 Experimental Settings", |
| "sec_num": "6" |
| }, |
| { |
| "text": "Two annotators annotated the documents for causal event pairs, using two simple notions for causality: the Cause event should temporally precede the Effect event, and the Effect event occurs because the Cause event occurs. However, sometimes it is debatable whether two events are involved in a causal relation, or whether they are simply involved in an uninteresting temporal relation. Hence, we allowed annotations of C to indicate causality, and R to indicate relatedness (for situations when the existence of causality is debatable). The annotators will simply identify and annotate the C or R relations between predicates of event pairs. Event arguments are not explicitly annotated, although the annotators are free to look at the entire document text while making their annotation decisions. Finally, they are free to annotate relations between predicates that have any number of sentences in between and are not restricted to a fixed sentence window-size. After adjudication, we obtained a total of 492 C +R relation annotations, and 414 C relation annotations on the evaluation documents. On the development documents, we obtained 92 C + R and 71 C relation annotations. The annotators overlapped on 10 evaluation documents. On these documents, the first (second) annotator annotated 215 (199) C + R relations, agreeing on 166 of these relations. Together, they annotated 248 distinct relations. Using this number, their agreement ratio would be 0.67 (166/248). The corresponding agreement ratio for C relations is 0.58. These numbers highlight that causality identification is a difficult task, as there could be as many as N 2 event pairs in a document (N is the number of events in the document). We plan to make this annotated dataset available soon. 8", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments 6.1 Experimental Settings", |
| "sec_num": "6" |
| }, |
| { |
| "text": "As mentioned in Section 5.1, to enable translating (the unbounded) CEA scores into binary causal, \u00accausal predictions, we need to rescale or calibrate these scores to range in [0, 1] . To do this, we first rank all the CEA scores of all event pairs in the development documents. Most of these event pairs will be \u00accausal. Based on the relation annotations in these development documents, we scanned through this ranked list of scores to locate the CEA score t that gives the highest F1-score (on the development documents) when used as a threshold between causal vs \u00accausal decisions. We then ranked all the CEA scores of all event pairs gathered from the 760K Gigaword documents, discretized all scores higher than t into B bins, and all scores lower than t into B bins. Together, these 2B bins represent the range [0, 1] . We used B = 500. Thus, consecutive bins represent a difference of 0.001 in calibrated scores.", |
| "cite_spans": [ |
| { |
| "start": 176, |
| "end": 179, |
| "text": "[0,", |
| "ref_id": null |
| }, |
| { |
| "start": 180, |
| "end": 182, |
| "text": "1]", |
| "ref_id": null |
| }, |
| { |
| "start": 816, |
| "end": 819, |
| "text": "[0,", |
| "ref_id": null |
| }, |
| { |
| "start": 820, |
| "end": 822, |
| "text": "1]", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "6.2" |
| }, |
| { |
| "text": "To measure the causality between a pair of events e i and e j , a simple baseline is to calculate P M I(p i , p j ). Using a similar thresholding and calibration process to translate P M I(p i , p j ) scores into binary causality decisions, we obtained a F1 score of 23.1 when measured over the causality C relations, as shown in the row P M I pp of Table 2 .", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 350, |
| "end": 357, |
| "text": "Table 2", |
| "ref_id": "TABREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "6.2" |
| }, |
| { |
| "text": "As mentioned in Section 2.1.2, Riaz and Girju (2010) proposed the ECD metric to measure causality between two events. Thus, as a point of comparison, we replaced s pp of Equation (1) with ECD(a, b) of Equation (5), substituting a = p i and b = p j . After thresholding and calibrating the scores of this approach, we obtained a F1-score of 29.7, as shown in the row ECD pp &P M I pa,aa of Table 2 .", |
| "cite_spans": [ |
| { |
| "start": 31, |
| "end": 52, |
| "text": "Riaz and Girju (2010)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 389, |
| "end": 396, |
| "text": "Table 2", |
| "ref_id": "TABREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "6.2" |
| }, |
| { |
| "text": "Next, we evaluated our proposed CEA approach and obtained a F1-score of 38.6, as shown in the row CEA of Table 2 . Thus, our proposed approach obtained significantly better performance than the PMI baseline and the ECD approach. Next, we performed joint inference with the discourse relation predictions as described in Section 5 and obtained an improved F1-score of 41.7. We note that we obtained improvements in both recall and precision. This means that with the aid of discourse relations, we are able to recover more causal relations, as well as reduce false-positive predictions.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 105, |
| "end": 112, |
| "text": "Table 2", |
| "ref_id": "TABREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "6.2" |
| }, |
| { |
| "text": "Constraint Equations (11) and (12) help to recover causal relations. For improvements in precision, as stated in the last paragraph of Section 5.2, identifying other discourse relations such as \"Comparison\", \"Contrast\", etc., provides counterevidence to causality. Together with constraint Equation 14, this helps to eliminate false-positive event pairs as classified by CEA and contributes towards CEA+Discourse having a higher precision than CEA.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "6.2" |
| }, |
| { |
| "text": "The corresponding results for extracting both causality and relatedness C + R relations are given in Table 3 . For these experiments, the aim was for a more relaxed evaluation and we simply collapsed C and R into a single label.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 101, |
| "end": 108, |
| "text": "Table 3", |
| "ref_id": "TABREF5" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "6.2" |
| }, |
| { |
| "text": "Finally, we also measured the precision of the top K causality C predictions, showing the precision trends in Figure 1 . As shown, CEA in general achieves higher precision when compared to P M I pp and ECD pp &P M I pa,aa . The trends for C + R predictions are similar.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 110, |
| "end": 118, |
| "text": "Figure 1", |
| "ref_id": "FIGREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "6.2" |
| }, |
| { |
| "text": "Thus far, we had included both verbal and nominal predicates in our evaluation. When we repeat the experiments for ECD pp &P M I pa,aa and CEA on just verbal predicates, we obtained the respective F1-scores of 31.8 and 38.3 on causality relations. The corresponding F1-scores for casuality and relatedness relations are 35.7 and 43.3. These absolute F1-scores are similar to those in Tables 2 and 3, differing by 1-2%.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "6.2" |
| }, |
| { |
| "text": "We randomly selected 50 false-positive predictions and 50 false-negative causality relations to analyze the mistakes made by CEA.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Analysis", |
| "sec_num": "7" |
| }, |
| { |
| "text": "Among the false-positives (precision errors), the most frequent error type (56% of the errors) is that CEA simply assigns a high score to event pairs that are not causal; more knowledge sources are required to support better predictions in these cases. The next largest group of error (22%) involves events containing pronouns (e.g. \"he\", \"it\") as arguments. Ap-plying coreference to replace these pronouns with their canonical entity strings or labeling them with semantic class information might be useful.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Analysis", |
| "sec_num": "7" |
| }, |
| { |
| "text": "Among the false-negatives (recall errors), 23% of the errors are due to CEA simply assigning a low score to causal event pairs and more contextual knowledge seems necessary for better predictions. 19% of the recall errors arises from causal event pairs involving nominal predicates that are not in our list of event evoking noun types (described in Section 3). A related 17% of recall errors involves nominal predicates without any argument. For these, less information is available for CEA to make predictions. The remaining group (15% of errors) involves events containing pronouns as arguments.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Analysis", |
| "sec_num": "7" |
| }, |
| { |
| "text": "Although prior work in event causality extraction in context is relatively sparse, there are many prior works concerning other semantic aspects of event extraction. Ji and Grishman (2008) extracts event mentions (belonging to a predefined list of target event types) and their associated arguments. In other prior work (Chen et al., 2009; Bejan and Harabagiu, 2010) , the authors focused on identifying another type of event pair semantic relation: event coreference. Chambers and Jurafsky (2008; chain events sharing a common (protagonist) participant. They defined events as verbs and given an existing chain of events, they predict the next likely event involving the protagonist. This is different from our task of detecting causality between arbitrary event pairs that might or might not share common arguments. Also, we defined events more broadly, as those that are triggered by either verbs or nouns. Finally, although our proposed CEA metric has resemblance the ECD metric in (Riaz and Girju, 2010) , our task is different from theirs and our work differs in many aspects. They focused on building a dataset of causal text spans, whereas we focused on identifying causal relations between events in a given text document. They considered text spans headed by verbs while we considered events triggered by both verbs and nouns. Moreover, we combined event causality prediction and discourse relation prediction through a global inference procedure to further improve the performance of event causality prediction.", |
| "cite_spans": [ |
| { |
| "start": 165, |
| "end": 187, |
| "text": "Ji and Grishman (2008)", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 319, |
| "end": 338, |
| "text": "(Chen et al., 2009;", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 339, |
| "end": 365, |
| "text": "Bejan and Harabagiu, 2010)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 468, |
| "end": 496, |
| "text": "Chambers and Jurafsky (2008;", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 985, |
| "end": 1007, |
| "text": "(Riaz and Girju, 2010)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "8" |
| }, |
| { |
| "text": "In this paper, using general tools such as the dependency and discourse parsers which are not trained specifically towards our target task, and a minimal set of development documents for threshold tuning, we developed a minimally supervised approach to identify causality relations between events in context. We also showed how to incorporate discourse relation predictions to aid event causality predictions through a global inference procedure. There are several interesting directions for future work, including the incorporation of other knowledge sources such as coreference and semantic class predictions, which were shown to be potentially important in our error analysis. We could also use discourse relations to aid in extracting other semantic relations between events.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "9" |
| }, |
| { |
| "text": "PMI is frequently used to measure association between variables.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "http://wordnet.princeton.edu/ 3 The WordNet resource provides derivational information on words that are in different syntactic (i.e. part-of-speech) categories, but having the same root (lemma) form and that are semantically related.4 The first WordNet sense of the noun \"event\" has the meaning: \"something that happens at a given place and time\"", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "http://en.wikipedia.org/wiki/Natural disaster", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "PDTB further refines these fine-grained relations into a final third level of relations, but we do not use them in this work.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "http://www.cnn.com", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "The authors thank the anonymous reviewers for their insightful comments and suggestions. University of Illinois at Urbana-Champaign gratefully acknowledges the support of Defense Advanced Research Projects Agency (DARPA) Machine Reading Program under Air Force Research Laboratory (AFRL) prime contract No. FA8750-09-C-0181. The first author thanks the Vietnam Education Foundation (VEF) for its sponsorship. Any opinions, findings, and conclusion or recommendations expressed in this material are those of the authors and do not necessarily reflect the view of the VEF, DARPA, AFRL, or the US government.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgments", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Using a bigram event model to predict causal potential", |
| "authors": [ |
| { |
| "first": "Brandon", |
| "middle": [], |
| "last": "Beamer", |
| "suffix": "" |
| }, |
| { |
| "first": "Roxana", |
| "middle": [], |
| "last": "Girju", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "CI-CLING", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Brandon Beamer and Roxana Girju. 2009. Using a bi- gram event model to predict causal potential. In CI- CLING.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Unsupervised event coreference resolution with rich linguistic features", |
| "authors": [ |
| { |
| "first": "Adrian", |
| "middle": [], |
| "last": "Cosmin", |
| "suffix": "" |
| }, |
| { |
| "first": "Sanda", |
| "middle": [], |
| "last": "Bejan", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Harabagiu", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Cosmin Adrian Bejan and Sanda Harabagiu. 2010. Un- supervised event coreference resolution with rich lin- guistic features. In ACL.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Unsupervised learning of narrative event chains", |
| "authors": [ |
| { |
| "first": "Nathanael", |
| "middle": [], |
| "last": "Chambers", |
| "suffix": "" |
| }, |
| { |
| "first": "Dan", |
| "middle": [], |
| "last": "Jurafsky", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nathanael Chambers and Dan Jurafsky. 2008. Unsuper- vised learning of narrative event chains. In ACL-HLT.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Unsupervised learning of narrative schemas and their participants", |
| "authors": [ |
| { |
| "first": "Nathanael", |
| "middle": [], |
| "last": "Chambers", |
| "suffix": "" |
| }, |
| { |
| "first": "Dan", |
| "middle": [], |
| "last": "Jurafsky", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nathanael Chambers and Dan Jurafsky. 2009. Unsuper- vised learning of narrative schemas and their partici- pants. In ACL.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Learning and inference with constraints", |
| "authors": [ |
| { |
| "first": "Ming-Wei", |
| "middle": [], |
| "last": "Chang", |
| "suffix": "" |
| }, |
| { |
| "first": "Lev", |
| "middle": [], |
| "last": "Ratinov", |
| "suffix": "" |
| }, |
| { |
| "first": "Nicholas", |
| "middle": [], |
| "last": "Rizzolo", |
| "suffix": "" |
| }, |
| { |
| "first": "Dan", |
| "middle": [], |
| "last": "Roth", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "AAAI", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ming-Wei Chang, Lev Ratinov, Nicholas Rizzolo, and Dan Roth. 2008. Learning and inference with con- straints. In AAAI.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "A pairwise event coreference model, feature impact and evaluation for event coreference resolution", |
| "authors": [ |
| { |
| "first": "Zheng", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Ji", |
| "middle": [], |
| "last": "Heng", |
| "suffix": "" |
| }, |
| { |
| "first": "Robert", |
| "middle": [], |
| "last": "Haralick", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "RANLP workshop on Events in Emerging Text Types", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Zheng Chen, Heng Ji, and Robert Haralick. 2009. A pairwise event coreference model, feature impact and evaluation for event coreference resolution. In RANLP workshop on Events in Emerging Text Types.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Automatic detection of causal relations for question answering", |
| "authors": [ |
| { |
| "first": "Roxana", |
| "middle": [], |
| "last": "Girju", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "ACL workshop on Multilingual Summarization and Question Answering", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Roxana Girju. 2003. Automatic detection of causal re- lations for question answering. In ACL workshop on Multilingual Summarization and Question Answering.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Deverbal nouns in knowledge representation", |
| "authors": [ |
| { |
| "first": "Olga", |
| "middle": [], |
| "last": "Gurevich", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Crouch", |
| "suffix": "" |
| }, |
| { |
| "first": "Tracy", |
| "middle": [ |
| "Holloway" |
| ], |
| "last": "King", |
| "suffix": "" |
| }, |
| { |
| "first": "Valeria", |
| "middle": [], |
| "last": "De Paiva", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Journal of Logic and Computation", |
| "volume": "18", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Olga Gurevich, Richard Crouch, Tracy Holloway King, and Valeria de Paiva. 2008. Deverbal nouns in knowl- edge representation. Journal of Logic and Computa- tion, 18, June.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Refining event extraction through unsupervised cross-document inference", |
| "authors": [ |
| { |
| "first": "Heng", |
| "middle": [], |
| "last": "Ji", |
| "suffix": "" |
| }, |
| { |
| "first": "Ralph", |
| "middle": [], |
| "last": "Grishman", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Heng Ji and Ralph Grishman. 2008. Refining event ex- traction through unsupervised cross-document infer- ence. In ACL.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Combining Local Context and WordNet Similarity for Word Sense Identification", |
| "authors": [ |
| { |
| "first": "Claudia", |
| "middle": [], |
| "last": "Leacock", |
| "suffix": "" |
| }, |
| { |
| "first": "Martin", |
| "middle": [], |
| "last": "Chodorow", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Claudia Leacock and Martin Chodorow, 1998. Combin- ing Local Context and WordNet Similarity for Word Sense Identification. MIT Press.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "A pdtb-styled end-to-end discourse parser", |
| "authors": [ |
| { |
| "first": "Ziheng", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| }, |
| { |
| "first": "Min-Yen", |
| "middle": [], |
| "last": "Hwee Tou Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Kan", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ziheng Lin, Hwee Tou Ng, and Min-Yen Kan. 2010. A pdtb-styled end-to-end discourse parser. Tech- nical report. http://www.comp.nus.edu.sg/ linzi- hen/publications/tech2010.pdf.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Generating typed dependency parses from phrase structure parses", |
| "authors": [ |
| { |
| "first": "Bill", |
| "middle": [], |
| "last": "Marie-Catherine De Marneffe", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [ |
| "D" |
| ], |
| "last": "Maccartney", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Manning", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "LREC", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Marie-catherine De Marneffe, Bill Maccartney, and Christopher D. Manning. 2006. Generating typed dependency parses from phrase structure parses. In LREC.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "The penn discourse treebank 2.0 annotation manual", |
| "authors": [ |
| { |
| "first": "Rashmi", |
| "middle": [], |
| "last": "Prasad", |
| "suffix": "" |
| }, |
| { |
| "first": "Eleni", |
| "middle": [], |
| "last": "Miltsakaki", |
| "suffix": "" |
| }, |
| { |
| "first": "Nikhil", |
| "middle": [], |
| "last": "Dinesh", |
| "suffix": "" |
| }, |
| { |
| "first": "Alan", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Aravind", |
| "middle": [], |
| "last": "Joshi", |
| "suffix": "" |
| }, |
| { |
| "first": "Livio", |
| "middle": [], |
| "last": "Robaldo", |
| "suffix": "" |
| }, |
| { |
| "first": "Bonnie", |
| "middle": [], |
| "last": "Webber", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Rashmi Prasad, Eleni Miltsakaki, Nikhil Dinesh, Alan Lee, Aravind Joshi, Livio Robaldo, and Bonnie Webber. 2007. The penn discourse tree- bank 2.0 annotation manual. Technical report.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Another look at causality: Discovering scenario-specific contingency relationships with no supervision", |
| "authors": [ |
| { |
| "first": "Mehwish", |
| "middle": [], |
| "last": "Riaz", |
| "suffix": "" |
| }, |
| { |
| "first": "Roxana", |
| "middle": [], |
| "last": "Girju", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "ICSC", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Mehwish Riaz and Roxana Girju. 2010. Another look at causality: Discovering scenario-specific contingency relationships with no supervision. In ICSC.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Learning based java for rapid development of nlp systems", |
| "authors": [ |
| { |
| "first": "N", |
| "middle": [], |
| "last": "Rizzolo", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Roth", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "LREC", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "N. Rizzolo and D. Roth. 2010. Learning based java for rapid development of nlp systems. In LREC.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "A linear programming formulation for global inference in natural language tasks", |
| "authors": [ |
| { |
| "first": "Dan", |
| "middle": [], |
| "last": "Roth", |
| "suffix": "" |
| }, |
| { |
| "first": "Wen", |
| "middle": [], |
| "last": "Tau", |
| "suffix": "" |
| }, |
| { |
| "first": "Yih", |
| "middle": [], |
| "last": "", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "CoNLL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dan Roth and Wen Tau Yih. 2004. A linear program- ming formulation for global inference in natural lan- guage tasks. In CoNLL.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Global inference for entity and relation identification via a linear programming formulation", |
| "authors": [ |
| { |
| "first": "Dan", |
| "middle": [], |
| "last": "Roth", |
| "suffix": "" |
| }, |
| { |
| "first": "Wen", |
| "middle": [], |
| "last": "Tau", |
| "suffix": "" |
| }, |
| { |
| "first": "Yih", |
| "middle": [], |
| "last": "", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Lise Getoor and Ben Taskar, editors, Introduction to Statistical Relational Learning", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dan Roth and Wen Tau Yih. 2007. Global inference for entity and relation identification via a linear program- ming formulation. In Lise Getoor and Ben Taskar, ed- itors, Introduction to Statistical Relational Learning. MIT Press.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "FrameNet II: Extended Theory and Practice", |
| "authors": [ |
| { |
| "first": "Josef", |
| "middle": [], |
| "last": "Ruppenhofer", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Ellsworth", |
| "suffix": "" |
| }, |
| { |
| "first": "Miriam", |
| "middle": [ |
| "R L" |
| ], |
| "last": "Petruck", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [ |
| "R" |
| ], |
| "last": "Johnson", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Josef Ruppenhofer, Michael Ellsworth, Miriam R. L. Petruck, Christopher R. Johnson, and Jan Scheffczyk. 2010. FrameNet II: Extended Theory and Practice. http://framenet.icsi.berkeley.edu.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Causal relation of queries from temporal logs", |
| "authors": [ |
| { |
| "first": "Yizhou", |
| "middle": [], |
| "last": "Sun", |
| "suffix": "" |
| }, |
| { |
| "first": "Ning", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Kunqing", |
| "middle": [], |
| "last": "Xie", |
| "suffix": "" |
| }, |
| { |
| "first": "Shuicheng", |
| "middle": [], |
| "last": "Yan", |
| "suffix": "" |
| }, |
| { |
| "first": "Benyu", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "Zheng", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "WWW", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yizhou Sun, Ning Liu, Kunqing Xie, Shuicheng Yan, Benyu Zhang, and Zheng Chen. 2007. Causal rela- tion of queries from temporal logs. In WWW.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "A Probabilistic Theory of Causality", |
| "authors": [ |
| { |
| "first": "Patrick", |
| "middle": [], |
| "last": "Suppes", |
| "suffix": "" |
| } |
| ], |
| "year": 1970, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Patrick Suppes. 1970. A Probabilistic Theory of Causal- ity. Amsterdam: North-Holland Publishing Company.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "num": null, |
| "uris": null, |
| "text": "The police arrested him] because [he killed someone].", |
| "type_str": "figure" |
| }, |
| "FIGREF1": { |
| "num": null, |
| "uris": null, |
| "text": "Cause: [The first priority is search and rescue] because [many people are trapped under the rubble].", |
| "type_str": "figure" |
| }, |
| "FIGREF2": { |
| "num": null, |
| "uris": null, |
| "text": "Connectives evoking Expansion discourse relations expand the discourse, such as by providing additional information, illustrating alternative situations, etc. An example sentence is: Conjunction: [Over the past decade, x women were killed] and [y went missing].", |
| "type_str": "figure" |
| }, |
| "FIGREF3": { |
| "num": null, |
| "uris": null, |
| "text": "Synchrony: [He was sitting at his home] when [the whole world started to shake].", |
| "type_str": "figure" |
| }, |
| "FIGREF4": { |
| "num": null, |
| "uris": null, |
| "text": "8 http://cogcomp.cs.illinois.edu/page/publication view/663 Precision of the top K causality C predictions.", |
| "type_str": "figure" |
| }, |
| "TABREF1": { |
| "type_str": "table", |
| "html": null, |
| "text": "Coarse-grained and fine-grained discourse relations.4.1 Discourse RelationsPDTB contains annotations for four coarse-grained discourse relation types, as shown in the left column ofTable 1. Each of these are further refined into several fine-grained discourse relations, as shown in the right column of the table. 6 Next, we briefly describe these relations, highlighting those that could potentially help to determine event causality.", |
| "num": null, |
| "content": "<table/>" |
| }, |
| "TABREF4": { |
| "type_str": "table", |
| "html": null, |
| "text": "Performance of baseline systems and our approaches on extracting Causal event relations.", |
| "num": null, |
| "content": "<table><tr><td>System</td><td colspan=\"2\">Rec% Pre% F1%</td></tr><tr><td colspan=\"2\">27.8 ECD pp &P M I pa,aa 42.4 P M I pp 63.1 CEA CEA+Discourse 65.3</td><td>24.9 26.2 28.5 34.1 33.7 43.9 36.5 46.9</td></tr></table>" |
| }, |
| "TABREF5": { |
| "type_str": "table", |
| "html": null, |
| "text": "Performance of the systems on extracting Causal and Related event relations.", |
| "num": null, |
| "content": "<table/>" |
| } |
| } |
| } |
| } |