| { |
| "paper_id": "U08-1011", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T03:11:53.063113Z" |
| }, |
| "title": "Automatic Event Reference Identification", |
| "authors": [ |
| { |
| "first": "Olivia", |
| "middle": [], |
| "last": "March", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "CSSE University of Melbourne", |
| "location": { |
| "postCode": "3010", |
| "settlement": "Victoria", |
| "country": "Australia" |
| } |
| }, |
| "email": "oliviacm@csse.unimelb.edu.au" |
| }, |
| { |
| "first": "Timothy", |
| "middle": [], |
| "last": "Baldwin", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "NICTA Victoria Research Laboratories", |
| "institution": "University of Melbourne", |
| "location": { |
| "postCode": "3010", |
| "settlement": "Victoria", |
| "country": "Australia" |
| } |
| }, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "Event reference identification is often treated as a sentence level classification task. However, several different event references can occur within a single sentence. We present a set of experiments involving real world event reference identification at the word level in newspaper and newswire documents, addressing the issue of effective text representation for classification of events using support vector machines. Our final system achieved an F-score of 0.764, significantly exceeding that of our baseline system. Additionally we achieved a marginally higher performance than a more complex comparable system.", |
| "pdf_parse": { |
| "paper_id": "U08-1011", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "Event reference identification is often treated as a sentence level classification task. However, several different event references can occur within a single sentence. We present a set of experiments involving real world event reference identification at the word level in newspaper and newswire documents, addressing the issue of effective text representation for classification of events using support vector machines. Our final system achieved an F-score of 0.764, significantly exceeding that of our baseline system. Additionally we achieved a marginally higher performance than a more complex comparable system.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Multi-document summarization seeks to combine the salient information from multiple independent documents into a single coherent summary. The Document Understanding Conference (DUC) shared tasks are a measured attempt to apply multi-document summarization to newswire document sets, to produce a fluent 250 word summary (Dang, 2006) . The resultant summaries are then scored and assessed, depending on the specifics of the task.", |
| "cite_spans": [ |
| { |
| "start": 320, |
| "end": 332, |
| "text": "(Dang, 2006)", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Newswire and newspaper document sets such as those used in DUC often describe the same series of events across multiple sources, for example, media coverage of an election in Bolivia. Here, a time-line can be a more effective way of structuring a multi-document summary than a traditional paragraph-style summary. More generally, historical and biographical information is frequently presented as a time-line of events, and the ability to produce such time-lines from independent sources is valuable for online content providers such as Wikipedia. The following is an illustration of the utility of time-lines in analysing topics (events highlighted in bold to illustrate their importance in time-line generation):", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The Russian Revolution in 1917 was triggered by a combination of economic breakdown, war weariness, and discontent with the autocratic system of government, and it first brought a coalition of liberals and moderate socialists to power, but their failed policies led to seizure of power by the Communist Bolsheviks on October 25.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Existing summarization systems, such as the DUC 07 main task system of Stokes et al. (2007) , are effective at general summary tasks but have difficulty with event based summaries. For example, Stokes et al. (2007) performs well on general questions centred on a specified topic such as:", |
| "cite_spans": [ |
| { |
| "start": 71, |
| "end": 91, |
| "text": "Stokes et al. (2007)", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 194, |
| "end": 214, |
| "text": "Stokes et al. (2007)", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "What attacks have occurred against tourists in Luxor, Egypt?", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "but not so well on topics requiring a listing of events, such as:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "What have been the most recent significant events in the life and career of actress Angelina Jolie?", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this paper, we identify the most effective representation of text for real-world event reference identification in newspaper and newswire text using support vector machines (SVMs). The work presented here is the basis of an event-based summary system which will in turn form a component of the Stokes et al. (2007) system, where each question will be classified as an event-based question or a general question. The corresponding summary will then be generated by the appropriate sub-system, i.e. the original general-purpose system or a dedicated event-based system.", |
| "cite_spans": [ |
| { |
| "start": 297, |
| "end": 317, |
| "text": "Stokes et al. (2007)", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In creating an event-based time-line, three main tasks need to be considered: (1) how to identify and group references to events in text; (2) how to ground the identified events to a timestamp; and finally (3) how to identify salient events for inclusion in the final fixed-length time-line summary. In this paper we specifically focus on the identification of event references in text.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Often the task of event identification is linked to that of temporal resolution, whereby a system resolves the time or date of the occurrence of an event. Many systems treat events as any sentence containing an event reference and focus efforts on identifying temporal cues in the text. As there can be multiple event references within a sentence we identify events at the word level.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The following example is an illustration of the type of event-based identification we aim to achieve in this paper, relative to a plain-text input sentence (events highlighted in bold):", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Amid reports that thousands of Iraqi soldiers had surrendered, administration aides were also upbeat in private, with one even talking of victory within a week.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The focus of this paper is an exploration of the most appropriate feature representation to use in a purely statistical approach to word-level event identification in newswire text.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In the remainder of this paper, we review the notion of \"event\" in the literature, and present an overview of the different approaches to event identification. We then describe our experimental methodology, including the TimeBank corpus. Finally we describe a comparative evaluation.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The definition of an event varies in granularity depending on the desired application of event extraction.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Event Definition", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "The Scenario Template task of the Message Understanding Conference (MUC) (Grishman and Sundheim, 1996) relied on specified domaindependent templates which define events for the purpose of information extraction. The events that are extracted depend on the templates and documents supplied. MUC consists of domainspecific scenario (topic) based templates that systems are required to fill. An example domain is financial news in the Wall Street Journal, e.g. with the scenario of change in the corporate executive management personnel of an organization (Grishman and Sundheim, 1996) .", |
| "cite_spans": [ |
| { |
| "start": 73, |
| "end": 102, |
| "text": "(Grishman and Sundheim, 1996)", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 553, |
| "end": 582, |
| "text": "(Grishman and Sundheim, 1996)", |
| "ref_id": "BIBREF8" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Event Definition", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "In the context of Topic Detection and Tracking (TDT) (Fiscus et al., 1998) , an event corresponds to an instance of a topic identified at the document level, where the aim is to cluster documents on the same topic rather than individual events within the document. For instance, given an earthquake in the Solomon Islands on August 13, in TDT earthquakes would be the topic and this particular earthquake instance would be the event. The task would then be to cluster all documents that refer to this specific earthquake.", |
| "cite_spans": [ |
| { |
| "start": 53, |
| "end": 74, |
| "text": "(Fiscus et al., 1998)", |
| "ref_id": "BIBREF7" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Event Definition", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "The aim of the Automatic Content Extraction (ACE) (Doddington et al., 2004) task is to identify and classify events of predefined semantic types. An event reference may be a noun, verb or phrase, however events are tagged at the sentence level (LDC, 2005 ). In the above-mentioned documents referring to the Solomon Islands earthquake, the ACE data would have each reference marked as an event, in addition to announcements made by officials and any other words/phrases referring to events within the documents.", |
| "cite_spans": [ |
| { |
| "start": 50, |
| "end": 75, |
| "text": "(Doddington et al., 2004)", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 244, |
| "end": 254, |
| "text": "(LDC, 2005", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Event Definition", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Setzer (2001) (and later TimeML ) define an event as something that happens with a defined beginning and ending within the scope of a document. By comparison, a state is something that holds between entities without a definite beginning or end. The identification of events is based on verb and noun analysis similar in granularity to that in the ACE task. ACE and TimeML differ primarily on the semantic class labels for the events, and the annotated attributes and entities associated with each event. Where TimeML has seven semantic class labels (Occurrence, State, Reporting, I-Action, I-State, Aspectual and Perception), ACE has five more specific labels (Destruction/Damage, Movement, Creation/Improvement, Transfer of Possession or Control and Interaction of agents) (LDC, 2005) .", |
| "cite_spans": [ |
| { |
| "start": 774, |
| "end": 785, |
| "text": "(LDC, 2005)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Event Definition", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "For our purposes in this research, we define an event to be something that happens or occurs within the context of the document, similar to Setzer (2001) . However, we do not limit events to being denoted by only verbs or nouns. We assume that an event reference is contained within a sentence and does not cross sentence boundaries. Section 3.1 contains a detailed description of how an event is defined in our development data.", |
| "cite_spans": [ |
| { |
| "start": 140, |
| "end": 153, |
| "text": "Setzer (2001)", |
| "ref_id": "BIBREF14" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Event Definition", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Approaches to event extraction vary from ontology-and frame-based systems to combined rule-based and statistical methods. In this section we describe a number of representative approaches to event extraction from the literature.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Previous Work in Event Classification", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "The REES system (Aone and Ramos-Santacruz, 2000) uses hand-crafted event and relation ontologies to extract 100 relation and event types, 61 of which are events. It is purported to be extensible, however only so far as the ontology is extended. Events and relations relate to templates with type, person, time, place, etc, as defined in the ontology. The system consists of three parts: a tagging module (containing NameTagger, NPTagger and EventTagger), a rule-based co-reference resolution module, and a template generation module (described as a nonhard-coded approach that uses declarative rules to generate and merge templates automatically to \"achieve portability\"). MUC style data was used for training and testing. The system achieved a 0.70 F-score over 26 event types.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Previous Work in Event Classification", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Jones 2003applied Hidden Markov Models to the automatic generation of scripts to identify events. HMMs were used to capture correlations of significant events in text: the nodes correspond to events in the text, and the arcs indicate which clauses occur together in text. The context of neighbouring events and textual similarity are used to decide whether or not to group clauses.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Previous Work in Event Classification", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "As a component of TARSQI, aimed at interpreting entities and events in temporally-based questions, Verhagen and Mani (2005) developed the EVITA event recognition tool. EVITA is used for event recognition in newswire text and analysis of grammatical features, such as tense and aspect. Events are identified using lexical anal-ysis, context analysis of verbs, lexical lookup of adjectival events, and machine learning to determine whether an ambiguous noun is used in an event sense. They combine linguistic and statistical methods to obtain a 74.0% precision and 87.3% recall (Sauir et al., 2005) . Time references and events are defined by the TimeML annotation scheme.", |
| "cite_spans": [ |
| { |
| "start": 99, |
| "end": 123, |
| "text": "Verhagen and Mani (2005)", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 576, |
| "end": 596, |
| "text": "(Sauir et al., 2005)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Previous Work in Event Classification", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Preprocessing of the data was carried out using the Alembic Workbench for part of speech tagging, lematizing and chunking. A shallow parser is used to retrieve event referring expressions which are conveyed by verbs, nouns and adjectives. Verb-based events are identified by lexical lookup and contextual parsing of the verbal chunks. Events denoted by nouns are identified via WordNet and disambiguation with a Bayesian classifier trained on SemCor. Adjectives are tagged as events when they come at the head of a predictive complement, such as:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Previous Work in Event Classification", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "A Philippine volcano, dormant for six centuries, ... (TimeML, 2006) EVITA also classifies events semantically, which we do not consider in this paper. Bethard and Martin (2006) describe a system to identify events and their semantic class for the purposes of question answering. They define events as a description of situations that involve some internal structure, as opposed to states which describe a situation that is static or unchanging. Bethard and Martin (2006) view event identification as a classification task. Their system STEP is able to identify events with a precision of 82% and recall of 71%. They use a suite of syntactic and semantic features as input to YamCha -a general purpose chunker -and TinySVM. Each word in the document is classified as either beginning (B), inside (I) or outside (O) an event.", |
| "cite_spans": [ |
| { |
| "start": 53, |
| "end": 67, |
| "text": "(TimeML, 2006)", |
| "ref_id": "BIBREF16" |
| }, |
| { |
| "start": 151, |
| "end": 176, |
| "text": "Bethard and Martin (2006)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 445, |
| "end": 470, |
| "text": "Bethard and Martin (2006)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Previous Work in Event Classification", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Given that our notion of event closely mirrors that of TimeBank, we were able to use the TimeBank 1.1 corpus for system development and evaluation. TimeBank 1.1 consists of 186 news articles marked up with TimeML standard 1.1. The articles are sourced primarily from the Wall Street Journal and Associated Press newswire articles. They contain EVENT tags denoting events at the phrase and word level.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data: TimeML and TimeBank", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "TimeML is a specification language for event and temporal expressions in text , and builds on TIDES TIMEX2 and the temporal annotation language presented in Setzer (2001) .", |
| "cite_spans": [ |
| { |
| "start": 157, |
| "end": 170, |
| "text": "Setzer (2001)", |
| "ref_id": "BIBREF14" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data: TimeML and TimeBank", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "TimeML addresses four basic problems in event-time identification:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data: TimeML and TimeBank", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "\u2022 Time stamping of events (anchor event in time).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data: TimeML and TimeBank", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "\u2022 Chronological ordering of events with respect to one another.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data: TimeML and TimeBank", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "\u2022 Reasoning with contextually under-specified temporal expressions (e.g. last week).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data: TimeML and TimeBank", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "\u2022 Reasoning about the persistence of events (how long does an event or outcome of an event last).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data: TimeML and TimeBank", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "TimeML has four major event-time structures: EVENT (see Section 3.1.1), TIMEX3 (used to denote explicit time references), SIGNAL (used primarily to annotate function words indicating how temporal objects relate to one another) and LINK (denoting the relationship between an event and other events or temporal expressions).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data: TimeML and TimeBank", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "Events are a cover term for situations that happen or occur, and can be punctual or last for a time period. Events are generally expressed by tensed or untensed verbs, nominalisations, adjectives, predicative clauses or prepositional phrases. An EVENT in TimeML contains several attributes: a unique identifier, a class, tense, aspect and optionally another event. It may also contain another event or time to which it is temporally related.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "EVENT", |
| "sec_num": "3.1.1" |
| }, |
| { |
| "text": "There are seven event class types (each of which is illustrated with a set of examples):", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "EVENT", |
| "sec_num": "3.1.1" |
| }, |
| { |
| "text": "\u2022 Occurrence: die, crash, build, merger, sell We currently disregard the event classes and use a binary classification, identifying each word as an event or non-event.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "EVENT", |
| "sec_num": "3.1.1" |
| }, |
| { |
| "text": "We preprocessed the TimeBank data to remove extraneous XML tags, decompose it into sentences and words, and discard punctuation. Finally we assign each word with a POS tag using the Stanford POS tagger (Toutanova et al., 2003) .", |
| "cite_spans": [ |
| { |
| "start": 202, |
| "end": 226, |
| "text": "(Toutanova et al., 2003)", |
| "ref_id": "BIBREF17" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System Description", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "In our experiments, we use three classification algorithms: decision trees, naive Bayes and support vector machines. The chosen training and evaluation platform was WEKA (Witten and Frank, 2005) . Stratified 10-fold cross-validation was used throughout, where the labeled text fragments were randomly allocated to different combinations of training and testing splits. We report on the average performance over the 10 runs.", |
| "cite_spans": [ |
| { |
| "start": 170, |
| "end": 194, |
| "text": "(Witten and Frank, 2005)", |
| "ref_id": "BIBREF20" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System Description", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Preliminary experiments over the three classification algorithms indicated that SVMs were more effective in the classification of event references in text. BSVM (Hsu and Lin, 2002) was used for subsequent experiments, again using 10-fold cross-validation.", |
| "cite_spans": [ |
| { |
| "start": 161, |
| "end": 180, |
| "text": "(Hsu and Lin, 2002)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System Description", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "In Section 3.3, we present a range of different feature types, designed to explore the impact of different information sources on event reference classification.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System Description", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "In our classifiers, we look first at sentencelevel event classification (i.e. does a given sentence contain one or more event references) and then word-level event classification (i.e. is a given word an event reference). The sentence-level classifier is used to filter out the non-event sentences, and only event-containing sentences are passed on to the word-level classifier.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System Description", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "We use a basic word feature vector to represent the words being classified. Each distinct word in the input text corresponds to a unique feature. The text is transformed into a vector of N values where N is the number of distinct words in the TimeBank corpus. The following are common text representation techniques in text classification, the effects of which we explore in the context of event identification.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Features", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "For identification of events at the word level we explore the use of a context window of the zero to three preceding words, in addition to the word being classified.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u2022 Context Window", |
| "sec_num": null |
| }, |
| { |
| "text": "We also look at the effect of representing the current and context window words as a bag of word types (binary representation), or list of position-specified words (representing each word by its positional value [current word = 1, previous word = 2, etc; other words = 0]).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u2022 Feature Representation", |
| "sec_num": null |
| }, |
| { |
| "text": "Stop words are short, frequently occurring words such as and, or, in, of. Stop word removal reduces the feature space with generally little impact on classification performance in document categorisation tasks, but has been shown to be detrimental in sentence-level classification tasks (Khoo et al., 2006) . In our research, we use the stop word list of van Rijsbergen (1979).", |
| "cite_spans": [ |
| { |
| "start": 287, |
| "end": 306, |
| "text": "(Khoo et al., 2006)", |
| "ref_id": "BIBREF10" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u2022 Stop Words", |
| "sec_num": null |
| }, |
| { |
| "text": "\u2022 POS Parts of speech (POSs) are assigned to each word using the Stanford maxent POS tagger, and optionally used as an alternative to word features for context words (i.e. each context word is represented as its POS).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u2022 Stop Words", |
| "sec_num": null |
| }, |
| { |
| "text": "We optionally group numbers, times and named entities into single features. Numbers and dollar amounts were identified using regular expressions, and named entities were tagged using the Alembic Workbench (Aberdeen et al., 1995) .", |
| "cite_spans": [ |
| { |
| "start": 205, |
| "end": 228, |
| "text": "(Aberdeen et al., 1995)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "\u2022 Feature Generalisation", |
| "sec_num": null |
| }, |
| { |
| "text": "We use F-score as our primary evaluation metric, as defined below relative to Precision and Recall: (2)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation Method", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "Precision = #", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation Method", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "F-Score = 2 \u00d7 Precision \u00d7 Recall Precision + Recall", |
| "eq_num": "(3)" |
| } |
| ], |
| "section": "Evaluation Method", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "Our baseline takes the form of a unigram tagger, where each word is labeled with its most frequent label. Table 1 shows the performance of the baseline system, in addition to the the comparable STEP (Aone and Ramos-Santacruz, 2000) and EVITA (Sauir et al., 2005 ) results -as described in Section 2.2 -both of which use the TimeBank data in the development and testing of their systems.", |
| "cite_spans": [ |
| { |
| "start": 199, |
| "end": 231, |
| "text": "(Aone and Ramos-Santacruz, 2000)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 242, |
| "end": 261, |
| "text": "(Sauir et al., 2005", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 106, |
| "end": 113, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Evaluation Method", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "In this section we first present the results of our sentence-level event identification system, where each sentence is classified as containing one or more event references, or alternatively containing no event references. Three learners -C4.5 decision tree, SVM and naive Bayes -were run over a random subset of the TimeBank data, the results of which are shown in Table 2 .", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 366, |
| "end": 373, |
| "text": "Table 2", |
| "ref_id": "TABREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Sentence Level Classification of Events", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "All three methods performed exceptionally well, at an F-score of or close to 1. This level of performance is due to 82% of the sentences in the corpus containing an event reference, and most of these containing more than one event reference. Therefore only one of the many event references within a sentence needs to be identified for the sentence to be correctly classified.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sentence Level Classification of Events", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "The naive Bayes classifier and decision tree models proved difficult to run over the entire TimeBank corpus due to resource constraints. As a result, we chose to use BSVM (Hsu and Lin, 2002) for all subsequent sentence-level experiments.", |
| "cite_spans": [ |
| { |
| "start": 171, |
| "end": 190, |
| "text": "(Hsu and Lin, 2002)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sentence Level Classification of Events", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "Algorithm Precision Recall F-Score Naive Bayes 0.995 0.998 0.996 SVM 1.000 1.000 1.000 C4.5", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sentence Level Classification of Events", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "1.000 1.000 1.000 As there can be multiple references to different events within a single sentence, sentence level classification is not fine-grained enough for event clustering and extraction in the context of eventbased summary generation. The sentence classifier is instead used as a pre-filter for word-level event classification, as discussed in the following section.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sentence Level Classification of Events", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "As stated above, in classifying individual words for event reference, we first use the sentence-level classifier to remove sentences that do not contain events. This proved to have no effect on the accuracy of the word level system, but did offer a slight reduction in the time required by the wordlevel classifier.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Word-level Classification of Events", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Both the naive Bayes classifier and C4.5 decision tree resulted in majority class classifiers (see Table 1 ), where all words were tagged as nonevents. The SVM, on the other hand, produced a variety of results, and thus forms the basis of all results presented in this section.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 99, |
| "end": 106, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Word-level Classification of Events", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Below, we will discuss our findings into the effects of stop word removal, feature representation, context window size, POS tagging and feature generalisation on word-level event classification.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Word-level Classification of Events", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Increasing the context window, using a binary feature representation, decreased the accuracy of the classifier, as shown in the leftmost segment of Figure 1 . The zero word context improves only 1% with an increase from 15 to 18 thousand training instances compared with an improvement of 7% for the three word prior context model. At this point, we have been unable to improve over the baseline system. ", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 148, |
| "end": 156, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Context Window Size", |
| "sec_num": "4.2.1" |
| }, |
| { |
| "text": "Replacing the binary bag of words representation with a position-specified binary word representation has a dramatic impact on our results, as indicated in the middle set of results in Figure 1. Maintaining the ordering of 2 word prior context has an F-score of 0.68, compared to a binary representation of 2 word prior context with an F-score of 0.41. Khoo et al. (2006) found that stop word removal had a detrimental effect on classification at the sentence level. Our results, however, demonstrate that stop word removal offers a slight increase in the accuracy of the word-level classifiers. Table 3 shows the results obtained using SVM to identify event references at the word level using different combinations of text representation techniques. As we can see from the table, stop word removal had a positive effect on the system irrespective of whether a binary or word order representation was used. With a one word POS context window and feature generalisation, stop word removal increases the F-score by 0.043. Similarly, using a binary representation and grouping, stop word removal increases the Fscore by 0.069.", |
| "cite_spans": [ |
| { |
| "start": 353, |
| "end": 371, |
| "text": "Khoo et al. (2006)", |
| "ref_id": "BIBREF10" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 185, |
| "end": 191, |
| "text": "Figure", |
| "ref_id": null |
| }, |
| { |
| "start": 596, |
| "end": 603, |
| "text": "Table 3", |
| "ref_id": "TABREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Feature Representation", |
| "sec_num": "4.2.2" |
| }, |
| { |
| "text": "Including the POS tags of the context and word being classified increases performance of Figure 2 : The learning curves for the models of different context window sizes, with feature generalisation and stop word removal the overall system. However, the greatest increase comes from the inclusion of the preceding context represented as a part of speech, without context words. This facilitates the generalisation of the patterns leading up to the event reference. Note that the word being tagged is retained as a binary feature, as its exclusion would leave a POS pattern feature that is too general given the limited combination of less than fifty distinct POS tags.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 89, |
| "end": 97, |
| "text": "Figure 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Part Of Speech", |
| "sec_num": "4.2.4" |
| }, |
| { |
| "text": "Grouping specific types of words such as numbers, named entities and time references into a single feature proved an effective way of reducing the feature space. However, closer analysis of the word-level event classification results showed that some of the terms being grouped together occurred as events whereas others did not. There are 610 NUMBER references in the TimeBank data. 247 of these are an event or part thereof and 363 are not, leading to most NUMBER references being tagged as non-events. For example, monetary amounts can be a state (type of event) if they are a milestone for a company, or alternatively they can simply be an amount lost by the company.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Feature Generalisation", |
| "sec_num": "4.2.5" |
| }, |
| { |
| "text": "The insurer's earnings from commercial property lines fell 59% in the latest quarter, while it lost $7.2 million in its personal property business, compared with earnings of $6.1 million 1 a year ago.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Feature Generalisation", |
| "sec_num": "4.2.5" |
| }, |
| { |
| "text": "The other groupings used, TIME and NAMED ENTITY, are never event references in the Time-Bank corpus, and thus their grouping reduces the feature space without loss of the ability to disambiguate. Further investigation into a more finegrained grouping of numerical terms is required before grouping numbers is included in our final system.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Feature Generalisation", |
| "sec_num": "4.2.5" |
| }, |
| { |
| "text": "Sentence-level classification proved a simple task with an almost perfect classification accuracy. The use of a sentence-level classifier for prior filtering of the word-level event system had no effect of the overall F-score.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussion", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Our final system combined POS representation of the two word prior context with context positional information and stop word removal to achieve an F-score of 0.764, exceeding our baseline of 0.653 and marginally exceeding the performance of the comparable STEP system with an F-score of 0.759. Over 50% of incorrectly classified words resulted from event words which occur only once in the TimeBank corpus. A further 20% are words that appear as both events and non-events in the training data. 7% of the remaining errors result from incorrect tokenisation dur-1 States that describe circumstances in which something obtains and holds true are called predictive states. The validity of this is dependent on the document creation time, and includes some quantitative statements such as those that appear in financial journals (TimeML, 2006) ", |
| "cite_spans": [ |
| { |
| "start": 825, |
| "end": 839, |
| "text": "(TimeML, 2006)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussion", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "[pg14].", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussion", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "ing preprocessing. There are 1540 words in the TimeBank data that occur only once as events.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussion", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "To investigate the impact of the amount of data on our results, we generate learning curves for word order-sensitive feature representation with stop word removal and feature generalisation, over diminishing amounts of the training data used in cross-validation. The curves are presented in Figure 2 over different context word window sizes. As we can see, the models based on larger context windows are still climbing steeply over the final increment of training instances, while the zero-word context model appears to be flattening out. As such, we would expect that an increase in the number of training instances to improve the performance of the models with larger context windows, although it remains to be seen how much they would improve relative to the smaller word windows.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 291, |
| "end": 299, |
| "text": "Figure 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Discussion", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Maintaining word order has the biggest effect on the performance of the classification system. Representing the preceding context as POS rather than words also increases the accuracy of the overall system and, contrary to expectations, the removal of stop words marginally increased the performance. As indicated by the learning curve (Figure 2 ), increasing the number of training instances has the potential to improve the performance of the models which include context.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 335, |
| "end": 344, |
| "text": "(Figure 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Discussion", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "The two-word POS context model misclassifies a different subset of words to the threeword POS context model. Therefore, combining different models may increase the performance of the overall system. The majority of errors came from previously unseen words in the training data. Future work will thus focus on how the system can better handle unseen words.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussion", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "We intend to look at the effect verb types obtained from WordNet (Fellbaum, 1998) or VERBOCEAN (Chklovski and Pantel, 2004) have on word-level event identification, to assist in better handling of unseen words. We will also look at how the system performance scales over a larger data set.", |
| "cite_spans": [ |
| { |
| "start": 65, |
| "end": 81, |
| "text": "(Fellbaum, 1998)", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 95, |
| "end": 123, |
| "text": "(Chklovski and Pantel, 2004)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Future Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "We presented a set of experiments involving realworld event reference identification at the word level in newspaper and newswire documents. We addressed the issue of effective text representation for classification of events using support vector machines. Our final system combined two word POS prior context, with positional information and stop word removal. It achieved an F-score of 0.764, significantly exceeding that of our baseline system with an F-score of 0.653.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "6" |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "NICTA is funded by the Australian government as represented by Department of Broadband, Communication and Digital Economy, and the Australian Research Council through the ICT Centre of Excellence programme.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgments", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "MITRE: Description of the Alembic system used for MUC-6", |
| "authors": [ |
| { |
| "first": "John", |
| "middle": [], |
| "last": "Aberdeen", |
| "suffix": "" |
| }, |
| { |
| "first": "John", |
| "middle": [ |
| "D" |
| ], |
| "last": "Burger", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [ |
| "S" |
| ], |
| "last": "Day", |
| "suffix": "" |
| }, |
| { |
| "first": "Lynette", |
| "middle": [], |
| "last": "Hirschman", |
| "suffix": "" |
| }, |
| { |
| "first": "Patricia", |
| "middle": [], |
| "last": "Robinson", |
| "suffix": "" |
| }, |
| { |
| "first": "Marc", |
| "middle": [ |
| "B" |
| ], |
| "last": "Vilain", |
| "suffix": "" |
| } |
| ], |
| "year": 1995, |
| "venue": "Proceedings of MUC-6", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "John Aberdeen, John D. Burger, David S. Day, Lynette Hirschman, Robinson, Patricia, and Marc B. Vilain. 1995. MITRE: Description of the Alembic system used for MUC-6. In Proceedings of MUC-6.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "REES: A large-scale relation and event extraction system", |
| "authors": [ |
| { |
| "first": "Chinatsu", |
| "middle": [], |
| "last": "Aone", |
| "suffix": "" |
| }, |
| { |
| "first": "Mila", |
| "middle": [], |
| "last": "Ramos-Santacruz", |
| "suffix": "" |
| } |
| ], |
| "year": 2000, |
| "venue": "Proceedings of the 6th Applied Natural Language Processing Conference", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Chinatsu Aone and Mila Ramos-Santacruz. 2000. REES: A large-scale relation and event extraction system. In Proceedings of the 6th Applied Natural Language Processing Conference, Seattle, USA.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Identification of event mentions and their semantic class", |
| "authors": [ |
| { |
| "first": "Steven", |
| "middle": [], |
| "last": "Bethard", |
| "suffix": "" |
| }, |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Martin", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of the 2006 Conference on Emprical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Steven Bethard and James Martin. 2006. Identifica- tion of event mentions and their semantic class. In Proceedings of the 2006 Conference on Emprical Methods in Natural Language Processing, Sydney, Australia.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "VER-BOCEAN: Mining the web for fine-grained semantic verb relations", |
| "authors": [ |
| { |
| "first": "Timothy", |
| "middle": [], |
| "last": "Chklovski", |
| "suffix": "" |
| }, |
| { |
| "first": "Patrick", |
| "middle": [], |
| "last": "Pantel", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proceedings of Conference on Empirical Methods in Natural Language Processing (EMNLP-04)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Timothy Chklovski and Patrick Pantel. 2004. VER- BOCEAN: Mining the web for fine-grained seman- tic verb relations. In Proceedings of Conference on Empirical Methods in Natural Language Process- ing (EMNLP-04), Barcelona, Spain.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Overview of DUC", |
| "authors": [ |
| { |
| "first": "Hoa", |
| "middle": [ |
| "Trang" |
| ], |
| "last": "Dang", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of the Document Understanding Conference Workshop", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hoa Trang Dang. 2006. Overview of DUC 2006. In Proceedings of the Document Understanding Con- ference Workshop, New York, USA.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Automatic content extraction (ACE) program -task definitions and performance measures", |
| "authors": [ |
| { |
| "first": "George", |
| "middle": [], |
| "last": "Doddington", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexis", |
| "middle": [], |
| "last": "Mitchell", |
| "suffix": "" |
| }, |
| { |
| "first": "Mark", |
| "middle": [], |
| "last": "Przybocki", |
| "suffix": "" |
| }, |
| { |
| "first": "Lance", |
| "middle": [], |
| "last": "Ramshaw", |
| "suffix": "" |
| }, |
| { |
| "first": "Stephanie", |
| "middle": [], |
| "last": "Strassel", |
| "suffix": "" |
| }, |
| { |
| "first": "Ralph", |
| "middle": [], |
| "last": "Weischedel", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proceedings of the Fourth International Conference on Language Resources and Evaluation", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "George Doddington, Alexis Mitchell, Mark Przy- bocki, Lance Ramshaw, Stephanie Strassel, and Ralph Weischedel. 2004. Automatic content ex- traction (ACE) program -task definitions and per- formance measures. In Proceedings of the Fourth International Conference on Language Resources and Evaluation. Lisbon, Portugal.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "WordNet An Electronic Lexical Database", |
| "authors": [ |
| { |
| "first": "Christiane", |
| "middle": [], |
| "last": "Fellbaum", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Christiane Fellbaum. 1998. WordNet An Electronic Lexical Database. The MIT Press, first edition.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "NIST's 1998 topic detection and tracking evaluation", |
| "authors": [ |
| { |
| "first": "Jon", |
| "middle": [], |
| "last": "Fiscus", |
| "suffix": "" |
| }, |
| { |
| "first": "George", |
| "middle": [], |
| "last": "Doddington", |
| "suffix": "" |
| }, |
| { |
| "first": "John", |
| "middle": [], |
| "last": "Garofolo", |
| "suffix": "" |
| }, |
| { |
| "first": "Alvin", |
| "middle": [], |
| "last": "Martin", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "Proceedings of the DARPA Broadcast News Workshop", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jon Fiscus, George Doddington, John Garofolo, and Alvin Martin. 1998. NIST's 1998 topic detec- tion and tracking evaluation. In Proceedings of the DARPA Broadcast News Workshop.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "A comparison on methods for multi-class support vector machines", |
| "authors": [ |
| { |
| "first": "Ralph", |
| "middle": [], |
| "last": "Grishman", |
| "suffix": "" |
| }, |
| { |
| "first": "Beth", |
| "middle": [], |
| "last": "Sundheim", |
| "suffix": "" |
| } |
| ], |
| "year": 1996, |
| "venue": "Proceedings of the 16th Conference on Computational Linguistics", |
| "volume": "13", |
| "issue": "", |
| "pages": "415--425", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ralph Grishman and Beth Sundheim. 1996. Message Understanding Conference 6: A brief history. In Proceedings of the 16th Conference on Computa- tional Linguistics, Copenhagen, Denmark. Chih-Wei Hsu and Chih-Jen Lin. 2002. A compari- son on methods for multi-class support vector ma- chines. IEEE Transactions on Neural Networks, 13:415-425.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Identifying events using similarity and context", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Dominic", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Jones", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Proceedings of the Seventh Conference on Natural Language Learning", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dominic R. Jones. 2003. Identifying events using similarity and context. In Proceedings of the Sev- enth Conference on Natural Language Learning, Edmonton, Canada.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Experiments with sentence classification", |
| "authors": [ |
| { |
| "first": "Anthony", |
| "middle": [], |
| "last": "Khoo", |
| "suffix": "" |
| }, |
| { |
| "first": "Yuval", |
| "middle": [], |
| "last": "Marom", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Albrecht", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of the 2006 Australasian Language Technology Workshop", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Anthony Khoo, Yuval Marom, and David Albrecht. 2006. Experiments with sentence classification. In Proceedings of the 2006 Australasian Language Technology Workshop, Sydney, Australia.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "ACE (automatic content extraction) English annotation guidelines for events", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Ldc", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "LDC. 2005. ACE (automatic content extraction) En- glish annotation guidelines for events.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "The specification language TimeML", |
| "authors": [ |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Pustejovsky", |
| "suffix": "" |
| }, |
| { |
| "first": "Robert", |
| "middle": [], |
| "last": "Ingria", |
| "suffix": "" |
| }, |
| { |
| "first": "Roser", |
| "middle": [], |
| "last": "Sauri", |
| "suffix": "" |
| }, |
| { |
| "first": "Jose", |
| "middle": [], |
| "last": "Castano", |
| "suffix": "" |
| }, |
| { |
| "first": "Jessica", |
| "middle": [], |
| "last": "Littman", |
| "suffix": "" |
| }, |
| { |
| "first": "Rob", |
| "middle": [], |
| "last": "Gaizauskas", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrea", |
| "middle": [], |
| "last": "Setzer", |
| "suffix": "" |
| }, |
| { |
| "first": "Graham", |
| "middle": [], |
| "last": "Katz", |
| "suffix": "" |
| }, |
| { |
| "first": "Inderjeet", |
| "middle": [], |
| "last": "Mani", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "The Language of Time: A Reader, chapter 27", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "James Pustejovsky, Robert Ingria, Roser Sauri, Jose Castano, Jessica Littman, Rob Gaizauskas, Andrea Setzer, Graham Katz, and Inderjeet Mani. 2005. The specification language TimeML. In The Lan- guage of Time: A Reader, chapter 27. Oxford Uni- versity Press.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "EVITA: A robust event recognizer for QA systems", |
| "authors": [ |
| { |
| "first": "Roser", |
| "middle": [], |
| "last": "Sauir", |
| "suffix": "" |
| }, |
| { |
| "first": "Robert", |
| "middle": [], |
| "last": "Knippen", |
| "suffix": "" |
| }, |
| { |
| "first": "Marc", |
| "middle": [], |
| "last": "Verhagen", |
| "suffix": "" |
| }, |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Pustejovsky", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Proceedings of Human Language Technology Conference and Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "700--707", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Roser Sauir, Robert Knippen, Marc Verhagen, and James Pustejovsky. 2005. EVITA: A robust event recognizer for QA systems. In Proceedings of Hu- man Language Technology Conference and Confer- ence on Empirical Methods in Natural Language Processing, pages 700-707, Vancouver, Canada.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Temporal Information In Newswire Articles: An Annotation Scheme and Corpus Study", |
| "authors": [ |
| { |
| "first": "Andrea", |
| "middle": [], |
| "last": "Setzer", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Andrea Setzer. 2001. Temporal Information In Newswire Articles: An Annotation Scheme and Corpus Study. Ph.D. thesis, University of Sheffield.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "NICTA's update and question-based summarisation systems at DUC", |
| "authors": [ |
| { |
| "first": "Nicola", |
| "middle": [], |
| "last": "Stokes", |
| "suffix": "" |
| }, |
| { |
| "first": "Jiawen", |
| "middle": [], |
| "last": "Rong", |
| "suffix": "" |
| }, |
| { |
| "first": "Lawrence", |
| "middle": [], |
| "last": "Cavedon", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proceedings of the Document Understanding Conference Workshop", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nicola Stokes, Jiawen Rong, and Lawrence Cavedon. 2007. NICTA's update and question-based sum- marisation systems at DUC 2007. In Proceedings of the Document Understanding Conference Work- shop, Rochester, USA.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "TimeML annotation guidelines version 1.2.1", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Timeml", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "TimeML. 2006. TimeML annotation guidelines ver- sion 1.2.1.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Feature-rich part-ofspeech tagging with a cyclic dependency network", |
| "authors": [ |
| { |
| "first": "Kristina", |
| "middle": [], |
| "last": "Toutanova", |
| "suffix": "" |
| }, |
| { |
| "first": "Dan", |
| "middle": [], |
| "last": "Klein", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [], |
| "last": "Manning", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoram", |
| "middle": [], |
| "last": "Singer", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Proceedings of HLT-NAACL 2003", |
| "volume": "", |
| "issue": "", |
| "pages": "252--259", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kristina Toutanova, Dan Klein, Christopher Manning, and Yoram Singer. 2003. Feature-rich part-of- speech tagging with a cyclic dependency network. In Proceedings of HLT-NAACL 2003, pages 252- 259, Edmonton, Canada.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Information Retrieval", |
| "authors": [ |
| { |
| "first": "C", |
| "middle": [ |
| "J" |
| ], |
| "last": "Keith Van Rijsbergen", |
| "suffix": "" |
| } |
| ], |
| "year": 1979, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "C.J. Keith van Rijsbergen. 1979. Information Re- trieval. Butterworth-Heinemann, 2nd edition.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Automating temporal annotaion with TARSQI", |
| "authors": [ |
| { |
| "first": "Marc", |
| "middle": [], |
| "last": "Verhagen", |
| "suffix": "" |
| }, |
| { |
| "first": "Inderjeet", |
| "middle": [], |
| "last": "Mani", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Proceedings of the ACL 2005 Interactive Poster and Demonstration Sessions", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Marc Verhagen and Inderjeet Mani. 2005. Automat- ing temporal annotaion with TARSQI. In Pro- ceedings of the ACL 2005 Interactive Poster and Demonstration Sessions, Ann Arbor, USA.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "Data Mining: Practical Machine Learning Tools and Techniques", |
| "authors": [ |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Ian", |
| "suffix": "" |
| }, |
| { |
| "first": "Eibe", |
| "middle": [], |
| "last": "Witten", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Frank", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ian H. Witten and Eibe Frank. 2005. Data Mining: Practical Machine Learning Tools and Techniques.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "text": "The F-score of models trained over different feature combinations. (* = statistically significant difference: two-tailed paired t-test, p < 0.05)", |
| "type_str": "figure", |
| "uris": null, |
| "num": null |
| }, |
| "TABREF0": { |
| "content": "<table/>", |
| "text": "State: on board, kidnaped, in love \u2022 Reporting: say, report, announce \u2022 I-Action: attempt, try, promise, offer \u2022 I-State: believe, intend, want \u2022 Aspectual: begin, finish, stop, continue \u2022 Perception: see, hear, watch, feel", |
| "html": null, |
| "num": null, |
| "type_str": "table" |
| }, |
| "TABREF2": { |
| "content": "<table/>", |
| "text": "Performance of sentence level classification with stop word removal and feature generalisation", |
| "html": null, |
| "num": null, |
| "type_str": "table" |
| }, |
| "TABREF4": { |
| "content": "<table><tr><td/><td/><td/><td/><td/><td colspan=\"2\">Learning Curve</td><td/><td/><td/></tr><tr><td/><td>0.75</td><td/><td/><td/><td/><td/><td/><td/><td/></tr><tr><td/><td>0.7</td><td/><td/><td/><td/><td/><td/><td/><td/></tr><tr><td/><td>0.65</td><td/><td/><td/><td/><td/><td/><td/><td/></tr><tr><td/><td>0.6</td><td/><td/><td/><td/><td/><td/><td/><td/></tr><tr><td>F-Score</td><td>0.5 0.55</td><td/><td/><td/><td/><td/><td/><td/><td/></tr><tr><td/><td>0.45</td><td/><td/><td/><td/><td/><td>Zero</td><td/><td/></tr><tr><td/><td/><td/><td/><td/><td/><td/><td>One</td><td/><td/></tr><tr><td/><td>0.4</td><td/><td/><td/><td/><td/><td>Two Three</td><td/><td/></tr><tr><td/><td/><td/><td/><td/><td/><td colspan=\"2\">Baseline</td><td/><td/></tr><tr><td/><td>0.35</td><td/><td/><td/><td/><td/><td/><td/><td/></tr><tr><td/><td>0.3</td><td/><td/><td/><td/><td/><td/><td/><td/></tr><tr><td/><td>0</td><td>2</td><td>4</td><td>6</td><td>8</td><td>10</td><td>12</td><td>14</td><td>16</td><td>18</td></tr><tr><td/><td/><td/><td colspan=\"6\">Number of Training Instances (Thousands)</td><td/></tr></table>", |
| "text": "The results of word-level event identification using an SVM and different combinations of features", |
| "html": null, |
| "num": null, |
| "type_str": "table" |
| } |
| } |
| } |
| } |