| { |
| "paper_id": "C02-1042", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T12:20:14.553380Z" |
| }, |
| "title": "Using Knowledge to Facilitate Factoid Answer Pinpointing", |
| "authors": [ |
| { |
| "first": "Eduard", |
| "middle": [], |
| "last": "Hovy", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Ulf Hermjakob", |
| "location": { |
| "addrLine": "Chin-Yew Lin" |
| } |
| }, |
| "email": "hovy@isi.edu" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "In order to answer factoid questions, the Webclopedia QA system employs a range of knowledge resources. These include a QA Typology with answer patterns, WordNet, information about typical numerical answer ranges, and semantic relations identified by a robust parser, to filter out likely-looking but wrong candidate answers. This paper describes the knowledge resources and their impact on system performance.", |
| "pdf_parse": { |
| "paper_id": "C02-1042", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "In order to answer factoid questions, the Webclopedia QA system employs a range of knowledge resources. These include a QA Typology with answer patterns, WordNet, information about typical numerical answer ranges, and semantic relations identified by a robust parser, to filter out likely-looking but wrong candidate answers. This paper describes the knowledge resources and their impact on system performance.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "The TREC evaluations of QA systems (Voorhees, 1999) require answers to be drawn from a given source corpus. Early QA systems used a simple filtering technique, question word density within a fixed n-word window, to pinpoint answers. Robust though this may be, the window method is not accurate enough. In response, factoid question answering systems have evolved into two types:", |
| "cite_spans": [ |
| { |
| "start": 35, |
| "end": 51, |
| "text": "(Voorhees, 1999)", |
| "ref_id": "BIBREF27" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": null |
| }, |
| { |
| "text": "\u2022 Use-Knowledge: extract query words from the input question, perform IR against the source corpus, possibly segment resulting documents, identify a set of segments containing likely answers, apply a set of heuristics that each consults a different source of knowledge to score each candidate, rank them, and select the best (Harabagiu et al., 2001; Hovy et al., 2001; Srihari and Li, 2000; Abney et al., 2000) . \u2022 Use-the-Web: extract query words from the question, perform IR against the web, extract likely answer-bearing sentences, canonicalize the results, and select the most frequent answer(s). Then, for justification, locate examples of the answers in the source corpus (Brill et al., 2001; Buchholz, 2001 ). Of course, these techniques can be combined: the popularity ratings from Use-the-Web can also be applied as a filtering criterion , or the knowledge resource heuristics can filter the web results. However, simply going to the web without using further knowledge (Brill et al., 2001 ) may return the web's majority opinions on astrology, the killers of JFK, the cancerous effects of microwave ovens, etc.-fun but not altogether trustworthy.", |
| "cite_spans": [ |
| { |
| "start": 325, |
| "end": 349, |
| "text": "(Harabagiu et al., 2001;", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 350, |
| "end": 368, |
| "text": "Hovy et al., 2001;", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 369, |
| "end": 390, |
| "text": "Srihari and Li, 2000;", |
| "ref_id": "BIBREF26" |
| }, |
| { |
| "start": 391, |
| "end": 410, |
| "text": "Abney et al., 2000)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 679, |
| "end": 699, |
| "text": "(Brill et al., 2001;", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 700, |
| "end": 714, |
| "text": "Buchholz, 2001", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 980, |
| "end": 999, |
| "text": "(Brill et al., 2001", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": null |
| }, |
| { |
| "text": "In this paper we describe the range of filtering techniques our system Webclopedia applies, from simplest to most sophisticated, and indicate their impact on the system.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": null |
| }, |
| { |
| "text": "As shown in Figure 1 , Webclopedia adopts the Use-Knowledge architecture. Its modules are described in more detail in (Hovy et al., 2001; Hovy et al., 1999) :", |
| "cite_spans": [ |
| { |
| "start": 118, |
| "end": 137, |
| "text": "(Hovy et al., 2001;", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 138, |
| "end": 156, |
| "text": "Hovy et al., 1999)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 12, |
| "end": 20, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Webclopedia Architecture", |
| "sec_num": "2." |
| }, |
| { |
| "text": "\u2022 Question parsing: Using BBN's IdentiFinder (Bikel et al., 1999) , the CONTEX parser (Hermjakob, 1997) produces a syntactic-semantic analysis of the question and determines the QA type.", |
| "cite_spans": [ |
| { |
| "start": 45, |
| "end": 65, |
| "text": "(Bikel et al., 1999)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 86, |
| "end": 103, |
| "text": "(Hermjakob, 1997)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Webclopedia Architecture", |
| "sec_num": "2." |
| }, |
| { |
| "text": "\u2022 Query formation: Single-and multi-word units (content words) are extracted from the analysis, and WordNet synsets (Fellbaum, 1998) are used for query expansion. A series of Boolean queries of decreasing specificity is formed. \u2022 IR: The publicly available IR engine MG (Witten et al., 1994) returns the top-ranked N documents.", |
| "cite_spans": [ |
| { |
| "start": 116, |
| "end": 132, |
| "text": "(Fellbaum, 1998)", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 270, |
| "end": 291, |
| "text": "(Witten et al., 1994)", |
| "ref_id": "BIBREF30" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Webclopedia Architecture", |
| "sec_num": "2." |
| }, |
| { |
| "text": "\u2022 Selecting and ranking sentences: For each document, the most promising K sentences are located and scored using a formula that rewards word and phrase overlap with the question and its expanded query words. Results are ranked.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Webclopedia Architecture", |
| "sec_num": "2." |
| }, |
| { |
| "text": "\u2022 Parsing candidates: CONTEX parses the top-ranked 300 sentences. \u2022 Pinpointing: As described in Section 3, a number of knowledge resources are used to perform filtering/pinpointing operations. \u2022 Ranking of answers: The candidate answers' scores are compared and the winner(s) are output.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Webclopedia Architecture", |
| "sec_num": "2." |
| }, |
| { |
| "text": "Unlike (Prager et al., 1999) , we do not first annotate the source corpus, but perform IR directly on the source text, using MG (Witten et al., 1994) . To determine goodness, we assign an initial base score to each retrieved sentence. We then compare the sentence to the question and adapt this score as follows:", |
| "cite_spans": [ |
| { |
| "start": 7, |
| "end": 28, |
| "text": "(Prager et al., 1999)", |
| "ref_id": "BIBREF21" |
| }, |
| { |
| "start": 128, |
| "end": 149, |
| "text": "(Witten et al., 1994)", |
| "ref_id": "BIBREF30" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 1: Question Word Matching", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "\u2022 exact matches of proper names double the base score. \u2022 matching an upper-cased term adds a 60% bonus of the base score for multi-words terms and 30% for single words (matching \"United States\" is better than just \"United\"). \u2022 matching a WordNet synonym of a term discounts by 10% (lower case) and 50% (upper case). (When \"Cage\" matches \"cage\", the former may be the last name of a person and the latter an object; the case mismatch signals less reliability.)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 1: Question Word Matching", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "\u2022 lower-case term matches after Porter stemming are discounted 30%; upper-case matches 70% (Porter stemming is more aggressive than WordNet stemming). \u2022 Porter stemmer matches of both question and sentence words with lower case are discounted 60%; with upper case, 80%. \u2022 if CONTEX indicates a term as being qsubsumed (see Section 3.9) the term is discouned 90% (in \"Which country manufactures weapons of mass destruction?\", \"country\" will be marked as qsubsumed). The top-scoring 300 sentences are passed on for further filtering.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 1: Question Word Matching", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "We classify desired answers by their semantic type, which have been taxonomized in the Webclopedia QA Typology (Hovy et al., 2002 ), http://www.isi.edu/natural-language/projects/we bclopedia/Taxonomy/taxonomy_toplevel.html). The currently approx. 180 classes, which we call qtargets, were developed after an analysis of over 17,000 questions (downloaded in 1999 from answers.com) and later enhancements to Webclopedia. They are of several types:", |
| "cite_spans": [ |
| { |
| "start": 111, |
| "end": 129, |
| "text": "(Hovy et al., 2002", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 2: Qtargets, the QA Typology, and the Semantic Ontology", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "\u2022 common semantic classes such as PROPER-PERSON, EMAIL-ADDRESS, LOCATION, PROPER-ORGANIZATION;", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 2: Qtargets, the QA Typology, and the Semantic Ontology", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "\u2022 classes particular to QA such as YES:NO, ABBREVIATION-EXPANSION, and WHY-FAMOUS;", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 2: Qtargets, the QA Typology, and the Semantic Ontology", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "\u2022 syntactic classes such as NP and NOUN, when no semnatic type can be determined (e.g., \"What does Peugeot manufacture?\"); \u2022 roles and slots, such as REASON and TITLE-P respectively, to indicate a desired relation with an anchoring concept. Given a question, the CONTEX parser uses a set of 276 hand-built rules to identify its most likely qtarget(s), and records them in a backoff scheme (allowing more general qtarget nodes to apply when more specific ones fail to find a match). The generalizations are captured in a typical concept ontology, a 10,000-node extract of WordNet.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 2: Qtargets, the QA Typology, and the Semantic Ontology", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "The recursive part of pattern matching is driven mostly by interrogative phrases. For example, the rule that determines the applicability of the qtarget WHY-FAMOUS requires the question word \"who\", followed by the copula, followed by a proper name. When there is no match at the current level, the system examines any interrogative constituent, or words in special relations to it. For example, the qtarget TEMPERATURE-QUANTITY (as in \"What is the melting point of X?\" requires as syntactic object something that in the ontology is subordinate to TEMP-QUANTIFIABLE-ABS-TRACT with, as well, the word \"how\" paired with \"warm\", \"cold\", \"hot\", etc., or the phrase \"how many degrees\" and a TEMPERATURE-UNIT (as defined in the ontology).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 2: Qtargets, the QA Typology, and the Semantic Ontology", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Often qtarget answers are expressed using rather stereotypical words or phrases. For example, the year of birth of a person is typically expressed using one of these phrases:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 3: Surface Pattern Matching", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "<name> was born in <birthyear> <name> (<birthyear>-<deathyear>) We have developed a method to learn such patterns automatically from text on the web (Ravichandran and Hovy, 2002) . We have added into the QA Typology the patterns for appropriate qtargets (qtargets with closed-list answers, such as PLANETS, require no patterns). Where some QA systems use such patterns exclusively (Soubbotin and Soubbotin, 2001) or partially (Wang et al., 2001; Lee et al., 2001 ), we employ them as an additional source of evidence for the answer. Preliminary results on for a range of qtargets, using the TREC-10 questions and the TREC corpus, are: ", |
| "cite_spans": [ |
| { |
| "start": 38, |
| "end": 63, |
| "text": "(<birthyear>-<deathyear>)", |
| "ref_id": null |
| }, |
| { |
| "start": 149, |
| "end": 178, |
| "text": "(Ravichandran and Hovy, 2002)", |
| "ref_id": "BIBREF23" |
| }, |
| { |
| "start": 381, |
| "end": 412, |
| "text": "(Soubbotin and Soubbotin, 2001)", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 426, |
| "end": 445, |
| "text": "(Wang et al., 2001;", |
| "ref_id": "BIBREF29" |
| }, |
| { |
| "start": 446, |
| "end": 462, |
| "text": "Lee et al., 2001", |
| "ref_id": "BIBREF17" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 3: Surface Pattern Matching", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "Quantity-targeting questions are often underspecified and rely on culturally shared cooperativeness rules and/or world knowledge: Q: How many people live in Chile? S1: \"From our correspondent comes good news about the nine people living in Chile\u2026\" A1: nine While certainly nine people do live in Chile, we know what the questioner intends. We have hand-implemented a rule that provides default range assumptions for POPULATION questions and biases quantity questions accordingly.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 4: Expected Numerical Ranges", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "Abbreviations often follow a pattern: Q: What does NAFTA stand for? S1: \"This range of topics includes the North American Free Trade Agreement, NAFTA, and the world trade agreement GATT.\" S2: \"The interview now changed to the subject of trade and pending economic issues, such as the issue of opening the rice market, NAFTA, and the issue of Russia repaying economic cooperation funds.\" After Webclopedia identifies the qtarget as", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 5: Abbreviation Expansion", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "ABBREVIATION-EXPANSION,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 5: Abbreviation Expansion", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "it extracts possible answer candidates, including \"North American Free Trade Agreement\" from S1 and \"the rice market\" from S2. Rules for acronym matching easily prefer the former.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 5: Abbreviation Expansion", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "Phone numbers, zip codes, email addresses, URLs, and different types of quantities obey lexicographic patterns that can be exploited for matching, as in Q: What is the zip code for Fremont, CA? S1: \"\u2026from Everex Systems Inc., 48431 Milmont Drive, Fremont, CA 94538.\" and Q: How hot is the core of the earth? S1. \"The temperature of Earth's inner core may be as high as 9,000 degrees Fahrenheit (5,000 degrees Celsius).\" Webclopedia identifies the qtargets respectively as ZIP-CODE and TEMPERATURE-QUANTITY. Approx. 30 heuristics (cascaded) apply to the input before parsing to mark up numbers and other orthographically recognizable units of all kinds, including (likely) zip codes, quotations, year ranges, phone numbers, dates, times, scores, cardinal and ordinal numbers, etc. Similar work is reported in (Kwok et al., 2001 ).", |
| "cite_spans": [ |
| { |
| "start": 808, |
| "end": 826, |
| "text": "(Kwok et al., 2001", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 6: Semantic Type Matching", |
| "sec_num": "3.6" |
| }, |
| { |
| "text": "We have found a 10% increase in accuracy in answering definition questions by using external glosses obtained from WordNet. For Q: What is the Milky Way? Webclopedia identified two leading answer candidates:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 7: Definitions from WordNet", |
| "sec_num": "3.7" |
| }, |
| { |
| "text": "A1: outer regions A2: the galaxy that contains the Earth Comparing these with the WordNet gloss:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 7: Definitions from WordNet", |
| "sec_num": "3.7" |
| }, |
| { |
| "text": "WordNet: \"Milky Way-the galaxy containing the solar system\" allows Webclopedia to straightforwardly match the candidate with the greater word overlap. Curiously, the system also needs to use WordNet to answer questions involving common knowledge, as in: Q: What is the capital of the United States? because authors of the TREC collection do not find it necessary to explain what Washington is: Ex: \"Later in the day, the president returned to Washington, the capital of the United States.\" While WordNet's definition Wordnet: \"Washington-the capital of the United States\" directly provides the answer to the matcher, it also allows the IR module to focus its search on passages containing \"Washington\", \"capital\", and \"United States\", and the matcher to pick a good motivating passage in the source corpus.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 7: Definitions from WordNet", |
| "sec_num": "3.7" |
| }, |
| { |
| "text": "Clearly, this capability can be extended to include (definitional and other) information provided by other sources, including encyclopedias and the web (Lin 2002) .", |
| "cite_spans": [ |
| { |
| "start": 152, |
| "end": 162, |
| "text": "(Lin 2002)", |
| "ref_id": "BIBREF19" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 7: Definitions from WordNet", |
| "sec_num": "3.7" |
| }, |
| { |
| "text": "So far, we have considered individual words and groups of words. But often this is insufficient to accurately score an answer. As also noted in (Buchholz, 2001) Kennedy from a sixth floor window as the presidential motorcade passed.\" The CONTEX parser (Hermjakob, 1997; provides the semantic relations. The parser uses machine learning techniques to build a robust grammar that produces semantically annotated syntax parses of English (and Korean and Chinese) sentences at approx. 90% accuracy (Hermjakob, 1999) .", |
| "cite_spans": [ |
| { |
| "start": 144, |
| "end": 160, |
| "text": "(Buchholz, 2001)", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 252, |
| "end": 269, |
| "text": "(Hermjakob, 1997;", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 494, |
| "end": 511, |
| "text": "(Hermjakob, 1999)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 8: Semantic Relation Matching", |
| "sec_num": "3.8" |
| }, |
| { |
| "text": "The matcher compares the parse trees of S1 and S2 to that of the question. Both S1 and S2 receive credit for matching question words \"Lee Harvey Oswald\" and \"kill\" (underlined), as well as for finding an answer (bold) of the proper qtarget type (PROPER-PERSON). However, is the answer \"Jack Ruby\" or \"President John F. Kennedy\"? The only way to determine this is to consider the semantic relationship between these candidates and the verb \"kill\" (parse trees simplified, and only portions shown here): Although the PREDs of both S1 and S2 match that of the question \"killed\", only S1 matches \"Lee Harvey Oswald\" as the head of the logical OBJect. Thus for S1, the matcher awards additional credit to node [2] (Jack Ruby) for being the logical SUBJect of the killing (using anaphora resolution). In S2, the parse tree correctly records that node [13] (\"John F. Kennedy\") is not the object of the killing. Thus despite its being closer to \"killed\", the candidate in S2 receives no extra credit from semantic relation matching.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 8: Semantic Relation Matching", |
| "sec_num": "3.8" |
| }, |
| { |
| "text": "It is important to note that the matcher awards extra credit for each matching semantic relationship between two constituents, not only when everything matches. This granularity improves robustness in the case of partial matches.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 8: Semantic Relation Matching", |
| "sec_num": "3.8" |
| }, |
| { |
| "text": "Semantic relation matching applies not only to logical subjects and objects, but also to all other roles such as location, time, reason, etc. (for additional examples see http://www.isi.edu/ natural-language/projects/webclopedia/sem-relexamples.html). It also applies at not only the sentential level, but at all levels, such as post-modifying prepositional and pre-modifying determiner phrases Additionally, Webclopedia uses 10 lists of word variations with a total of 4029 entries for semantically related concepts such as \"to invent\", \"invention\" and \"inventor\", and rules for handling them. For example, via coercing \"invention\" to \"invent\", the system can give \"Johan Vaaler\" extra credit for being a likely logical subject of \"invention\": Q: Who invented the paper clip? Qtargets: PROPER-PERSON & PROPER-NAME, PROPER-ORGANIZATION S1: \"The paper clip, weighing a desk-crushing 1,320 pounds, is a faithful copy of Norwegian Johan Vaaler's 1899 invention, said Per Langaker of the Norwegian School of Management.\" while \"David\" actually loses points for being outside of the clausal scope of the inventing:", |
| "cite_spans": [ |
| { |
| "start": 787, |
| "end": 815, |
| "text": "PROPER-PERSON & PROPER-NAME,", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 8: Semantic Relation Matching", |
| "sec_num": "3.8" |
| }, |
| { |
| "text": "S2: \"'Like the guy who invented the safety pin, or the guy who invented the paper clip,' David added.\"", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 8: Semantic Relation Matching", |
| "sec_num": "3.8" |
| }, |
| { |
| "text": "Webclopedia also includes a typical windowbased scoring module that moves a window over the text and assigns a score to each window position depending on a variety of criteria (Hovy et al., 1999) . Unlike Chen et al., 2001 ), we have not developed a very sophisticated scoring function, preferring to focus on the modules that employ information deeper than the word level.", |
| "cite_spans": [ |
| { |
| "start": 176, |
| "end": 195, |
| "text": "(Hovy et al., 1999)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 205, |
| "end": 222, |
| "text": "Chen et al., 2001", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 9: Word Window Scoring", |
| "sec_num": "3.9" |
| }, |
| { |
| "text": "This method is applied only when no other method provides a sufficiently high-scoring answer. The window scoring function is S = (500/(500+w))*(1/r) * \u03a3[(\u03a3I 1.5*q*e*b*u ) 1.5 ] Factors:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 9: Word Window Scoring", |
| "sec_num": "3.9" |
| }, |
| { |
| "text": "w: window width (modulated by gaps of various lengths: \"white house\" \u2260 \"white car and house\"), r: rank of qtarget in list returned by CONTEX, I: window word information content (inverse log frequency score of each word), summed, q: # different question words matched, plus specific rewards (bonus q=3.0), e: penalty if word matches one of question word's WordNet synset items (e=0.8), b: bonus for matching main verb, proper names, certain target words (b=2.0), u: (value 0 or 1) indicates whether a word has been qsubsumed (\"subsumed\" by the qtarget) and should not contribute (again) to the score. For example, \"In what year did Columbus discover America?\" the qsubsumed words are \"what\" and \"year\".", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Type 9: Word Window Scoring", |
| "sec_num": "3.9" |
| }, |
| { |
| "text": "In TREC-10's QA track, Webclopedia received an overall Mean Reciprocal Rank (MRR) score of 0.435, which put it among the top 4 performers of the 68 entrants (the average MRR score for the main QA task was about 0.234). The pinpointing heuristics are fairly accurate: when Webclopedia finds answers, it usually ranks them in the first place (1 st place: 35.5%; 2 nd : 8.94%; 3 rd : 5.69%; 4 th : 3.05%; 5 th : 5.28%; not found: 41.87%).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Performance Evaluation", |
| "sec_num": "4." |
| }, |
| { |
| "text": "We determined the impact of each knowledge source on system performance, using the TREC-10 test corpus using the standard MRR scoring. We applied the system to the questions of each knowledge type separately, with and without its specific knowledge source/algorithm. Results are shown in Table 1 , columns A (without) and B (with). To indicate overall effect, we also show (in columns C and D) the percentage of questions in TREC-10 and -9 respecively of each knowledge type.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 288, |
| "end": 295, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Performance Evaluation", |
| "sec_num": "4." |
| }, |
| { |
| "text": "It is tempting to search for a single technique that will solve the whole problem (for example, Ittycheriah et al. (2001) focus on the subset of factoid questions answerable by NPs, and train a statistical model to perform NP-oriented answer pinpointing). Our experience, however, is that even factoid QA is varied enough to require various special-purpose techniques and knowledge. The theoretical limits of the various techniques are not known, though Light et al.'s (2001) interesting work begins to study this.", |
| "cite_spans": [ |
| { |
| "start": 96, |
| "end": 121, |
| "text": "Ittycheriah et al. (2001)", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 454, |
| "end": 475, |
| "text": "Light et al.'s (2001)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5." |
| }, |
| { |
| "text": "Column A: % questions of the knowledge type answered correctly without using knowlege Column B: % questions, now using knowledge Column C: % questions of type in TREC-10 Column D: % questions of type in TREC-9 Table 1 . Performance of knowledge sources. Semantic relation scores measured only on questions in which they could logically apply.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 210, |
| "end": 217, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5." |
| }, |
| { |
| "text": "We conclude that factoid QA performance can be significantly improved by the use of knowledge attuned to specific question types and specific information characteristics. Most of the techniques for exploiting this knowledge require learning to ensure robustness.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5." |
| }, |
| { |
| "text": "To improve performance beyond this, we believe a combination of going to the web and turning to deeper world knowledge and automated inference (Harabagiu et al., 2001) to be the answer. It remains an open question how much work these techniques would require, and what their payoff limits are.", |
| "cite_spans": [ |
| { |
| "start": 143, |
| "end": 167, |
| "text": "(Harabagiu et al., 2001)", |
| "ref_id": "BIBREF8" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5." |
| } |
| ], |
| "back_matter": [], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Answer Extraction", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Abney", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Collins", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Singhal", |
| "suffix": "" |
| } |
| ], |
| "year": 2000, |
| "venue": "Proceedings of the Applied Natural Language Processing Conference (ANLP-NAACL-00)", |
| "volume": "", |
| "issue": "", |
| "pages": "296--301", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Abney, S., M. Collins, and A. Singhal. 2000. Answer Extraction. Proceedings of the Applied Natural Language Processing Conference (ANLP- NAACL-00), Seattle, WA, 296-301.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "An Algorithm that Learns What's in a Name", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Bikel", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Schwartz", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Weischedel", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "Machine Learning-Special Issue on NL Learning", |
| "volume": "34", |
| "issue": "", |
| "pages": "1--3", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bikel, D., R. Schwartz, and R. Weischedel. 1999. An Algorithm that Learns What's in a Name. Machine Learning-Special Issue on NL Learning, 34, 1-3.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Data-Intensive Question Answering", |
| "authors": [ |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Brill", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Banko", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Dumais", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Ng", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the TREC-10 Conference. NIST", |
| "volume": "", |
| "issue": "", |
| "pages": "183--189", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Brill, E., J. Lin, M. Banko, S. Dumais, and A. Ng. 2001. Data-Intensive Question Answering. Proceedings of the TREC-10 Conference. NIST, Gaithersburg, MD, 183-189.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Using Grammatical Relations, Answer Frequencies and the World Wide Web for TREC Question Answering", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Buchholz", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the TREC-10 Conference. NIST", |
| "volume": "", |
| "issue": "", |
| "pages": "496--503", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Buchholz, S. 2001. Using Grammatical Relations, Answer Frequencies and the World Wide Web for TREC Question Answering. Proceedings of the TREC-10 Conference. NIST, 496-503.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "CNLP at TREC-10 QA Track", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [ |
| "R" |
| ], |
| "last": "Diekema", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [ |
| "D" |
| ], |
| "last": "Taffet", |
| "suffix": "" |
| }, |
| { |
| "first": "N", |
| "middle": [], |
| "last": "Mccracken", |
| "suffix": "" |
| }, |
| { |
| "first": "N", |
| "middle": [ |
| "Ercan" |
| ], |
| "last": "Ozgencil", |
| "suffix": "" |
| }, |
| { |
| "first": "O", |
| "middle": [], |
| "last": "Yilmazel", |
| "suffix": "" |
| }, |
| { |
| "first": "E", |
| "middle": [ |
| "D" |
| ], |
| "last": "Liddy", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the TREC-10 Conference. NIST", |
| "volume": "", |
| "issue": "", |
| "pages": "480--490", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Chen, J., A.R. Diekema, M.D. Taffet, N. McCracken, N. Ercan Ozgencil, O. Yilmazel, and E.D. Liddy. 2001. CNLP at TREC-10 QA Track. Proceedings of the TREC-10 Conference. NIST, 480-490.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Web Reinforced Question Answering", |
| "authors": [ |
| { |
| "first": "C", |
| "middle": [ |
| "L A" |
| ], |
| "last": "Clarke", |
| "suffix": "" |
| }, |
| { |
| "first": "G", |
| "middle": [ |
| "V" |
| ], |
| "last": "Cormack", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [ |
| "R" |
| ], |
| "last": "Lynam", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [ |
| "M" |
| ], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "G", |
| "middle": [ |
| "L" |
| ], |
| "last": "Mclearn", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the TREC-10 Conference. NIST", |
| "volume": "", |
| "issue": "", |
| "pages": "620--626", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Clarke, C.L.A., G.V. Cormack, T.R. Lynam, C.M. Li, and G.L. McLearn. 2001. Web Reinforced Question Answering. Proceedings of the TREC- 10 Conference. NIST, 620-626.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Exploiting Redundancy in Question Answering", |
| "authors": [ |
| { |
| "first": "C", |
| "middle": [ |
| "L A" |
| ], |
| "last": "Clarke", |
| "suffix": "" |
| }, |
| { |
| "first": "G", |
| "middle": [ |
| "V" |
| ], |
| "last": "Cormack", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [ |
| "R" |
| ], |
| "last": "Lynam", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the SIGIR Conference", |
| "volume": "", |
| "issue": "", |
| "pages": "358--365", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Clarke, C.L.A., G.V. Cormack, and T.R. Lynam. 2001. Exploiting Redundancy in Question Answering. Proceedings of the SIGIR Conference. New Orleans, LA, 358-365.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "WordNet: An Electronic Lexical Database", |
| "authors": [ |
| { |
| "first": "Ch", |
| "middle": [], |
| "last": "Fellbaum", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Fellbaum, Ch. (ed). 1998. WordNet: An Electronic Lexical Database. Cambridge: MIT Press.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "FALCON: Boosting Knowledge for Answer Engines", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Harabagiu", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Moldovan", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Pasca", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Mihalcea", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Surdeanu", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Buneascu", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "G\u00eerju", |
| "suffix": "" |
| }, |
| { |
| "first": "V", |
| "middle": [], |
| "last": "Rus", |
| "suffix": "" |
| }, |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Morarescu", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the 9 th Text Retrieval Conference (TREC-9", |
| "volume": "", |
| "issue": "", |
| "pages": "479--488", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Harabagiu, S., D. Moldovan, M. Pasca, R. Mihalcea, M. Surdeanu, R. Buneascu, R. G\u00eerju, V. Rus and P. Morarescu. 2001. FALCON: Boosting Knowledge for Answer Engines. Proceedings of the 9 th Text Retrieval Conference (TREC-9), NIST, 479-488.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Learning Parse and Translation Decisions from Examples with Rich Context. Ph.D. dissertation", |
| "authors": [ |
| { |
| "first": "U", |
| "middle": [], |
| "last": "Hermjakob", |
| "suffix": "" |
| } |
| ], |
| "year": 1997, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hermjakob, U. 1997. Learning Parse and Translation Decisions from Examples with Rich Context. Ph.D. dissertation, University of Texas Austin. file://ftp.cs.utexas.edu/pub/mooney/paper s/hermjakob-dissertation 97.ps.gz.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Parsing and Question Classification for Question Answering", |
| "authors": [ |
| { |
| "first": "U", |
| "middle": [], |
| "last": "Hermjakob", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the Workshop on Question Answering at ACL-2001", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hermjakob, U. 2001. Parsing and Question Classification for Question Answering. Proceedings of the Workshop on Question Answering at ACL-2001. Toulouse, France.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Question Answering in Webclopedia", |
| "authors": [ |
| { |
| "first": "E", |
| "middle": [ |
| "H" |
| ], |
| "last": "Hovy", |
| "suffix": "" |
| }, |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Gerber", |
| "suffix": "" |
| }, |
| { |
| "first": "U", |
| "middle": [], |
| "last": "Hermjakob", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Junk", |
| "suffix": "" |
| }, |
| { |
| "first": "C.-Y.", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "Proceedings of the TREC-9", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hovy, E.H., L. Gerber, U. Hermjakob, M. Junk, and C.-Y. Lin. 1999. Question Answering in Webclopedia. Proceedings of the TREC-9", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "A Question/Answer Typology with Surface Text Patterns", |
| "authors": [ |
| { |
| "first": "E", |
| "middle": [ |
| "H" |
| ], |
| "last": "Hovy", |
| "suffix": "" |
| }, |
| { |
| "first": "U", |
| "middle": [], |
| "last": "Hermjakob", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Ravichandran", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "Poster in Proceedings of the DARPA Human Language Technology Conference (HLT)", |
| "volume": "", |
| "issue": "", |
| "pages": "234--238", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hovy, E.H., U. Hermjakob, and D. Ravichandran. 2002. A Question/Answer Typology with Surface Text Patterns. Poster in Proceedings of the DARPA Human Language Technology Conference (HLT). San Diego, CA, 234-238.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "The Use of External Knowledge in Factoid QA", |
| "authors": [ |
| { |
| "first": "E", |
| "middle": [ |
| "H" |
| ], |
| "last": "Hovy", |
| "suffix": "" |
| }, |
| { |
| "first": "U", |
| "middle": [], |
| "last": "Hermjakob", |
| "suffix": "" |
| }, |
| { |
| "first": "C.-Y.", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the TREC-10 Conference. NIST", |
| "volume": "", |
| "issue": "", |
| "pages": "166--174", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hovy, E.H., U. Hermjakob, and C.-Y. Lin. 2001. The Use of External Knowledge in Factoid QA. Proceedings of the TREC-10 Conference. NIST, Gaithersburg, MD, 166-174.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "IBM's Statistical Question Answering System", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Ittycheriah", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Franz", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Roukos", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the TREC-10 Conference. NIST", |
| "volume": "", |
| "issue": "", |
| "pages": "317--323", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ittycheriah, A., M. Franz, and S. Roukos. 2001. IBM's Statistical Question Answering System. Proceedings of the TREC-10 Conference. NIST, Gaithersburg, MD, 317-323.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "TREC2001 Question-Answer, Web and Cross Language experiments using PIRCS", |
| "authors": [ |
| { |
| "first": "K", |
| "middle": [ |
| "L" |
| ], |
| "last": "Kwok", |
| "suffix": "" |
| }, |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Grunfeld", |
| "suffix": "" |
| }, |
| { |
| "first": "N", |
| "middle": [], |
| "last": "Dinstl", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Chan", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the TREC-10 Conference. NIST", |
| "volume": "", |
| "issue": "", |
| "pages": "447--451", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kwok, K.L., L. Grunfeld, N. Dinstl, and M. Chan. 2001. TREC2001 Question-Answer, Web and Cross Language experiments using PIRCS. Proceedings of the TREC-10 Conference. NIST, Gaithersburg, MD, 447-451.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "SiteQ: Engineering High Performance QA System Using Lexico=Semantic Pattern Matching and Shallow NLP", |
| "authors": [ |
| { |
| "first": "G", |
| "middle": [ |
| "G" |
| ], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Seo", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Jung", |
| "suffix": "" |
| }, |
| { |
| "first": "B-H", |
| "middle": [], |
| "last": "Cho", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "B-K", |
| "middle": [], |
| "last": "Kwak", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Cha", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "-H. An", |
| "suffix": "" |
| }, |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the TREC-10 Conference. NIST", |
| "volume": "", |
| "issue": "", |
| "pages": "437--446", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lee, G.G., J. Seo, S. Lee, H. Jung, B-H. Cho, C. Lee, B-K. Kwak, J, Cha, D. Kim, J-H. An, H. Kim, and K. Kim. 2001. SiteQ: Engineering High Performance QA System Using Lexico=Semantic Pattern Matching and Shallow NLP. Proceedings of the TREC-10 Conference. NIST, Gaithersburg, MD, 437-446.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Analyses for Elucidating Current Question Answering Technology", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Light", |
| "suffix": "" |
| }, |
| { |
| "first": "G", |
| "middle": [ |
| "S" |
| ], |
| "last": "Mann", |
| "suffix": "" |
| }, |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Riloff", |
| "suffix": "" |
| }, |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Breck", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Natural Language Engineering", |
| "volume": "7", |
| "issue": "4", |
| "pages": "325--342", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Light, M., G.S. Mann, E. Riloff, and E. Breck. 2001. Analyses for Elucidating Current Question Answering Technology. Natural Language Engineering, 7:4, 325-342.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "The Effectiveness of Dictionary and Web-Based Answer Reranking", |
| "authors": [ |
| { |
| "first": "C.-Y", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "Proceedings of the 19th International Conference on Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lin, C.-Y. 2002. The Effectiveness of Dictionary and Web-Based Answer Reranking. Proceedings of the 19th International Conference on Computational Linguistics (COLING 2002), Taipei, Taiwan.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "TREC-10 Experiments at KAIST: Batch Filtering and Question Answering", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [ |
| "H" |
| ], |
| "last": "Oh", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Ks", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [ |
| "S" |
| ], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [ |
| "W" |
| ], |
| "last": "Chang", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [ |
| "S" |
| ], |
| "last": "Seo", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Choi", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the TREC-10 Conference. NIST", |
| "volume": "", |
| "issue": "", |
| "pages": "354--361", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Oh, JH., KS. Lee, DS. Chang, CW. Seo, and KS. Choi. 2001. TREC-10 Experiments at KAIST: Batch Filtering and Question Answering. Proceedings of the TREC-10 Conference. NIST, Gaithersburg, MD, 354-361.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "One Search Engine or Two for Question Answering", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Prager", |
| "suffix": "" |
| }, |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Brown", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [ |
| "R" |
| ], |
| "last": "Radev", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Czuba", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "Proceedings of the TREC-9", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Prager, J., E. Brown, D.R. Radev, and K. Czuba. 1999. One Search Engine or Two for Question Answering. Proceedings of the TREC-9", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "Learning Surface Text Patterns for a Question Answering System", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Ravichandran", |
| "suffix": "" |
| }, |
| { |
| "first": "E", |
| "middle": [ |
| "H" |
| ], |
| "last": "Hovy", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "Proceedings of the ACL conference", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ravichandran, D. and E.H. Hovy. 2002. Learning Surface Text Patterns for a Question Answering System. Proceedings of the ACL conference. Philadelphia, PA.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "Patterns of Potential Answer Expressions as Clues to the Right Answer", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [ |
| "M" |
| ], |
| "last": "Soubbotin", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [ |
| "M" |
| ], |
| "last": "Soubbotin", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the TREC-10", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Soubbotin, M.M. and S.M. Soubbotin. 2001. Patterns of Potential Answer Expressions as Clues to the Right Answer. Proceedings of the TREC-10", |
| "links": null |
| }, |
| "BIBREF26": { |
| "ref_id": "b26", |
| "title": "A Question Answering System Supported by Information Extraction", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Srihari", |
| "suffix": "" |
| }, |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| } |
| ], |
| "year": 2000, |
| "venue": "Proceedings of the 1 st Meeting of the North American Chapter of the Association for Computational Linguistics (ANLP-NAACL-00)", |
| "volume": "", |
| "issue": "", |
| "pages": "166--172", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Srihari, R. and W. Li. 2000. A Question Answering System Supported by Information Extraction. Proceedings of the 1 st Meeting of the North American Chapter of the Association for Computational Linguistics (ANLP-NAACL-00), Seattle, WA, 166-172.", |
| "links": null |
| }, |
| "BIBREF27": { |
| "ref_id": "b27", |
| "title": "Overview of the Question Answering Track", |
| "authors": [ |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Voorhees", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "Proceedings of the TREC-9", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Voorhees, E. 1999. Overview of the Question Answering Track. Proceedings of the TREC-9", |
| "links": null |
| }, |
| "BIBREF29": { |
| "ref_id": "b29", |
| "title": "TREC-10 Experiments at CAS-ICT: Filtering, Web, and QA", |
| "authors": [ |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Xu", |
| "suffix": "" |
| }, |
| { |
| "first": "Z", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "X", |
| "middle": [], |
| "last": "Cheng", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Bu", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Bai", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the TREC-10 Conference. NIST", |
| "volume": "", |
| "issue": "", |
| "pages": "229--241", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wang, B., H. Xu, Z. Yang, Y. Liu, X. Cheng, D. Bu, and S. Bai. 2001. TREC-10 Experiments at CAS- ICT: Filtering, Web, and QA. Proceedings of the TREC-10 Conference. NIST, 229-241.", |
| "links": null |
| }, |
| "BIBREF30": { |
| "ref_id": "b30", |
| "title": "Managing Gigabytes: Compressing and Indexing Documents and Images", |
| "authors": [ |
| { |
| "first": "I", |
| "middle": [ |
| "H" |
| ], |
| "last": "Witten", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Moffat", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [ |
| "C" |
| ], |
| "last": "Bell", |
| "suffix": "" |
| } |
| ], |
| "year": 1994, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Witten, I.H., A. Moffat, and T.C. Bell. 1994. Managing Gigabytes: Compressing and Indexing Documents and Images. New York: Van Nostrand Reinhold.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "type_str": "figure", |
| "num": null, |
| "uris": null, |
| "text": "Webclopedia architecture." |
| } |
| } |
| } |
| } |