| { |
| "paper_id": "E14-1037", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T10:40:42.205585Z" |
| }, |
| "title": "Empirically-motivated Generalizations of CCG Semantic Parsing Learning Algorithms", |
| "authors": [ |
| { |
| "first": "Jesse", |
| "middle": [], |
| "last": "Glass", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Temple University", |
| "location": { |
| "addrLine": "1801 N Broad Street Philadelphia", |
| "postCode": "19122", |
| "region": "PA", |
| "country": "USA" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Alexander", |
| "middle": [], |
| "last": "Yates", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Temple University", |
| "location": { |
| "addrLine": "1801 N Broad Street Philadelphia", |
| "postCode": "19122", |
| "region": "PA", |
| "country": "USA" |
| } |
| }, |
| "email": "ayates@gmail.com" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "Learning algorithms for semantic parsing have improved drastically over the past decade, as steady improvements on benchmark datasets have shown. In this paper we investigate whether they can generalize to a novel biomedical dataset that differs in important respects from the traditional geography and air travel benchmark datasets. Empirical results for two state-of-the-art PCCG semantic parsers indicates that learning algorithms are sensitive to the kinds of semantic and syntactic constructions used in a domain. In response, we develop a novel learning algorithm that can produce an effective semantic parser for geography, as well as a much better semantic parser for the biomedical dataset.", |
| "pdf_parse": { |
| "paper_id": "E14-1037", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "Learning algorithms for semantic parsing have improved drastically over the past decade, as steady improvements on benchmark datasets have shown. In this paper we investigate whether they can generalize to a novel biomedical dataset that differs in important respects from the traditional geography and air travel benchmark datasets. Empirical results for two state-of-the-art PCCG semantic parsers indicates that learning algorithms are sensitive to the kinds of semantic and syntactic constructions used in a domain. In response, we develop a novel learning algorithm that can produce an effective semantic parser for geography, as well as a much better semantic parser for the biomedical dataset.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Semantic parsing is the task of converting natural language utterances into formal representations of their meaning. In this paper, we consider in particular a grounded form of semantic parsing, in which the meaning representation language takes its logical constants from a given, fixed ontology. Several recent systems have demonstrated the ability to learn semantic parsers for domains like the GeoQuery database containing geography relations, or the ATIS database of air travel information. In these settings, existing systems can produce correct meaning representations with F1 scores approaching 0.9 (Wong and Mooney, 2007; Kwiatkowski et al., 2011) .", |
| "cite_spans": [ |
| { |
| "start": 607, |
| "end": 630, |
| "text": "(Wong and Mooney, 2007;", |
| "ref_id": "BIBREF23" |
| }, |
| { |
| "start": 631, |
| "end": 656, |
| "text": "Kwiatkowski et al., 2011)", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "These benchmark datasets have supported a diverse and influential line of research into semantic parsing learning algorithms for sophisticated semantic constructions, with continuing advances in accuracy. However, the focus on these datasets leads to a natural question -do other natural datasets have similar syntax and semantics, and if not, can existing algorithms handle the variability in syntax and semantics?", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In an effort to investigate and improve the generalization capacity of existing learning algorithms for semantic parsing, we develop a novel, natural experimental setting, and we test whether current semantic parsers generalize to the new setting. For our datset, we use descriptions of clinical trials of experimental drugs in the United States, available from the U.S. National Institutes of Health 1 . Much of the text in the description of these clinical trials can be mapped neatly onto biomedical ontologies, thus permitting grounded semantic analysis. Crucially, the dataset was not designed specifically with semantic parsing or question-answering in mind, and as a result, it provides a natural source for the variety and complexity of utterances that humans use in this domain. As an added benefit, a successful semantic parser in this domain could yield a variety of useful bioinformatics applications by permitting comparisons between and across clinical trials using structured representations of the data, rather than unstructured text.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this initial investigation of semantic parsing in this context, we ask:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 Can existing semantic parsing learning algorithms handle the variety and complexity of the clinical trials dataset? We show that two representative learning algorithms fare poorly on the clinical trials data: the best one achieves a 0.41 F1 in our tests.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 What types of constructions are the major cause of errors on the clinical trials dataset, and can semantic parsers be extended to handle them? While this initial investigation does not cover all types of constructions, we identify three important types of constructions that existing learning algorithms do not handle. We propose a new learning algorithm that can handle these types of constructions, and we demonstrate empirically that the new algorithm produces a semantic parser that improves by over 23 points in F1 on the clinical trials dataset compared with existing parsers.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The rest of this paper is organized as follows. The next section provides background information on CCG and semantic parsing. Section 3 describes the text and ontology that form the new clinical trials dataset for semantic parsing, as well as some of the problems that exising approaches have on this dataset. Sections 4 describes our semantic parsing model, and learning and inference algorithms. Section 5 presents our experiments and results, and Section 6 concludes.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Our approach to learning a semantic parser falls into the general framework of context-free Probabilistic Combinatory Categorial Grammars (PCCG) (Zettlemoyer and Collins, 2005) with typed lambda calculus expressions for the semantics. PCCG grammars involve lexical entries, which are weighted unary rewrite rules of the form Syntax : Semantics \u2192 Phrase. For example: Example Lexical Entries N P : melanoma \u2192 skin cancer S\\N P : \u03bbp\u03bbd.has condition(p, d) \u2192 patients with In addition to lexical rules, PCCG grammars involve weighted binary rewrite rules like the following:", |
| "cite_spans": [ |
| { |
| "start": 145, |
| "end": 176, |
| "text": "(Zettlemoyer and Collins, 2005)", |
| "ref_id": "BIBREF26" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background on Semantic Parsing with CCG", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Example CCG Grammar Rules X : f (g) \u2192 X/Y : f Y : g (function application) X : f (g) \u2192 Y : g X\\Y : f (backward application)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background on Semantic Parsing with CCG", |
| "sec_num": "2" |
| }, |
| { |
| "text": "These rules apply for any syntactic categories X and Y , and any logical forms f and g. The rules specify mechanisms for deducing syntactic categories for whole phrases based on their constituent parts. They also specify mechanisms for identifying semantics (logical forms) for phrases and sentences based on combinations of the semantics for the constituent parts. Besides function application, other ways to combine the semantics of children typically include conjunction, disjunction, function composition, and substitution, among others. Inference algorithms for PCCG can identify the best parse and logical form for a given sentence using standard dynamic programming algorithms for context-free grammars (Clark and Curran, 2007) .", |
| "cite_spans": [ |
| { |
| "start": 710, |
| "end": 734, |
| "text": "(Clark and Curran, 2007)", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background on Semantic Parsing with CCG", |
| "sec_num": "2" |
| }, |
| { |
| "text": "As a baseline in our experiments, we use a learning algorithm for semantic parsing known as Unification Based Learning (UBL) (Kwiatkowski et al., 2010) . Source code for UBL is freely available. Its authors found that the semantic parsers it learns achieve results competitive with the state-of-the-art on a variety of standard semantic parsing data sets, including GeoQuery (0.882 F1). UBL uses a log-linear probabilistic model P (L, T |S) over logical forms L and parse tree derivations T , given sentences S. During training, only S and L are observed, and UBL's gradient-based parameter estimation algorithm tries to maximize T P (L, T |S) over the training dataset. To learn lexicon entries, it adopts a search procedure that involves unification in higher-order logic. The objective of the search procedure is to identify lexical entries for the words in a sentence that, when combined with the lexical entries for other words in the sentence, will produce the observed logical form in the training data. For each training sentence, UBL heuristically explores the space of all possible lexical entries to produce a set of promising candidates, and adds them to the lexicon.", |
| "cite_spans": [ |
| { |
| "start": 125, |
| "end": 151, |
| "text": "(Kwiatkowski et al., 2010)", |
| "ref_id": "BIBREF14" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background on Semantic Parsing with CCG", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Our second baseline is an extension of this work, called Factored Unification Based Learning (FUBL) (Kwiatkowski et al., 2011) . Again, source code is freely available. FUBL factors the lexicon into a set of base lexical entries, and a set of templates that can construct more complex lexical entries from the base entries. This allows for a significantly more compact lexicon, as well as the ability to handle certain linguistic constructions, like ellipsis, that appear frequently in the ATIS dataset and which UBL struggles with. FUBL achieves an F1 of 0.82 on ATIS (compared with 66.3 for UBL), and an F1 of 0.886 on GeoQuery; both results are at or very near the best-reported results for those datasets.", |
| "cite_spans": [ |
| { |
| "start": 100, |
| "end": 126, |
| "text": "(Kwiatkowski et al., 2011)", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background on Semantic Parsing with CCG", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Many supervised learning frameworks have been applied to the task of learning a semantic parser, including inductive logic programming (Zelle and Mooney, 1996; Thompson and Mooney, 1999; Thompson and Mooney, 2003) , support vec-tor machine-based kernel approaches (Kate et al., 2005; Kate and Mooney, 2006; Kate and Mooney, 2007) , machine translation-style synchronous grammars (Wong and Mooney, 2007) , and context-free grammar-based approaches like probabilistic Combinatory Categorial Grammar (Zettlemoyer and Collins, 2005; Zettlemoyer and Collins, 2007; Zettlemoyer and Collins, 2009; Kwiatkowski et al., 2010; Kwiatkowski et al., 2011; Lu et al., 2008) and discriminative reranking (Ge and Mooney, 2006; Ge and Mooney, 2009) . These approaches have yielded steady improvements on standard test sets like GeoQuery. As far as we are aware, such systems have not been tested on domains besides ATIS and GeoQuery.", |
| "cite_spans": [ |
| { |
| "start": 135, |
| "end": 159, |
| "text": "(Zelle and Mooney, 1996;", |
| "ref_id": "BIBREF25" |
| }, |
| { |
| "start": 160, |
| "end": 186, |
| "text": "Thompson and Mooney, 1999;", |
| "ref_id": "BIBREF21" |
| }, |
| { |
| "start": 187, |
| "end": 213, |
| "text": "Thompson and Mooney, 2003)", |
| "ref_id": "BIBREF22" |
| }, |
| { |
| "start": 264, |
| "end": 283, |
| "text": "(Kate et al., 2005;", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 284, |
| "end": 306, |
| "text": "Kate and Mooney, 2006;", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 307, |
| "end": 329, |
| "text": "Kate and Mooney, 2007)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 379, |
| "end": 402, |
| "text": "(Wong and Mooney, 2007)", |
| "ref_id": "BIBREF23" |
| }, |
| { |
| "start": 497, |
| "end": 528, |
| "text": "(Zettlemoyer and Collins, 2005;", |
| "ref_id": "BIBREF26" |
| }, |
| { |
| "start": 529, |
| "end": 559, |
| "text": "Zettlemoyer and Collins, 2007;", |
| "ref_id": "BIBREF27" |
| }, |
| { |
| "start": 560, |
| "end": 590, |
| "text": "Zettlemoyer and Collins, 2009;", |
| "ref_id": "BIBREF28" |
| }, |
| { |
| "start": 591, |
| "end": 616, |
| "text": "Kwiatkowski et al., 2010;", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 617, |
| "end": 642, |
| "text": "Kwiatkowski et al., 2011;", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 643, |
| "end": 659, |
| "text": "Lu et al., 2008)", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 689, |
| "end": 710, |
| "text": "(Ge and Mooney, 2006;", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 711, |
| "end": 731, |
| "text": "Ge and Mooney, 2009)", |
| "ref_id": "BIBREF7" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Previous Work", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Because of the complexity involved in building a training dataset for a supervised semantic parser, there has been a recent push towards developing techniques which reduce the annotation cost or the data complexity of the models. Models have been developed which can handle some ambiguity in terms of which logical form is the correct label for each training sentence (Chen et al., 2010; Liang et al., 2009) . Another set of approaches have investigated the case where no logical forms are provided, but instead some form of feedback or response from the world is used as evidence for what the correct logical form must have been (Clarke et al., 2010; Liang et al., 2011; Artzi and Zettlemoyer, 2011) . Several projects have investigated unsupervised (Goldwasser et al., 2011; Poon, 2013; Krishnamurthy and Mitchell, 2012) and semi-supervised (Yahya et al., 2012; Cai and Yates, 2013) approaches. These techniques tend to handle either the same benchmark domains, or simpler questions over larger ontologies. While such techniques are important, their (unlabeled and labeled) sample complexity is higher than it could be, because the underlying grammars involved are not as general as they could be. Our work investigates techniques that will reduce this sample complexity.", |
| "cite_spans": [ |
| { |
| "start": 368, |
| "end": 387, |
| "text": "(Chen et al., 2010;", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 388, |
| "end": 407, |
| "text": "Liang et al., 2009)", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 630, |
| "end": 651, |
| "text": "(Clarke et al., 2010;", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 652, |
| "end": 671, |
| "text": "Liang et al., 2011;", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 672, |
| "end": 700, |
| "text": "Artzi and Zettlemoyer, 2011)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 751, |
| "end": 776, |
| "text": "(Goldwasser et al., 2011;", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 777, |
| "end": 788, |
| "text": "Poon, 2013;", |
| "ref_id": "BIBREF20" |
| }, |
| { |
| "start": 789, |
| "end": 822, |
| "text": "Krishnamurthy and Mitchell, 2012)", |
| "ref_id": null |
| }, |
| { |
| "start": 843, |
| "end": 863, |
| "text": "(Yahya et al., 2012;", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 864, |
| "end": 884, |
| "text": "Cai and Yates, 2013)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Previous Work", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Clinical trials are scientific experiments that measure the effects of a medical procedure, instrument, or product on humans. Since September 2009 in the United States, any clinical trial that is funded by the federal government must make its results publicly available online at clinicaltrials.gov. This site provides a wealth of biomedical text and structured data, which we use to produce a novel test set for semantic parsing.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Clinical Trials Dataset", |
| "sec_num": "3" |
| }, |
| { |
| "text": "We collected our utterances from a set of 47 random documents from clinicaltrials.gov. Many aspects of each study are reported in structured format; for example, the number of participants who were given a placebo and the number of participants who were given the intervention under consideration are both reported in a table in a standard format. However, certain crucial aspects of each study are reported only in text. Perhaps the most critical aspect of each study that is described only in text is the set of criteria for deciding who will be admitted to the study and who cannot be; these criteria are called inclusion criteria and exclusion criteria. We focus our semantic parsing tests on these criteria because they often form the longest portion of unstructured text for a given clinical trial report; because their meaning can be represented using a concise set of logical constants from a biomedical ontology; and because the criteria have a great deal of significance in the clinical trials domain. For example, these criteria are crucial for understanding why the results of two related studies about the same intervention might differ.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The text and ontology", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "The criteria for a study can be logically represented as a function of candidate test subjects that returns true if they match the study criteria, and false otherwise. We use a variant of lambda calculus over a typed ontology to represent each inclusion and exclusion criterion in our dataset. We randomly collected 803 utterances and manually labeled each using our representation language. 401 were used for training, 109 for development, and 293 for our final tests.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The text and ontology", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "To keep our semantic parsing study simple, we eschewed existing ontologies like UMLS (Bodenreider, 2004) that are large and overly-complex for this problem. We instead developed an ontology of 10 types, 38 relations and functions, and a dictionary of 591 named-entities to build the logical forms. The five most common types and relations in our dataset are listed in Table 1 . On average, the logical forms in our dataset involved 3.7 relations per logical form, typically joined with conjunction, implication, or disjunction. If accepted, both the full ontology and dataset will be made publicly available.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 368, |
| "end": 375, |
| "text": "Table 1", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "The text and ontology", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "We applied two state-of-the-art learning algorithms for learning PCCG semantic parsers -UBL and its extension, FUBL-to our training AST > 3 mg (i.e., include patients with a level of the AST enzyme in the blood of greater than 3 mg) \u03bbp . > (result(p, AST), unit(3, mg)) To help understand why state-of-the-art systems' performance differs so much from performance on benchmark datasets like GeoQuery, we performed an error analysis. Table 3 describes the most common errors we observed. The most common errors occurred on sentences containing coordination constructions, nested function applications, and for UBL, ellipsis, although a long tail of less common errors exists. FUBL manages to handle the elliptical constructions in Clinical Trials well, but not coordination or nested functions. Both systems tend to learn many, overly-specific lexical entries that include too much of the logical form in one lexical entry. For instance, from the coordination example in Table 3 , UBL learns a lexical entry for the word \"or\" that includes the logical form \u03bbp\u03bbd1\u03bbd2 . or(has condition(p, d1), has condition(p, d2)). While this entry works well when coordinating two diseases or conditions that patients must have, it will not work for coor-dinations between treatments or dates, or coordinations between diseases that patients should not have. UBL learns over 250 lexical entries for the word \"or\" from our training dataset of 401 sentences, each one with limited applicability to development sentences.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 433, |
| "end": 440, |
| "text": "Table 3", |
| "ref_id": "TABREF4" |
| }, |
| { |
| "start": 970, |
| "end": 977, |
| "text": "Table 3", |
| "ref_id": "TABREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Problems with semantic parsing the clinical trials data", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Based on these observed error types, we next develop novel learning procedures that properly handle coordination, nested function constructions, and ellipsis.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Problems with semantic parsing the clinical trials data", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Constructions in Clinical Trials Data", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning to Handle Complex", |
| "sec_num": "4" |
| }, |
| { |
| "text": "We introduce the GLL system for learning a semantic parser that generalizes to both GeoQuery and Clinical Trials data. The semantic parsing model involves a grammar that consists of a fixed set of binary CCG rewrite rules, a learned lexicon \u039b, and a new set T of learned templates for constructing unary type-raising rules. We call these templates for type-raising rules T-rules; these are described below in Section 4.4. Following Kwiatkowsi et al. 2010, we assign a probability P (L, T |S, \u03b8, G) to a logical form and parse tree for a sentence licensed by grammar G using a log-linear model with parameters \u03b8. We use a set of feature functions", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model and Inference", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "F (L, T, S) = (f 1 (L, T, S), . . . , f K (L, T, S)),", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model and Inference", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "where each f i counts the number of times that the ith grammar rule is used in the derivation of T and S. The probability of a particular logical form given a sentence and \u03b8 is given by:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model and Inference", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "P (L|S, \u03b8, G) = T exp( \u03b8 \u2022 F (L, T, S)) T ,L exp( \u03b8 \u2022 F (L , T , S))", |
| "eq_num": "(1)" |
| } |
| ], |
| "section": "Model and Inference", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "where the trees T (and T ) are restricted to those that are licensed by G and which produce L (L') as the logical form for the parent node of the tree. Inference is performed using standard dynamic programming algorithms for context-free parsing.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model and Inference", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "The input for the task of learning a semantic parser is a set of sentences S, where each S i \u2208 S has been labeled with a logical form L(S i ). We assume a fixed set of binary grammar productions, and use the training data to learn lexical entries, T-rules, and parameters. The training objective is to maximize the likelihood of the observed logical Clinical trials data has more coordination, especially noun phrase and adjective phrase coordination, than GeoQuery. forms, or to find G * and \u03b8 * such that:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Input: set of labeled sentences {(S i , L(S i ))}, ini- tial grammar G 0 , number of iterations M AX, learning rate \u03b1 \u039b \u2190 \u2205 \u2200i : \u039b \u2190 \u039b \u222a {S : L(S i ) \u2192 S i } G \u2190 G 0 \u222a \u039b \u03b8 \u2190 0 For iteration := 1 to M AX: T R \u2190 TRLEARN(G) Add dimension \u03b8 t to \u03b8 for t \u2208 T R \u2212 G G \u2190 G \u222a T R For each sentence S i : \u039b \u2190 LEXENTLEARN(S i , L(S i ), G) Add dimension \u03b8 \u03bb to \u03b8 for all \u03bb \u2208 \u039b\u2212G G \u2190 G \u222a \u039b \u03b8 \u2190 \u03b8 + \u03b1\u2207 i CLL Return G, \u03b8", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "G * , \u03b8 * = arg max G, \u03b8 i P (L(S i )|S i , \u03b8, G)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "This is a non-convex optimization problem. We use a greedy optimization procedure that iteratively updates G and \u03b8. Figure 1 shows an overview of the full algorithm. We use stochastic gradient updates to estimate parameters (LeCun et al., 1998) . For each example sentence S i in training, we compute the local gradient of the conditional log likelihood function CLL = log P (L(S i )|S i , \u03b8, G), and update \u03b8 by a step in the direction of this local gradient. The partial derivatives for this local gradient are:", |
| "cite_spans": [ |
| { |
| "start": 224, |
| "end": 244, |
| "text": "(LeCun et al., 1998)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 116, |
| "end": 124, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Learning", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "\u2202CLL \u2202\u03b8 j = E P (T |L(S i ),S i , \u03b8,G) f j (L(S i ), T, S i )\u2212 E P (T |S i , \u03b8,G) f j (L(S i ), T, S i )", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "We adopt a greedy approach to learning new lexical entries. We first identify in our current parse any high-scoring lexical entries that cover multiple words, and then look for new lexical rules for the sub-phrases covered by these lexical entries that could combine to create the current parse chart entry using the existing grammar rules. This requires searching through the grammar rules to find children nodes that the nonterminal could be the parent of. In general, this produces an intractably large set, because it requires taking the inverse of function application and function composition for forming the semantics of the nonterminal, and those inverses are intractably large. Figure 2 shows our algorithm for learning lexical entries, and Figure 3 shows the details of the critical component that generates the semantics of new potential lexical entries. For brevity, we omit the details of how we learn the syntax and mappings from semantics to words or phrases for new lexical entries, but these are borrowed from the existing techniques in UBL. The crucial difference from existing techniques is that the SPLITLEARN algorithm focuses on inverse function composition, while existing techniques focus on inverse The SPLITLEARN algorithm uses a GET-SUBEXPR subroutine to heuristically select only certain subexpressions of the input logical form for computing inverse composition. This is to avoid a combinatorial explosion in the number of learned splits of the input semantics. Mostly we consider any subexpression that forms an argument to some function in le.sem, but we take care to also include abstracted versions of these subexpressions, in which some of their arguments are in turn replaced by variables. The subroutine FREEVARS identifies all variables in a logical form that have no quantifier; REPEATVARS identifies all variables that appear at least twice. PC-SCORE looks for any entry in the parse chart that has a matching semantics and returns the score of that entry, or 0 if no matches are found. We use PCSCORE to measure the improvement (delta) in the score of the parse if it uses the two new lexical entries, rather than the previous single lexical entry. SPLITLEARN returns the set of lexical entries that tie for the largest improvement in the score of the parse.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 687, |
| "end": 695, |
| "text": "Figure 2", |
| "ref_id": null |
| }, |
| { |
| "start": 750, |
| "end": 758, |
| "text": "Figure 3", |
| "ref_id": "FIGREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Learning Lexical Entries with Inverse Function Composition", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Figures 4 and 5 illustrate the difference be-Input: lexical entry le, parse chart P C Entries \u2190 \u2205 For s \u2208GETSUBEXPR(le.sem): t \u2190 copy of s sem \u2190 copy of le.sem", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning Lexical Entries with Inverse Function Composition", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Apply[t] \u2190 \u2205 For v \u2208 FREEVARS(s)\u2229 REPEATVARS(sem ): Create variable v , t \u2190 t v sub for v", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning Lexical Entries with Inverse Function Composition", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Concatenate \"\u03bbv \" onto front of t", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning Lexical Entries with Inverse Function Composition", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Apply[t] \u2190 Apply[t] \u222a {v} For v \u2208 FREEVARS(t):", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning Lexical Entries with Inverse Function Composition", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Remove \"\u03bbv\" from front of sem Concatenate \"\u03bbv\" onto front of t Create new variable w sub \u2190 \"(w\" + each a \u2208 Apply[t] + \")\" sem \u2190 sem sub sub for s Concatenate \"\u03bbw\" onto front of sem tween SPLITLEARN and lexical entry learning for UBL and FUBL. For both example sentences, there is a point in the learning process where a logical form must be split using inverse function composition in order for useful lexical entries to be learned. At those points, UBL and FUBL split the logical forms using inverse function application, resulting in splits where the semantics of different lexemes are mixed together in the two resulting subexpressions. In Figure 4 , all three systems take the logical form \u03bbu.\u03bbp. > (result(p, bilirubin), unit(1.5, u)) and split it by removing some aspect of the final argument, unit(1.5, u), from the full expression. In UBL and FUBL, the term that is left behind in the full expression is something that unifies with \u03bbu.unit(1.5, u). In GLL, however, only a variable is left behind, since that variable can be replaced by \u03bbu. unit(1.5, u) through function composition to obtain the original expression. Thus GLL's split yields one significantly simpler subexpression, which in the end yields simpler lexical entries. In both figures, and in general for most parses we have observed, inverse function composition yields simpler and cleaner subexpressions. Figure 4 : An example of a sentence with nestedfunction semantics. GLL's lexical entry learning procedure correctly identifies the most general semantics for the lexeme >, while UBL and FUBL learn more specific and complex semantics.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 643, |
| "end": 651, |
| "text": "Figure 4", |
| "ref_id": null |
| }, |
| { |
| "start": 1378, |
| "end": 1386, |
| "text": "Figure 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Learning Lexical Entries with Inverse Function Composition", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Entries \u2190 Entries \u222a {t, sem } delta[t], delta[sem ] \u2190 PCSCORE(t) + PCSCORE(sem ) -PCSCORE(le) max \u2190 max x delta[x] Return max, {s \u2208 Entries | delta[s] = max}", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning Lexical Entries with Inverse Function Composition", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "We use T-rules to handle elliptical constructions. They are essentially a simplification of the factored lexicon used in FUBL that yields very similar results. Each T-rule \u03c4 \u2208 T is a function of the form \u03bbe . if type(e) then return Syn : f (e) \u2192 Syn : e, where type is a type from our ontology, Syn and Syn are two syntactic CCG categories or variables, and f is an arbitrary lambda calculus expression. For example, consider the T-rule \u03c4 = (\u03bbe . if disease(e) then return S\\N : \u03bbp . has condition(p, e) \u2192 N : e). When applied to the entity diabetes, this Trule results in an ordinary CCG rule: S\\N : \u03bbp . has condition(p, diabetes) \u2192 N : diabetes. Thus each T-rule is a template for constructing unary (type-raising) CCG grammar rules from an entity of the appropriate type. TRLEARN works by first identifying a set of entity symbols E that appear in multiple lexical entries in the input grammar G. Let the lexical entries for entity e \u2208 E be denoted by \u039b(e); thus, E consists of all entities where |\u039b(e)| \u2265 2. TRLEARN then looks for patterns in each of these sets of lexical entries. If one of the lexical entries in \u039b(e) has a semantics that consists of just e (for example, the lexical entry N : diabetes \u2192 diabetes), we create candidate T-rules from every other lexical entry l \u2208 \u039b(e) that has the same child, such as S\\N : \u03bbp . has condition(p, diabetes) \u2192 diabetes.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning T-rules", |
| "sec_num": "4.4" |
| }, |
| { |
| "text": "From this lexical entry, we create the candidate T-rule \u03c4 = Figure 5 : An example of a sentence with coordination semantics. GLL's lexical entry learning procedure correctly identifies the semantics for the lexeme or, while UBL and FUBL learn incorrect semantics.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 60, |
| "end": 68, |
| "text": "Figure 5", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Learning T-rules", |
| "sec_num": "4.4" |
| }, |
| { |
| "text": "(\u03bbx . if disease(x) then return S\\N : \u03bbp . has condition(p, x) \u2192 N : x). In general, the test in the if statement in the T-rule contains a check for the type of entity e. The right-hand side of the implication contains a unary grammar rule whose parent matches the parent of the rule in l , except that entity e has been replaced by a variable x. The child of the grammar rule matches the parent of the basic lexical entry N : e, except again that the entity e has been replaced by the variable x.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning T-rules", |
| "sec_num": "4.4" |
| }, |
| { |
| "text": "Having constructed a set of candidate T-rules from this process, TRLEARN must select the ones that will actually be added to the grammar. We use a test of selecting T-rules that cover at least M IN existing grammar rules in the input grammar G. In our implementation, we set M IN = 2. When parsing a sentence, the parser checks any parse chart entry for semantics that consist solely of an entity; for any such entry, it looks in a hashbased index for applicable T-rules, applies them to the entity to construct new unary grammar rules, and then applies the unary grammar rules to the parse chart entry to create new nonterminal nodes.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning T-rules", |
| "sec_num": "4.4" |
| }, |
| { |
| "text": "In our experiments, we test the generality of our learning algorithm by testing its ability to handle both GeoQuery and the Clinical Trials datasets.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments", |
| "sec_num": "5" |
| }, |
| { |
| "text": "The clinical trials dataset is described above in Section 3. Table 5 : On the clinical trials dataset, GLL outperforms UBL and FUBL by more than 23 points in F1, for a reduction in error (i.e., 1-F1) of nearly 40% over FUBL.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 61, |
| "end": 68, |
| "text": "Table 5", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Experimental setup", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "2400 geographical entities, such as nations, rivers, and mountains, as well as 8 geography relations, such as the location of a mountain, and whether one state borders another. The text for semantic parsing consists of a set of 880 geography questions, labeled with a lambda-calculus representation of the sentence's meaning. We follow the procedure described by Kwiatkowski et al.. (2010) in splitting these sentences into training, development, and test sentences. This dataset allows us to provide a comparison with other semantic parsers on a well-known dataset. We measured performance based on exact-match of the full logical form, modulo re-ordering of arguments to symmetric relations (like conjunction and disjunction).", |
| "cite_spans": [ |
| { |
| "start": 363, |
| "end": 389, |
| "text": "Kwiatkowski et al.. (2010)", |
| "ref_id": "BIBREF14" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental setup", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "Tables 4 and 5 show the results of semantic parsers learned by the UBL, FUBL, and GLL learning algorithms on the GeoQuery and clinical trials datasets, respectively. On the GeoQuery dataset, all three parsers perform very similarly, although GLL's performance is slightly worse. However, on the clinical trials dataset, GLL significantly outperforms both UBL and FUBL in terms of precision, recall, and F1. Of course, there clearly remain many syntactic and semantic constructions that none of these algorithms can currently handle, as all systems perform significantly worse on clinical trials than on GeoQuery. Tables 6 shows the overall size of UBL's and GLL's learned lexicons, and Table 7 shows the number of learned entries for selected lexical System GeoQuery Clinical Trials UBL 5,149 49,635 GLL 4,528 36,112 Table 6 : GLL learns a lexicon that is 27% smaller than UBL's lexicon on clinical trials data.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 686, |
| "end": 693, |
| "text": "Table 7", |
| "ref_id": null |
| }, |
| { |
| "start": 751, |
| "end": 833, |
| "text": "System GeoQuery Clinical Trials UBL 5,149 49,635 GLL 4,528 36,112 Table 6", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results and Discussion", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "Lexeme UBL meanings GLL meanings > 36 2 < 28 2 = 35 2 and 6 4 or 254 9 Table 7 : For certain common and critical lexical items in the clinical trials dataset, GLL learns far fewer (but more general) lexical entries; for the word \"or\", GLL learns only 3.5% of the entries that UBL learns.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 71, |
| "end": 78, |
| "text": "Table 7", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Lexicon Size", |
| "sec_num": null |
| }, |
| { |
| "text": "items that appear frequently in the clinical trials corpus. FUBL uses a factored lexicon in which the semantics of a logical form is split across two data structures. As a result, FUBL's lexicon is not directly comparable to the other systems, so for these comparisons we restrict our attention to UBL and GLL. UBL tends to learn far more lexical entries than GLL, particularly for words that appear in multiple sentences. Yet the poorer performance of UBL on clinical trials is an indication that these lexical entries are overly specific.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lexicon Size", |
| "sec_num": null |
| }, |
| { |
| "text": "We have introduced the clinical trials dataset, a naturally-occurring set of text where existing learning algorithms for semantic parsing struggle.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "6" |
| }, |
| { |
| "text": "Our new GLL algorithm uses a novel inverse function composition algorithm to handle coordination and nested function constructions, and pattern learning to handle elliptical constructions. These innovations allow GLL to handle GeoQuery and improve on clinical trials. Many sources of error on clinical trials remain for future research, including long-distance dependencies, attachment ambiguities, and coreference. In addition, further investigation is necessary to test how these algorithms handle additional domains and other types of natural linguistic constructions.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "6" |
| }, |
| { |
| "text": "clinicaltrials.gov", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "This work was supported by National Science Foundation grant 1218692. The authors appreciate the help that Anjan Nepal, Qingqing Cai, Avirup Sil, and Fei Huang provided.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgments", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Bootstrapping Semantic Parsers from Conversations", |
| "authors": [ |
| { |
| "first": "Yoav", |
| "middle": [], |
| "last": "Artzi", |
| "suffix": "" |
| }, |
| { |
| "first": "Luke", |
| "middle": [], |
| "last": "Zettlemoyer", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yoav Artzi and Luke Zettlemoyer. 2011. Bootstrap- ping Semantic Parsers from Conversations. In Pro- ceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP).", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "The Unified Medical Language System (UMLS): integrating biomedical terminology", |
| "authors": [ |
| { |
| "first": "Olivier", |
| "middle": [], |
| "last": "Bodenreider", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Nucleic Acids Research", |
| "volume": "32", |
| "issue": "", |
| "pages": "267--270", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Olivier Bodenreider. 2004. The Unified Medical Lan- guage System (UMLS): integrating biomedical ter- minology. Nucleic Acids Research, 32:D267-D270.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Large-scale Semantic Parsing via Schema Matching and Lexicon Extension", |
| "authors": [ |
| { |
| "first": "Qingqing", |
| "middle": [], |
| "last": "Cai", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexander", |
| "middle": [], |
| "last": "Yates", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the Annual Meeting of the Association for Computational Linguistics (ACL)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Qingqing Cai and Alexander Yates. 2013. Large-scale Semantic Parsing via Schema Matching and Lexi- con Extension. In Proceedings of the Annual Meet- ing of the Association for Computational Linguistics (ACL).", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Training a Multilingual Sportscaster: Using Perceptual Context to Learn Language", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [ |
| "L" |
| ], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Joohyun", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| }, |
| { |
| "first": "Raymond", |
| "middle": [ |
| "J" |
| ], |
| "last": "Mooney", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Journal of Artificial Intelligence Research", |
| "volume": "37", |
| "issue": "", |
| "pages": "397--435", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "David L. Chen, Joohyun Kim, and Raymond J. Mooney. 2010. Training a Multilingual Sportscaster: Using Perceptual Context to Learn Language. Journal of Artificial Intelligence Re- search, 37:397-435.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Widecoverage efficient statistical parsing with ccg and log-linear models", |
| "authors": [ |
| { |
| "first": "Stephen", |
| "middle": [], |
| "last": "Clark", |
| "suffix": "" |
| }, |
| { |
| "first": "James", |
| "middle": [ |
| "R" |
| ], |
| "last": "Curran", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Computational Linguistics", |
| "volume": "33", |
| "issue": "4", |
| "pages": "493--552", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Stephen Clark and James R. Curran. 2007. Wide- coverage efficient statistical parsing with ccg and log-linear models. Computational Linguistics, 33(4):493-552.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Driving semantic parsing from the world's response", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Clarke", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Goldwasser", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Chang", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Roth", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Computational Natural Language Learning (CoNLL)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "J. Clarke, D. Goldwasser, M. Chang, and D. Roth. 2010. Driving semantic parsing from the world's re- sponse. In Computational Natural Language Learn- ing (CoNLL).", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Discriminative Reranking for Semantic Parsing", |
| "authors": [ |
| { |
| "first": "Ruifang", |
| "middle": [], |
| "last": "Ge", |
| "suffix": "" |
| }, |
| { |
| "first": "Raymond", |
| "middle": [ |
| "J" |
| ], |
| "last": "Mooney", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of the 21st International Conference on Computational Linguistics and 44th Annual Meeting of the Association for Computational Linguistics (COLING/ACL-06)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ruifang Ge and Raymond J. Mooney. 2006. Discrim- inative Reranking for Semantic Parsing. In Pro- ceedings of the 21st International Conference on Computational Linguistics and 44th Annual Meet- ing of the Association for Computational Linguistics (COLING/ACL-06).", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Learning a Compositional Semantic Parser using an Existing Syntactic Parser", |
| "authors": [ |
| { |
| "first": "Ruifang", |
| "middle": [], |
| "last": "Ge", |
| "suffix": "" |
| }, |
| { |
| "first": "Raymond", |
| "middle": [ |
| "J" |
| ], |
| "last": "Mooney", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Joint Conference of the 47th Annual Meeting of the Association for Computational Linguistics and the 4th International Joint Conference on Natural Language Processing of the Asian Federation of Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ruifang Ge and Raymond J. Mooney. 2009. Learning a Compositional Semantic Parser using an Existing Syntactic Parser. In Joint Conference of the 47th An- nual Meeting of the Association for Computational Linguistics and the 4th International Joint Confer- ence on Natural Language Processing of the Asian Federation of Natural Language Processing (ACL- IJCNLP 2009).", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Confidence driven unsupervised semantic parsing", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Goldwasser", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Reichart", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Clarke", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Roth", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Association for Computational Linguistics (ACL)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "D. Goldwasser, R. Reichart, J. Clarke, and D. Roth. 2011. Confidence driven unsupervised semantic parsing. In Association for Computational Linguis- tics (ACL).", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Using String-Kernels for Learning Semantic Parsers", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Rohit", |
| "suffix": "" |
| }, |
| { |
| "first": "Raymond", |
| "middle": [ |
| "J" |
| ], |
| "last": "Kate", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Mooney", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of the 21st International Conference on Computational Linguistics and the 44th annual meeting of the ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Rohit J. Kate and Raymond J. Mooney. 2006. Us- ing String-Kernels for Learning Semantic Parsers. In Proceedings of the 21st International Conference on Computational Linguistics and the 44th annual meeting of the ACL.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Semi-Supervised Learning for Semantic Parsing using Support Vector Machines", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Rohit", |
| "suffix": "" |
| }, |
| { |
| "first": "Raymond", |
| "middle": [ |
| "J" |
| ], |
| "last": "Kate", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Mooney", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proceedings of the Human Language Technology Conference of the North American Chapter of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Rohit J. Kate and Raymond J. Mooney. 2007. Semi- Supervised Learning for Semantic Parsing using Support Vector Machines. In Proceedings of the Hu- man Language Technology Conference of the North American Chapter of the Association for Compu- tational Linguistics, Short Papers (NAACL/HLT- 2007).", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Learning to Transform Natural to Formal Languages", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Rohit", |
| "suffix": "" |
| }, |
| { |
| "first": "Yuk", |
| "middle": [ |
| "Wah" |
| ], |
| "last": "Kate", |
| "suffix": "" |
| }, |
| { |
| "first": "Raymond", |
| "middle": [ |
| "J" |
| ], |
| "last": "Wong", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Mooney", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Proceedings of the Twentieth National Conference on Artificial Intelligence (AAAI-05)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Rohit J. Kate, Yuk Wah Wong, and Raymond J. Mooney. 2005. Learning to Transform Natural to Formal Languages. In Proceedings of the Twen- tieth National Conference on Artificial Intelligence (AAAI-05).", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Weakly Supervised Training of Semantic Parsers", |
| "authors": [], |
| "year": null, |
| "venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Weakly Supervised Training of Semantic Parsers. In Proceedings of the Conference on Empirical Meth- ods in Natural Language Processing (EMNLP).", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Inducing Probabilistic CCG Grammars from Logical Form with Higherorder Unification", |
| "authors": [ |
| { |
| "first": "Tom", |
| "middle": [], |
| "last": "Kwiatkowski", |
| "suffix": "" |
| }, |
| { |
| "first": "Luke", |
| "middle": [], |
| "last": "Zettlemoyer", |
| "suffix": "" |
| }, |
| { |
| "first": "Sharon", |
| "middle": [], |
| "last": "Goldwater", |
| "suffix": "" |
| }, |
| { |
| "first": "Mark", |
| "middle": [], |
| "last": "Steedman", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tom Kwiatkowski, Luke Zettlemoyer, Sharon Goldwa- ter, and Mark Steedman. 2010. Inducing Probabilis- tic CCG Grammars from Logical Form with Higher- order Unification. In Proceedings of the Conference on Empirical Methods in Natural Language Pro- cessing (EMNLP).", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Lexical Generalization in CCG Grammar Induction for Semantic Parsing", |
| "authors": [ |
| { |
| "first": "Tom", |
| "middle": [], |
| "last": "Kwiatkowski", |
| "suffix": "" |
| }, |
| { |
| "first": "Luke", |
| "middle": [], |
| "last": "Zettlemoyer", |
| "suffix": "" |
| }, |
| { |
| "first": "Sharon", |
| "middle": [], |
| "last": "Goldwater", |
| "suffix": "" |
| }, |
| { |
| "first": "Mark", |
| "middle": [], |
| "last": "Steedman", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tom Kwiatkowski, Luke Zettlemoyer, Sharon Gold- water, and Mark Steedman. 2011. Lexical Gen- eralization in CCG Grammar Induction for Seman- tic Parsing. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP).", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Gradient-based learning applied to document recognition", |
| "authors": [ |
| { |
| "first": "Y", |
| "middle": [], |
| "last": "Lecun", |
| "suffix": "" |
| }, |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Bottou", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| }, |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Haffner", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "Proceedings of the IEEE", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. 1998. Gradient-based learning applied to document recog- nition. In Proceedings of the IEEE.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Learning semantic correspondences with less supervision", |
| "authors": [ |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Liang", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [ |
| "I" |
| ], |
| "last": "Jordan", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Klein", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Association for Computational Linguistics and International Joint Conference on Natural Language Processing (ACL-IJCNLP)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "P. Liang, M. I. Jordan, and D. Klein. 2009. Learning semantic correspondences with less supervision. In Association for Computational Linguistics and In- ternational Joint Conference on Natural Language Processing (ACL-IJCNLP).", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Learning dependency-based compositional semantics", |
| "authors": [ |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Liang", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [ |
| "I" |
| ], |
| "last": "Jordan", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Klein", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Association for Computational Linguistics (ACL)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "P. Liang, M. I. Jordan, and D. Klein. 2011. Learn- ing dependency-based compositional semantics. In Association for Computational Linguistics (ACL).", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "A Generative Model for Parsing Natural Language to Meaning Representations", |
| "authors": [ |
| { |
| "first": "Wei", |
| "middle": [], |
| "last": "Lu", |
| "suffix": "" |
| }, |
| { |
| "first": "Wee", |
| "middle": [], |
| "last": "Hwee Tou Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "Luke", |
| "middle": [ |
| "S" |
| ], |
| "last": "Sun Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Zettlemoyer", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proceedings of The Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wei Lu, Hwee Tou Ng, Wee Sun Lee, and Luke S. Zettlemoyer. 2008. A Generative Model for Parsing Natural Language to Meaning Representations. In Proceedings of The Conference on Empirical Meth- ods in Natural Language Processing (EMNLP).", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "Grounded Unsupervised Semantic Parsing", |
| "authors": [ |
| { |
| "first": "Hoifung", |
| "middle": [], |
| "last": "Poon", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the Annual Meeting of the Association for Computational Linguistics (ACL)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hoifung Poon. 2013. Grounded Unsupervised Se- mantic Parsing. In Proceedings of the Annual Meet- ing of the Association for Computational Linguistics (ACL).", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "Automatic construction of semantic lexicons for learning natural language interfaces", |
| "authors": [ |
| { |
| "first": "C", |
| "middle": [ |
| "A" |
| ], |
| "last": "Thompson", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [ |
| "J" |
| ], |
| "last": "Mooney", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "Proc. 16th National Conference on Artificial Intelligence (AAAI-99)", |
| "volume": "", |
| "issue": "", |
| "pages": "487--493", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "C.A. Thompson and R.J. Mooney. 1999. Automatic construction of semantic lexicons for learning natu- ral language interfaces. In Proc. 16th National Con- ference on Artificial Intelligence (AAAI-99), pages 487-493.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "Acquiring Word-Meaning Mappings for Natural Language Interfaces", |
| "authors": [ |
| { |
| "first": "Cynthia", |
| "middle": [ |
| "A" |
| ], |
| "last": "Thompson", |
| "suffix": "" |
| }, |
| { |
| "first": "Raymond", |
| "middle": [ |
| "J" |
| ], |
| "last": "Mooney", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Journal of Artificial Intelligence Research (JAIR)", |
| "volume": "18", |
| "issue": "", |
| "pages": "1--44", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Cynthia A. Thompson and Raymond J. Mooney. 2003. Acquiring Word-Meaning Mappings for Natural Language Interfaces. Journal of Artificial Intelli- gence Research (JAIR), 18:1-44.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "Learning Synchronous Grammars for Semantic Parsing with Lambda Calculus", |
| "authors": [ |
| { |
| "first": "Yuk", |
| "middle": [ |
| "Wah" |
| ], |
| "last": "Wong", |
| "suffix": "" |
| }, |
| { |
| "first": "Raymond", |
| "middle": [ |
| "J" |
| ], |
| "last": "Mooney", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proceedings of the 45th Annual Meeting of the Association for Computational Linguistics (ACL-2007)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yuk Wah Wong and Raymond J. Mooney. 2007. Learning Synchronous Grammars for Semantic Parsing with Lambda Calculus. In Proceedings of the 45th Annual Meeting of the Association for Com- putational Linguistics (ACL-2007).", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "Natural Language Questions for the Web of Data", |
| "authors": [ |
| { |
| "first": "Mohamed", |
| "middle": [], |
| "last": "Yahya", |
| "suffix": "" |
| }, |
| { |
| "first": "Klaus", |
| "middle": [], |
| "last": "Berberich", |
| "suffix": "" |
| }, |
| { |
| "first": "Shady", |
| "middle": [], |
| "last": "Elbassuoni", |
| "suffix": "" |
| }, |
| { |
| "first": "Maya", |
| "middle": [], |
| "last": "Ramanath", |
| "suffix": "" |
| }, |
| { |
| "first": "Gerhard", |
| "middle": [], |
| "last": "Volker Tresp", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Weikum", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Mohamed Yahya, Klaus Berberich, Shady Elbas- suoni, Maya Ramanath, Volker Tresp, and Gerhard Weikum. 2012. Natural Language Questions for the Web of Data. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP).", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "Learning to Parse Database Queries using Inductive Logic Programming", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "John", |
| "suffix": "" |
| }, |
| { |
| "first": "Raymond", |
| "middle": [ |
| "J" |
| ], |
| "last": "Zelle", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Mooney", |
| "suffix": "" |
| } |
| ], |
| "year": 1996, |
| "venue": "AAAI/IAAI", |
| "volume": "", |
| "issue": "", |
| "pages": "1050--1055", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "John M. Zelle and Raymond J. Mooney. 1996. Learn- ing to Parse Database Queries using Inductive Logic Programming. In AAAI/IAAI, pages 1050-1055.", |
| "links": null |
| }, |
| "BIBREF26": { |
| "ref_id": "b26", |
| "title": "Learning to Map Sentences to Logical Form: Structured Classification with Probabilistic Categorial Grammars", |
| "authors": [ |
| { |
| "first": "Luke", |
| "middle": [ |
| "S" |
| ], |
| "last": "Zettlemoyer", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Collins", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Proceedings of the Twenty First Conference on Uncertainty in Artificial Intelligence (UAI)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Luke S. Zettlemoyer and Michael Collins. 2005. Learning to Map Sentences to Logical Form: Struc- tured Classification with Probabilistic Categorial Grammars. In Proceedings of the Twenty First Conference on Uncertainty in Artificial Intelligence (UAI).", |
| "links": null |
| }, |
| "BIBREF27": { |
| "ref_id": "b27", |
| "title": "Online Learning of Relaxed CCG Grammars for Parsing to Logical Form", |
| "authors": [ |
| { |
| "first": "Luke", |
| "middle": [ |
| "S" |
| ], |
| "last": "Zettlemoyer", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Collins", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proceedings of the Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning (EMNLP-CoNLL)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Luke S. Zettlemoyer and Michael Collins. 2007. On- line Learning of Relaxed CCG Grammars for Pars- ing to Logical Form. In Proceedings of the Joint Conference on Empirical Methods in Natural Lan- guage Processing and Computational Natural Lan- guage Learning (EMNLP-CoNLL).", |
| "links": null |
| }, |
| "BIBREF28": { |
| "ref_id": "b28", |
| "title": "Learning Context-dependent Mappings from Sentences to Logical Form", |
| "authors": [ |
| { |
| "first": "Luke", |
| "middle": [ |
| "S" |
| ], |
| "last": "Zettlemoyer", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Collins", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Proceedings of the Joint Conference of the Association for Computational Linguistics and International Joint Conference on Natural Language Processing (ACL-IJCNLP)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Luke S. Zettlemoyer and Michael Collins. 2009. Learning Context-dependent Mappings from Sen- tences to Logical Form. In Proceedings of the Joint Conference of the Association for Computational Linguistics and International Joint Conference on Natural Language Processing (ACL-IJCNLP).", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "num": null, |
| "type_str": "figure", |
| "text": "The GLL Learning Algorithm. \u2207 i CLL indicates the local gradient of the conditional log likelihood at sentence S i .", |
| "uris": null |
| }, |
| "FIGREF1": { |
| "num": null, |
| "type_str": "figure", |
| "text": "Input: training sentence Sent, its logical form L, current grammar G Initialize: P C \u2190 parse chart from parsing Sent with G splits \u2190 \u2205For len := length(Sent) to 1: For pos := 0 to length(Sent) \u2212 len: e = arg max entry\u2208P C[len][pos] entry.score if e's only derivation is a lexical rule in G: (score, \u039b) \u2190 SPLITLEARN(e, P C) splits \u2190 splits \u222a {(score, \u039b)} split * \u2190 arg max split\u2208splits split.score Return split * .\u039bFigure 2: LEXENTLEARN Algorithm for learning lexical entries function application. While a priori both techniques are reasonable choices (and both work well on GeoQuery), our empirical results show that inverse function composition can learn the same semantic forms as inverse function application, but in addition can handle nested functions (which are function compositions) and coordination -a form of function composition if one views logical connectives like or as boolean functions.", |
| "uris": null |
| }, |
| "FIGREF2": { |
| "num": null, |
| "type_str": "figure", |
| "text": "SPLITLEARN Algorithm for generating (the semantics of) new lexical entries.", |
| "uris": null |
| }, |
| "FIGREF3": { |
| "num": null, |
| "type_str": "figure", |
| "text": "(has condition(p,pregnant), has condition(p,lactating)) \u03bbd'.\u03bbF.\u03bbd.\u03bbp.or(has condition(p,d'),F(p,d)) \u03bbF.\u03bbd.\u03bbp.or(has condition(p,pregnant),F(p,d)) .or(has condition(p,pregnant), has condition(p,d))\u03bbG.\u03bbd.or(G(pregnant),G(d))", |
| "uris": null |
| }, |
| "TABREF1": { |
| "num": null, |
| "type_str": "table", |
| "text": "Common types and functions in our ontology. In the example functions, t indicates boolean type, i indicates real values, p indicates person, d disease, and so on.", |
| "content": "<table><tr><td>patients with acute lymphoma</td></tr><tr><td>\u03bbp . has condition(p, acute(lymphoma))</td></tr><tr><td>hypertension</td></tr><tr><td>(i.e., include patients with hypertension)</td></tr><tr><td>\u03bbp . has condition(p, hypertension)</td></tr></table>", |
| "html": null |
| }, |
| "TABREF2": { |
| "num": null, |
| "type_str": "table", |
| "text": "Example utterances from the clinical trials dataset, and their logical forms. Paraphrases in parentheses do not appear in the actual data.data and tested the resulting parsers on development data. Results indicate that both systems have difficulty with the clinical trials datasets: FUBL achieves an F1 of 0.413, and UBL of just 0.281.", |
| "content": "<table/>", |
| "html": null |
| }, |
| "TABREF3": { |
| "num": null, |
| "type_str": "table", |
| "text": "27% patients > 18 years of age \u03bbp . > (result(age, p), unit(18, year))Many logical forms involve functions as arguments to other functions or relations.", |
| "content": "<table><tr><td>Error Type</td><td>Freq. Example</td><td>Description</td></tr><tr><td>Nested Funcs. Ellipsis</td><td>26% diabetes</td><td>Many examples in the inclusion (exclu-</td></tr><tr><td/><td>\u03bbp . has condition(p, diabetes)</td><td>sion) criteria simply list a disease or treat-</td></tr><tr><td/><td/><td>ment, with the understanding that a patient</td></tr><tr><td/><td/><td>p should be included (excluded) if p has</td></tr><tr><td/><td/><td>the disease or is undergoing the treatment.</td></tr><tr><td>Coordination</td><td>16% patient is pregnant or lactating</td><td/></tr><tr><td/><td>\u03bbp . or(has condition(p, pregnant),</td><td/></tr><tr><td/><td>has condition(p, lactating))</td><td/></tr></table>", |
| "html": null |
| }, |
| "TABREF4": { |
| "num": null, |
| "type_str": "table", |
| "text": "Three common kinds of utterances in the clinical trials development set that caused UBL and FUBL to make errors. Frequency indicates the percentage of all development examples that exhibited that type of construction.", |
| "content": "<table/>", |
| "html": null |
| } |
| } |
| } |
| } |