| { |
| "paper_id": "W93-0108", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T04:42:18.842831Z" |
| }, |
| "title": "Detecting Dependencies between Semantic Verb Subclasses and Subcategorization Frames in Text Corpora", |
| "authors": [ |
| { |
| "first": "Victor", |
| "middle": [], |
| "last": "Poznadski", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| }, |
| { |
| "first": "Antonio", |
| "middle": [], |
| "last": "Sanfilippo", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "We present a method for individuating dependencies between the semantic class of predicates and their associated subcategorization frames, and describe an implementation which allows the acquisition of such dependencies from bracketed texts.", |
| "pdf_parse": { |
| "paper_id": "W93-0108", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "We present a method for individuating dependencies between the semantic class of predicates and their associated subcategorization frames, and describe an implementation which allows the acquisition of such dependencies from bracketed texts.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "There is a widespread belief among linguists that a predicate's subcategorization frames are largely determined by its lexical-semantic properties [23, 11, 12] . Consider the domain of movement verbs. Following Talmy [23] , these can he semantically classified with reference to the meaning components: MOTION, MANNER, CAUSATION, THEME (MOVING ENTITY), PATH AND REFERENCE LOCATIONS (GOAL, SOURCE). Lexicalization patterns which arise from identifying clusters of such meaning components in verb senses can be systematically related to distinct subcategorization frames. 1 For example, the arguments of a verb expressing directed caused motion (e.g. bring, put, give) are normally a causative subject (agent), a theme direct object (moving entity) and a directional argument expressing path and reference location (goal), e.g.", |
| "cite_spans": [ |
| { |
| "start": 147, |
| "end": 151, |
| "text": "[23,", |
| "ref_id": "BIBREF22" |
| }, |
| { |
| "start": 152, |
| "end": 155, |
| "text": "11,", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 156, |
| "end": 159, |
| "text": "12]", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 217, |
| "end": 221, |
| "text": "[23]", |
| "ref_id": "BIBREF22" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "(1) Jackie will bring a bottle of retsina to the party CAUSER THEME PATH GOAL However, a motion verb which is not amenable to direct external causation [13] , will typically take a theme subject, with the possible addition of a directional argument, e.g.", |
| "cite_spans": [ |
| { |
| "start": 152, |
| "end": 156, |
| "text": "[13]", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "(2) The baby crawled (across the room)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Co-occurrence restrictions between meaning components may also preempt subcategorization options; for example, manner of motion verbs in Italian cannot integrate a completed path component and therefore never subcategorize for a directional argument, e.g.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "(3)*Carlo ha camminato a casa Carlo walked home These generalizations are important for NLP since they frequently cover large subclasses of lexical items and can be used both to reduce redundancy and elucidate significant aspects of lexical structure. Moreover, a precise characterization of the relation between semantic subclasses and subcategorization properties of verbs can aid lexical disambiguation. For example, the verb accord can be used in either one of two senses: agree or give, e,g.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "(4) a The two alibis do not accord Your alibi does not accord with his b They accorded him a warm welcome Accord is intransitive in the agree senses shown in (4a), and ditransitive in the give sense shown in (4b).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The manual encoding of subcategorization options for each choice of verb subclass in the language is very costly to develop and maintain. This problem can be alleviated by automatically extracting collocational information, e.g. grammar codes, from Machine Readable Dictionaries (MRDs). However, most of these dictionaries are not intended for such processing; their readership rarely require or desire such exhaustive and exacting precision. More specifically, the information available is in most cases compiled manually according to the lexicographer's intuitions rather than (semi-)automatically derived from texts recording actual language use. As a source of lexical information for NLP, MRDs are therefore liable to suffer from omissions, inconsistencies and occasional errors as well as being unable to cope with evolving usage [1, 4, 2, 6] . Ultimately, the maintenance costs involved in redressing such inadequacies are likely to reduce the initial appeal of generating subcategorization lists from MRDs.", |
| "cite_spans": [ |
| { |
| "start": 836, |
| "end": 839, |
| "text": "[1,", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 840, |
| "end": 842, |
| "text": "4,", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 843, |
| "end": 845, |
| "text": "2,", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 846, |
| "end": 848, |
| "text": "6]", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In keeping with these observations, we implemented a suite of programs which provide an integrated approach to lexical knowledge acquisition. The programs elicit dependencies between semantic verb classes and their admissible subcategorization frames using machine readable thesauri to assist in semantic tagging of texts.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Currently available dictionaries do not provide a sufficiently reliable source of lexical knowlege for NLP systems. This has led an increasing number of researchers to look at text corpora as a source of information [8, 22, 9, 6, 3] . For example, Brent [6] describes a program which retrieves subcategorization frames from untagged text. Brent's approach relies on detecting nominal, clausal and infinitive complements after identification of proper nouns and pronouns using predictions based on GB's Case Filter [16] --e.g. in English, a noun phrase occurs to the immediate left of a tensed verb, or the immediate right of a main verb or preposition. Brent's results are impressive considering that no text preprocessing (e.g. tagging or bracketing) is assumed. However, the number of subeategorization options recognized is minimal, 2 and it is hard to imagine how the approach could be extended to cover the full range of subcategorization possibilities without introducing some form of text preprocessing. Also, the phrasal patterns extracted are too impoverished to infer selectional restrictions as they only contain proper nouns and pronouns.", |
| "cite_spans": [ |
| { |
| "start": 216, |
| "end": 219, |
| "text": "[8,", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 220, |
| "end": 223, |
| "text": "22,", |
| "ref_id": "BIBREF21" |
| }, |
| { |
| "start": 224, |
| "end": 226, |
| "text": "9,", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 227, |
| "end": 229, |
| "text": "6,", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 230, |
| "end": 232, |
| "text": "3]", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 254, |
| "end": 257, |
| "text": "[6]", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 514, |
| "end": 518, |
| "text": "[16]", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background", |
| "sec_num": "2" |
| }, |
| { |
| "text": "2Brent's program recognizes five suhcategorization frames built out of three kinds of constituents: noun phrase, clause, infinitive.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Lexical acquisition of collocational information from preprocessed text is now becoming more popular as tools for analyzing corpora are getting to be more reliable [9] . For example, Basili el al. [3] present a method for acquiring sublanguage-specific selectional restrictions from corpora which uses text processing techniques such as morphological tagging and shallow syntactic analysis. Their approach relies on extracting word pairs and triples which represent crucial environments for the acquisition of selectional restrictions (e.g. V_prep_N(go, to, Boston)). They then replace words with semantic tags (V_prep_N(PHYSICAL_ACT-to-PLACE)) and compute co-occurrence preferences among them. Semantic tags are crucial for making generalizations about the types of words which can appear in a given context (e.g. as the argument of a verb or preposition). However, Basili et al. rely on manual encoding in the assignment of semantic tags; such a practice is bound to become more costly as the text under consideration grows in size and may prove prohibitively expensive with very large corpora. Furthermore, the semantic tags are allowed to vary from domain to domain (e.g. commercial and legal corpora) and are not hierarchically structured. With no consequent notion of subsumption, it might be impossible to identify \"families\" of tags relating to germane concepts across sublanguages (e.g. PHYSICAL_ACT, ACT; BUILDING, REAL_ESTATES).", |
| "cite_spans": [ |
| { |
| "start": 164, |
| "end": 167, |
| "text": "[9]", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 197, |
| "end": 200, |
| "text": "[3]", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background", |
| "sec_num": "2" |
| }, |
| { |
| "text": "In developing CorPSE (Corpus-based Predicate Structure Extractor) we followed Basili et al.'s idea of extracting semantically tagged phrasal frames from preprocessed text, but we used the Longman Lexicon of Contemporary English (LLOCE [15] ) to automate semantic tagging. LLOCE entries are similar to those of learner's dictionaries, but are arranged in a thesaurus-like fashion using semantic codes which provide a linguisticallymotivated classification of words. For example, [19] show that the semantic codes of LLOCE are instrumental in identifying members of the six subclasses of psychological predicates described in (5) [12, 11] . As shown in (6) , each verb representing a subclass has a code which often provides a uniform characterization of the subclass. We assigned semantic tags to Subcategorizatio, Frame tokens (SF tokens) extracted from the Penn Treebank [14, 20, 21] to produce Subcategorization Frame types (SF types).", |
| "cite_spans": [ |
| { |
| "start": 235, |
| "end": 239, |
| "text": "[15]", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 478, |
| "end": 482, |
| "text": "[19]", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 628, |
| "end": 632, |
| "text": "[12,", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 633, |
| "end": 636, |
| "text": "11]", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 651, |
| "end": 654, |
| "text": "(6)", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 872, |
| "end": 876, |
| "text": "[14,", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 877, |
| "end": 880, |
| "text": "20,", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 881, |
| "end": 884, |
| "text": "21]", |
| "ref_id": "BIBREF20" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "CorPSE: a Body of Programs for Acquiring Semantically Tagged Subcategorization Frames from Bracketed Texts", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Each SF type consists of a verb stem associated with one or more semantic tags, and a list of its (non-subject) complements, if any. The head of noun phrase complements were also semantically tagged. We used LLOCE collocational information --grammar codes --to reduce or remove semantic ambiguity arising from multiple assignment of tags to verb and noun stems. The structures below exemplify these three stages.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "CorPSE: a Body of Programs for Acquiring Semantically Tagged Subcategorization Frames from Bracketed Texts", |
| "sec_num": "3" |
| }, |
| { |
| "text": "SF type: ((\"deny\" (\"C193\"-refuse \"G127\"-reject)) ((*NP* (\"C\"-people_and_family)) (*NP* (\"N\"-general_and_abstract_terms)))) Disambiguated SF type: ((\"deny\" (\"C193\")) ((,NP, (\"c\")) (,NP, (\"r))))", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "SF token: ((DENY VB) (NP (ALIENS NNS)) (NP (*COMPOUND-NOUN* (STATE NN) (BENEFITS NNS))))", |
| "sec_num": null |
| }, |
| { |
| "text": "CorPSE is conceptually segmented into 2 parts: a predicate structure extractor, and a semantic processor. The predicate structure extractor takes bracketed text as input, and outputs SF tokens. The semantic processor converts SF tokens into SF types and disambiguates them.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "CorPSE's General Functionality", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "The predicate structure extractor elicits SF tokens from a bracketed input corpus. These tokens are formed from phrasal fragments which correspond to a subcategorization frame, factoring out the most relevant information. In the case of verbs, such fragments correspond to verb phrases where the following simplificatory changes have been applied: \u2022 clausal complements are reduced to a mention of the complementizer which introduces them, e.g. ((ARGUED VBD) (SBAR THAT))", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Extracting SF Tokens", |
| "sec_num": "3.1.1" |
| }, |
| { |
| "text": "\u2022", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Extracting SF Tokens", |
| "sec_num": "3.1.1" |
| }, |
| { |
| "text": "An important step in the extraction of SF tokens is to distinguish passive and active verb phrases. Passives are discriminated by locating a past participle following an auxiliary be.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Extracting SF Tokens", |
| "sec_num": "3.1.1" |
| }, |
| { |
| "text": "The semantic processor operates on the output of the predicate structure extractor. Inflected words in input SF tokens are first passed through a general purpose morphological analyser [17] and reduced to bare stems suitable for automated dictionary and lexicon searches. The next phase is to supplement SF tokens with semantic tags from LLOCE using the facilities of the ACQUILEX LDB [5, 7] and DCK [17] ; LLOCE tags are associated with verb stems and simply replace noun stems.", |
| "cite_spans": [ |
| { |
| "start": 185, |
| "end": 189, |
| "text": "[17]", |
| "ref_id": "BIBREF16" |
| }, |
| { |
| "start": 385, |
| "end": 388, |
| "text": "[5,", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 389, |
| "end": 391, |
| "text": "7]", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 400, |
| "end": 404, |
| "text": "[17]", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Converting SF Tokens into SF Types", |
| "sec_num": "3.1.2" |
| }, |
| { |
| "text": "The resulting SF structures are finally converted into SF types according to the representation system whose syntax is sketched in 7where: stem is the verb stem, parts a Coalescing merges SF types with isomorphic structure and identical verb stem, combining the semantic codes of NP-frames, e.g.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Converting SF Tokens into SF Types", |
| "sec_num": "3.1.2" |
| }, |
| { |
| "text": "((\"accord\" (\"D101 .... N226\") ) ((*PP* TO (*NP* (\"C\"))))) ((\"accord\" (\"D101 .... N226\")) =~ ((\"accord\" (\"D101 .... N226\")) ((*ee* TO (*NP* (\"G\"))))) ((*PP* TO (*liP* (\"C .... G\"))))) ((\"accord\" (\"D101 .... N226\")) ((*PP* TO (*NP* (\"C .... G\") ", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 224, |
| "end": 242, |
| "text": "(*NP* (\"C .... G\")", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Converting SF Tokens into SF Types", |
| "sec_num": "3.1.2" |
| }, |
| { |
| "text": "))))", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Converting SF Tokens into SF Types", |
| "sec_num": "3.1.2" |
| }, |
| { |
| "text": "This process can be performed in linear time when the input is lexicographically sorted.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Converting SF Tokens into SF Types", |
| "sec_num": "3.1.2" |
| }, |
| { |
| "text": "We employ two tag reduction methods. The first eliminates equivalent tags, the second applies syntactico-semantic restrictions using LLOCE grammar codes.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Converting SF Tokens into SF Types", |
| "sec_num": "3.1.2" |
| }, |
| { |
| "text": "More than one LLOCE code can apply to a particular entry. Under these circumstances, it may be possible to ignore one or more of them. We exploit the correlation between descriptors and semantic codes in the LLOCI'; index, reducing multiple codes indexed by the same descriptor to just one. More precisely, tile reduction involves substitution of all codes having equal descriptors with a new code which represents the logical conjunction of the substituted codes. This is shown in (8) where \"I28+N123\" is defined as the intersection of \"128\" and \"N123\" in the LLOCE hierarchy of semantics codes as indicated in (9).", |
| "cite_spans": [ |
| { |
| "start": 482, |
| "end": 485, |
| "text": "(8)", |
| "ref_id": "BIBREF7" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Converting SF Tokens into SF Types", |
| "sec_num": "3.1.2" |
| }, |
| { |
| "text": "((\"function\" (\"I28 .... N123\")) ((\"function\" (\"I28+N123\")) (8) ((*PP* LIKE (*NP* (\"C\"))))) ~ ((*PP* LIKE (*NP* (\"C\")))))", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Converting SF Tokens into SF Types", |
| "sec_num": "3.1.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "T I N \u00b0\u00b0\u00b0 \u2022 .. I28 N123 \"-\" I28+N123", |
| "eq_num": "(9)" |
| } |
| ], |
| "section": "Converting SF Tokens into SF Types", |
| "sec_num": "3.1.2" |
| }, |
| { |
| "text": "The second means for disambiguating SF types consists of filtering out the codes of verb stems which are incompatible with the type of subcategorization frame in which they occur. This is done by using collocational information provided in LLOCE. For example, the verb deny is assigned two distinct semantic codes which cannot be reduced to one as they have different descriptors:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Converting SF Tokens into SF Types", |
| "sec_num": "3.1.2" |
| }, |
| { |
| "text": "deny ... refuse v C193 reject v G127", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Converting SF Tokens into SF Types", |
| "sec_num": "3.1.2" |
| }, |
| { |
| "text": "The difference in semantic code entails distinct subcategorization options: deny can have a ditransitive subcategorization frame only in the refuse sense, e.g. Semantic codes which are incompatible with the SF types in which they occur, such as G127 in (12) , can thus be filtered out by enforcing constraints between SF type complement structures and LLOCE grammar codes. A set of 2-tuples of the form (verb-stem-semantic-code, grammar-codes) is formed by noting the LLOCE grammar codes for each semantic code that could apply to the verb stem. If the grammar codes of any 2-tuple have no intersection with the grammatical restrictions GC, we conclude that the associated verb-stem-semantic code is not possible. 3 For example, C193 in the SF type for deny in (13) is paired up with the grammar codes {D1;T1} and G127 with (T1,4,5;V3} according to the LLOCE entries for deny shown in 3This procedure is only effective if the corpus subcategorization information is equally or more precise than the dictionary information. For our corpus, it proved to be the case. (12) . The constraints in (14) would thus license automatic removal of semantic code G 127 from the SF type for ditransitive deny as shown in (15) . (14) ((\"deny\" ((C193, {D1, T1}) ~ GC = {D1} (G127, {T1, T4, T5, V3})) ) ((*NP* (\"C\")) (,NP, (\"r)))) (15) ((\"deny\" (\"C193\"-refuse \"G127\"-reject)) ~ ((\"deny\" (\"C193\")) ((*NP* (\"c\")) ((.Np. (\"c\"))", |
| "cite_spans": [ |
| { |
| "start": 253, |
| "end": 257, |
| "text": "(12)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 1065, |
| "end": 1069, |
| "text": "(12)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 1091, |
| "end": 1095, |
| "text": "(14)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 1207, |
| "end": 1211, |
| "text": "(15)", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 1214, |
| "end": 1218, |
| "text": "(14)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 1314, |
| "end": 1318, |
| "text": "(15)", |
| "ref_id": "BIBREF14" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Converting SF Tokens into SF Types", |
| "sec_num": "3.1.2" |
| }, |
| { |
| "text": "It may appear that there is a certain circularity in our work. We use grammar codes to help disambiguate SF types, but it might be argued that the corpus could not have been bracketed without some prior grammatical information: subcategorisation frames. This picture is inaccurate because our SF types provide collocational information which is not in LLOCE. For example, the SF type shown in (16a) captures the use of link in (16b); this subcategorization cannot be inferred from the LLOCE entry where no PP headed by to is mentioned. Indeed, another possible use for our system would be to provide feedback to an on-line dictionary. We also provide a partial indication of selectional restrictions, i.e. the semantic tags of NP complements. Furthermore, text can be bracketed using techniques such as stochastic and semi-automatic parsing which need not rely on exhaustive lists of subcategorisations.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "(,NP, ('T')))) (*NP* (\"r))))", |
| "sec_num": null |
| }, |
| { |
| "text": "In testing CorPSE, our main objectives were:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Using CorPSE: Emerging Trends and Current Limitations", |
| "sec_num": "4" |
| }, |
| { |
| "text": "\u2022 to assess the functionality of text pre-processing techniques involving automated semantic tagging and lexical disambiguation, and", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Using CorPSE: Emerging Trends and Current Limitations", |
| "sec_num": "4" |
| }, |
| { |
| "text": "\u2022 to show that such techniques may yield profitable results in capturing regularities in the syntax-semantics interface", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Using CorPSE: Emerging Trends and Current Limitations", |
| "sec_num": "4" |
| }, |
| { |
| "text": "In order to do this, we ran CorPSE on a section of the Penn Treebank comprising 576 bracketed sentences from radio transcripts. /,From these sentences, CorPSE extracted 1335 SF tokens comprising 1245 active VPs and 90 passives. The SF tokens were converted into 817 SF types. The coalescence process reduced the 817 SF types to 583, which are representative of 346 distinct verb stems. The verb stern of 308 of these 583 SF types was semantically ambiguous as it was associated with more than one semantic tag. In some cases, this ambiguity was appropriate because the semantic codes assigned to the stem were all compatible with the complement structure of their SF type. For example, the verb call can occur in either one of two senses, summon and phone, with no change in subcategorization structure:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Using CorPSE: Emerging Trends and Current Limitations", |
| "sec_num": "4" |
| }, |
| { |
| "text": "(17) a Supper is ready, call the kids b Call me when you land in Paris", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Using CorPSE: Emerging Trends and Current Limitations", |
| "sec_num": "4" |
| }, |
| { |
| "text": "In this case, CorPSE correctly maintains the ambiguity as shown in (18) . In other cases, the ambiguity was in need of resolution as some of the verb-stem's semantic codes referred to the same LLOCE entry or were incompatible with the complement structure in the SF type (see \u00a73.1.3). Disambiguation using semantic tag equivalence reduced the ambiguity of 206 types, totally disambiguating 31 stems. Applying collocation restrictions further reduced 38 stems, totally disambiguating 24 of them.", |
| "cite_spans": [ |
| { |
| "start": 67, |
| "end": 71, |
| "text": "(18)", |
| "ref_id": "BIBREF17" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Using CorPSE: Emerging Trends and Current Limitations", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Taking into account that the amount of data processed was too small to use statistical techniques for disambiguation, the results achieved are very promising: we managed to reduce ambiguity in over half the SF types and totally disambiguated 16 percent, thus providing a unique correspondence between semantic verb class and subcategorization frame in 346 cases. Of the remaining 179 SF frames, 106 had verb stems with two semantic codes, 72 had verb stems with 3-5 semantic codes and the verb stem of one SF type had 6. Needless to say, the number of ambiguous SF types is bound to increase as more texts are processed. However, as we accumulate more data, we will be able to apply statistical techniques to reduce ]exical ambiguity, e.g. by computing co-occurrence restrictions between the semantic codes of the verb stem and complement heads in SF types.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Using CorPSE: Emerging Trends and Current Limitations", |
| "sec_num": "4" |
| }, |
| { |
| "text": "The table below summarizes some of the results concerning the correlation of semantic codes and subcategorization options obtained by running CorPse on the Penn Treebank fragment. The first column lists the LLOCE semantic codes which are explained in (20) . The second column indicates the number of unique subcategorization occurrences for each code. A major difficulty in computing this relation was the presence of certain constituents as arguments that are usually thought of as adjuncts. For example, purpose clauses and time adverbials such as yesterday, all day, in March, on Friday had often been bracketed as arguments (i.e. sisters to a V node). Our solution was to filter out inadequately parsed arguments semi-automatically. Certain constituents were automatically filtered from SF types as their status as adjuncts was manifest, e.g. complements introduced by prepositions and complementizers such as without, as, since and because. Other suspect constituents, such as infinitive VPs which could represent purpose clauses, were processed by direct query. A second problem was the residual ambiguities in SF types mentioned above. These biased the significance of occurrences since one or more codes in an ambiguous SF type could be inconsistent with the subcategorization of the SF type. A measure of the \"noise\" factor introduced by ambiguous SF types is given in the third column of (19) , where ambiguity rate is computed by dividing the number of codes associated with the same complement structure by the number of occurrences of that code with any complement structure. This ambiguity measure allows the significance of the figures in the second column to be assessed. For example, since the occurrences of \"E\" instances were invariably ambiguous, it is difficult to draw reliable conclusions about them. Indeed, on referring most of these SF types (e.g. beat, bolt and have) back to their source texts, the \"Food & Drink\" connotation proved incorrect. The figuresin column 1 were normalised as percentages of the total number of occurrences in order to provide a measure of the statistical significance of the results in the remaining columns. We thus conclude that the results for B, E, H, and I are unlikely to be significant as they occur with low relative frequency and are highly ambiguous. The final three columns quantify the relative frequency of occurrence for VP, SBAR and PP complements in SF types for each semantic code. For example, the low frequency of VP and SBAR complements with code \"M\" reflects the relatively rare incidence of clausal arguments ill the semantics of motion and location verbs. By contrast, the relatively high frequency of PP complements with this code can be related to the semantic propensity of motion and location verbs to take spatial arguments.", |
| "cite_spans": [ |
| { |
| "start": 251, |
| "end": 255, |
| "text": "(20)", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 1398, |
| "end": 1402, |
| "text": "(19)", |
| "ref_id": "BIBREF18" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Using CorPSE: Emerging Trends and Current Limitations", |
| "sec_num": "4" |
| }, |
| { |
| "text": "The \"A\" verbs (eg. create, live and murder) appear to be strongly biased towards taking a direct object complement only. This might be due to the fact that these verbs involve creating, destroying or manipulating life rather than events. Finally, the overwhelmingly high frequency of SBAR complements with \"G\" verbs is related to the fact that thought and communication verbs typically involve individuals and states of affairs. We also found interesting results concerning the distribution of subcategorization options among specializations of the same general code. For example, 23 out of 130 occurrences of \"M\" verbs exhibited an \"NP PP\" complement structure; 17 of these were found in SF types with codes \"M50-M65\" which largely characterize verbs of caused directed motion: Putting and Taking, Pulling ~4 Pushing. This trend confirms some of the observations discussed in the introduction. It is now premature to report results of this kind more fully since the corpus data used was too small and genre-specific to make more reliable and detailed inferences about the relation between subcategorization and semantic verb subclass. We hope that further work with larger corpora will uncover new patterns and corroborate current correlations which at present can only be regarded as providing suggestive evidence. Other than using substantially larger texts, improvements could also be obtained by enriching SF types, e.g. by adding information about subject constituents.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Using CorPSE: Emerging Trends and Current Limitations", |
| "sec_num": "4" |
| }, |
| { |
| "text": "We have provided the building blocks for a system that combines the advantages of freetext processing of corpora with the more organised information found in MRDs, such as semantic tags and collocational information. We have shown how such a system can be used to acquire lexical knowledge in the form of semantically tagged subcategorization frames. These results can assist the automatic construction of lexicons for NLP, semantic tagging for data retrieval from textual databases as well as to help maintain, refine and augment MRDs.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5" |
| }, |
| { |
| "text": "1Following Levin[12] and Sanfilippo[18], we maintain that valency reduction processes (e.g. the causative-inchoative alternation) are semantically governed and thus do not weaken the correlation between verb semantics and subcategorization properties.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "Some of the work discussed in this paper was carried out at the Computer Laboratory in Cambridge within the context of the ACQUILEX project. The Penn-Treebank-data used were provided in CD-ROM format by the University of Pennsylvania through the ACL Data Collection Initiative (ACL/DCI CD-ROM I, September 1991). We are indebted to Inn Johnson for helpful comments and encouragement, and to John Beaven and Pete Whitelock for providing feedback on previous versions of this paper. Many thanks also to Ann Copestake and Victor Lesk for their invaluable contribution towards mounting LLOCE on the LDB.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgements", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Explicit and Implicit Information in Dictionaries", |
| "authors": [ |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Atkins", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Kegl", |
| "suffix": "" |
| }, |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Levin", |
| "suffix": "" |
| } |
| ], |
| "year": 1986, |
| "venue": "Advances in Lexicology, Proceedings of the Second Annual Conference of the Centre for the New OED", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Atkins, B., Kegl, J. ~z Levin, B. (1986) Explicit and Implicit Information in Dictionar- ies. In Advances in Lexicology, Proceedings of the Second Annual Conference of the Centre for the New OED, University of Waterloo, Waterloo, Ontario.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Admitting Impediments", |
| "authors": [ |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Atkins", |
| "suffix": "" |
| }, |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Levin", |
| "suffix": "" |
| } |
| ], |
| "year": 1991, |
| "venue": "Lexical Acquisition: Using On-Line Resources to Build a Lexicon", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Atkins, B. & Levin, B. (1991) Admitting Impediments. In Zernik, U. (ed.) Lexical Acquisition: Using On-Line Resources to Build a Lexicon., Lawrence Erlbaum Asso- ciates, Hillsdale, New Jersey.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Computational Lexicography: the Neat Examples and the Odd Exemplars", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Basili", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [ |
| "T" |
| ], |
| "last": "Pazienza", |
| "suffix": "" |
| }, |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Velardi", |
| "suffix": "" |
| } |
| ], |
| "year": 1992, |
| "venue": "Proc. 3rd Conference on Applied NLP", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Basili, R. and Pazienza, M. T. and Velardi, P. (1992). Computational Lexicography: the Neat Examples and the Odd Exemplars. In Proc. 3rd Conference on Applied NLP, Trento, Italy.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Utilising the LDOCE Grammar Codes", |
| "authors": [ |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Boguraev", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Briscoe", |
| "suffix": "" |
| } |
| ], |
| "year": 1989, |
| "venue": "Computational Lexicography for Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Boguraev, B. & Briscoe, T. (1989) Utilising the LDOCE Grammar Codes. In Boguraev, B. & Briscoe, T. (eds.) Computational Lexicography for Natural Language Processing. Longman, London.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Database Models for Computational Lexicography", |
| "authors": [ |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Boguraev", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Briscoe", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Carroll", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Copestake", |
| "suffix": "" |
| } |
| ], |
| "year": 1990, |
| "venue": "Proceedings off EURALEX IV", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Boguraev, B., Briscoe, T., Carroll, J. and Copestake, A. (1990) Database Models for Computational Lexicography. In Proceedings off EURALEX IV, M~ilaga, Spain.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Automatic Semantic Classification of Verbs from their Syntactic Contexts: An Implemented Classifier for Stativity", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [ |
| "R" |
| ], |
| "last": "Brent", |
| "suffix": "" |
| } |
| ], |
| "year": 1991, |
| "venue": "Proc 29th ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Brent, M R. (1991) Automatic Semantic Classification of Verbs from their Syntactic Contexts: An Implemented Classifier for Stativity. In Proc 29th ACL, University of California, Berkeley, California.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "The ACQUILEX Lexical Database System: System Description and User Manual", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Carroll", |
| "suffix": "" |
| } |
| ], |
| "year": 1992, |
| "venue": "The (Other) Cambridge ACQUILEX Papers", |
| "volume": "253", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Carroll, J. (1992). The ACQUILEX Lexical Database System: System Description and User Manual. In The (Other) Cambridge ACQUILEX Papers, Tit 253, University of Cambridge, Cambridge, UK.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Word Association Norms, Mutual Information and Lexicography", |
| "authors": [ |
| { |
| "first": "Kenneth", |
| "middle": [], |
| "last": "Church", |
| "suffix": "" |
| }, |
| { |
| "first": "Patrick", |
| "middle": [], |
| "last": "Hanks", |
| "suffix": "" |
| } |
| ], |
| "year": 1989, |
| "venue": "Proc 23rd ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "76--83", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Church, Kenneth and Hanks, Patrick (1989) Word Association Norms, Mutual Infor- mation and Lexicography. Proc 23rd ACL. pp. 76 -83.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Using Statistics in Lexical Analysis", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Church", |
| "suffix": "" |
| }, |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Gale", |
| "suffix": "" |
| }, |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Hanks", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Hindle", |
| "suffix": "" |
| } |
| ], |
| "year": 1991, |
| "venue": "In Lexical Acquisition, Zernik", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Church, K and Gale, W. and Hanks, P. and Hindle, D (1991) Using Statistics in Lexical Analysis. In Lexical Acquisition, Zernik, Uri, Ed. Erlbaum, Hillsdale, NJ.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Noun Classification from Predicate Argument Structures", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Hindle", |
| "suffix": "" |
| } |
| ], |
| "year": 1990, |
| "venue": "Proc 28th ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "268--275", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hindle, D. (1990). Noun Classification from Predicate Argument Structures. In Proc 28th ACL. pp. 268 -275.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Semantic Structures", |
| "authors": [ |
| { |
| "first": "1~", |
| "middle": [], |
| "last": "Jackendoff", |
| "suffix": "" |
| } |
| ], |
| "year": 1990, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jackendoff, 1~. (1990) Semantic Structures. MIT Press, Cambridge, Mass.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Towards a Lexical Organization of English Verbs. Ms", |
| "authors": [ |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Levin", |
| "suffix": "" |
| } |
| ], |
| "year": 1989, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Levin, B. (1989) Towards a Lexical Organization of English Verbs. Ms., Dept. of Linguistics, Northwestern University", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "The Lexical Semantics of Verbs in Motion: The Perspective from Unaccusativity", |
| "authors": [ |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Levin", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Rappaport", |
| "suffix": "" |
| } |
| ], |
| "year": 1991, |
| "venue": "Thematic Structure: Its Role in Grammar, Forts", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Levin, B. and Rappaport, M. (1991) The Lexical Semantics of Verbs in Motion: The Perspective from Unaccusativity. To appear in Roca, I. (ed.) Thematic Structure: Its Role in Grammar, Forts, Dordrecht.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Very Large Text Corpora: Whai Yon Can Do with Them, and How to Do It", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Liberman", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Marcus", |
| "suffix": "" |
| } |
| ], |
| "year": 1992, |
| "venue": "Tutorial notes, 30th ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Liberman, M. and Marcus, M (1992) Very Large Text Corpora: Whai Yon Can Do with Them, and How to Do It. Tutorial notes, 30th ACL, University of Delaware, Newark, Delaware.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Longman Lexicon of Contemporary English", |
| "authors": [ |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Mcarthur", |
| "suffix": "" |
| } |
| ], |
| "year": 1981, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "McArthur, T. (1981) Longman Lexicon of Contemporary English. Longman, London.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Specifying Reference to the Subject", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Rouvret", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [ |
| "R" |
| ], |
| "last": "Vergnaud", |
| "suffix": "" |
| } |
| ], |
| "year": 1980, |
| "venue": "In Linguistic Enquiry", |
| "volume": "", |
| "issue": "1", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Rouvret, A. and Vergnaud, J R. (1980) Specifying Reference to the Subject. In Lin- guistic Enquiry, 11(1).", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "A Morphological Analyser for English and Italian", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Sanfilippo", |
| "suffix": "" |
| } |
| ], |
| "year": 1992, |
| "venue": "The (Other) Cambridge ACQUILEX Papers", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sanfilippo, A (1992) A Morphological Analyser for English and Italian. In The (Other) Cambridge ACQUILEX Papers, TR 253, University of Cambridge, Cambridge, UK.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Verbal Diathesis: Knowledge Acquisition, Lexicon Construction and Dictionary Compilation. TR SLE/IT/93-11", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Sanfilippo", |
| "suffix": "" |
| } |
| ], |
| "year": 1993, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sanfilippo, A (1993) Verbal Diathesis: Knowledge Acquisition, Lexicon Construction and Dictionary Compilation. TR SLE/IT/93-11, Sharp Laboratories of Europe, Ox- ford, UK.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "The Acquisition of Lexical Knowledge from Combined Machine-Readable Dictionary Sources", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Sanfilippo", |
| "suffix": "" |
| }, |
| { |
| "first": "V", |
| "middle": [], |
| "last": "Poznafiski", |
| "suffix": "" |
| } |
| ], |
| "year": 1992, |
| "venue": "Proceedings off the 3rd Conference on Applied Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sanfilippo, A and Poznafiski, V. (1992) The Acquisition of Lexical Knowledge from Combined Machine-Readable Dictionary Sources. In Proceedings off the 3rd Conference on Applied Natural Language Processing, Trento.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Bracketing Guidelines for the Penn Treebank Project. Ms. University of Pennsylvania", |
| "authors": [ |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Santorini", |
| "suffix": "" |
| } |
| ], |
| "year": 1991, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Santorini, B. (1991) Bracketing Guidelines for the Penn Treebank Project. Ms. Uni- versity of Pennsylvania.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "Part-of-Spech Tagging Guidelines for the Penn Treebank Project", |
| "authors": [ |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Santorini", |
| "suffix": "" |
| } |
| ], |
| "year": 1991, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Santorini, B. (1991) Part-of-Spech Tagging Guidelines for the Penn Treebank Project. Ms. University of Pennsylvania.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "Automatically Extracting and Representing Collocations for Language Generation", |
| "authors": [ |
| { |
| "first": "F", |
| "middle": [ |
| "A" |
| ], |
| "last": "Smajda", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [ |
| "R" |
| ], |
| "last": "Mckeown", |
| "suffix": "" |
| } |
| ], |
| "year": 1990, |
| "venue": "Proc PSth A CL", |
| "volume": "", |
| "issue": "", |
| "pages": "252--259", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Smajda, F A. and McKeown, K. R. (1990) Automatically Extracting and Represent- ing Collocations for Language Generation. In Proc PSth A CL. pp. 252 -259.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "(ed) Language Typology and Syntactic Description 3. Grammatical Categories and the Lexicon", |
| "authors": [ |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Talmy", |
| "suffix": "" |
| } |
| ], |
| "year": 1985, |
| "venue": "CUP", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Talmy, L. Lexicalization Patterns: Semantic Structure in Lexical Form. In Shopen, T. (ed) Language Typology and Syntactic Description 3. Grammatical Categories and the Lexicon, CUP, 1985.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "num": null, |
| "text": "NP complements have been reduced to the head noun (or head nouns in the case of coordinated NP's or nominal compounds), e.g. ((FACES VBZ) (NP (CHARGES NNS))) \u2022 PP complements have been reduced to the head preposition plus the head of the complement noun phrase, e.g. ((RIDES VBZ) (PP IN ((VAIl Nil)))) * VP complements are reduced to a mention of the VFORM of the head verb, e.g. ((TRY VB) (VP TO))", |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "FIGREF1": { |
| "num": null, |
| "text": "possil)ly empty sequence of particles associated with the verb stem, {A ... N } is the set of LLOCE semantic codes, pforrn thehead of a prepositional phrase, compform the possibly empty complementizer of a clausal complement, and cat any category not covered by np-, pp-, sbar-and vp-frames. np-frame ::= pp-frame ::= sbar-frame ::= vp-frame ::= cat-frame ::= ( stem parts sem comps ) ({A...N }*) comp* ( { np-frame I pp-frame ] sbar-frame I vp-frame I cat-frame } ) ( *NP* sem ) ( *PP* pform comp ) ( *SBAIt* compform ) ( *VP* vform ) ( *CAT* cat ) 3.1.3 Disambiguating SF Types The disambiguation module of the semantic processor coalesces SF types, and reduces semantic tags when verb stems have several codes.", |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "FIGREF2": { |
| "num": null, |
| "text": "Republican senator David Lock's bill would permanently { deny (refuse) illegal *deny (reject) J aliens all State benefits The codependency between semantic verb class and subcategorization can often be inferred by the grammar code of LLOCE entries. For example, only the entry for the refuse sense of deny in LLOCE includes the grammar code D1 which signals a ditransitive subcategorization frame: (ll) C193 verbs: not letting or allowing deny [D1;T1] ... G127 verbs: rejecting... deny 1 [T1,4,5;V3] ...2 IT1] ...", |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "FIGREF3": { |
| "num": null, |
| "text": "( (\"deny\" (\"C193 .... G127\")) ((*NP* (\"C\")) (,~e, (\"~\"))))To automate this process, we first form a set GC of compatible grammar codes for each choice of complement structure in SF types. For example, the set of compatible grammar codes GC for any SF type with two noun phrase complements is restricted to the singleton set {D1}, e.g.", |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "FIGREF4": { |
| "num": null, |
| "text": "a ((\"link\" NIL (\"N\")) ((*NP, (\"c\")) (*PP* TO (*NP* (\"B .... N\"))))) b The arrest warrant issued in Florida ]inks the attorney to a government probe of the Medhyin drug cartel ...", |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "FIGREF5": { |
| "num": null, |
| "text": "((\"call\" (\"G\"-s-,mon \"M\"-phone))", |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "TABREF1": { |
| "num": null, |
| "text": ", experience... admire, respect, look up to ... fear, fear for, be frightened ... attract, interest, concern... fascinate, enthrall, enchant... frighten, scare, terrify... Moreover, LLOCE codes are conveniently arranged into a 3-tier hierarchy according to specificity, e.g. F Feelings, Emotions, Attitudes and Sensations F20-F40 Liking and not Liking F26 Attracting and Interesting very much fascinate, enthrall, enchant, charm, captivate The bottom layer of the hierarchy contains over 1500 domain-specific tags, the middle layer has 129 tags and the top (most general) layer has 14. Domain-specific tags are always linked to intermediate tags which are, in turn, linked to general tags. Thus we can tag sublanguages using domain-specific semantic codes (as do Basili et ai.) without generating unrelated sets of such codes.", |
| "type_str": "table", |
| "content": "<table><tr><td>Co~</td><td>Group Header</td><td>Entries</td><td>H</td></tr><tr><td/><td>Relating to feeling</td><td>feel, sense</td><td/></tr><tr><td/><td>Admiring and honouring</td><td/><td/></tr><tr><td/><td>Fear and Dread</td><td/><td/></tr><tr><td/><td>Attracting and interesting</td><td/><td/></tr><tr><td/><td>Attracting and interesting very much</td><td/><td/></tr><tr><td/><td>Frighten and panic</td><td/><td/></tr></table>", |
| "html": null |
| }, |
| "TABREF2": { |
| "num": null, |
| "text": "For example, the verb function is assigned two distinct codes in LLOCE: 128 functioning and serving, and N123 functioning and performing. Although I-and N:codes may in principle differ considerably, in this case they are very similar; indeed, the entries for the two codes are identical. This", |
| "type_str": "table", |
| "content": "<table><tr><td>give v D101</td></tr><tr><td>agree v N226</td></tr></table>", |
| "html": null |
| } |
| } |
| } |
| } |