ACL-OCL / Base_JSON /prefixU /json /U07 /U07-1021.json
Benjamin Aw
Add updated pkl file v3
6fa4bc9
{
"paper_id": "U07-1021",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T03:08:51.341560Z"
},
"title": "Distributional Similarity of Multi-Word Expressions",
"authors": [
{
"first": "Laura",
"middle": [],
"last": "Ingram",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "The University of Sydney NSW 2006",
"location": {
"country": "Australia"
}
},
"email": ""
},
{
"first": "James",
"middle": [
"R"
],
"last": "Curran",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "The University of Sydney NSW 2006",
"location": {
"country": "Australia"
}
},
"email": ""
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "Most existing systems for automatically extracting lexical-semantic resources neglect multi-word expressions (MWEs), even though approximately 30% of gold-standard thesauri entries are MWEs. We present a distributional similarity system that identifies synonyms for MWEs. We extend Grefenstette's SEXTANT shallow parser to first identify bigram MWEs using collocation statistics from the Google WEB1T corpus. We extract contexts from WEB1T to increase coverage on the sparser bigrams.",
"pdf_parse": {
"paper_id": "U07-1021",
"_pdf_hash": "",
"abstract": [
{
"text": "Most existing systems for automatically extracting lexical-semantic resources neglect multi-word expressions (MWEs), even though approximately 30% of gold-standard thesauri entries are MWEs. We present a distributional similarity system that identifies synonyms for MWEs. We extend Grefenstette's SEXTANT shallow parser to first identify bigram MWEs using collocation statistics from the Google WEB1T corpus. We extract contexts from WEB1T to increase coverage on the sparser bigrams.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "Lexical-semantic resources, such as WordNet (Fellbaum, 1998) , are used in many applications in Natural Language Processing (NLP). Unfortunately, they are expensive and time-consuming to produce and are prone to bias and limited coverage. Automatically extracting these resources is crucial to overcoming the knowledge bottleneck in NLP.",
"cite_spans": [
{
"start": 44,
"end": 60,
"text": "(Fellbaum, 1998)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Existing distributional approaches to semantic similarity focus on unigrams, with very little work on extracting synonyms for multi-word expressions (MWEs). In this work, we extend an existing system to support MWEs by identifying bigram MWEs using collocation statistics (Manning and Sch\u00fctze, 1999) . These are calculated using n-gram counts from the Google WEB1T corpus (Brants and Franz, 2006) .",
"cite_spans": [
{
"start": 272,
"end": 299,
"text": "(Manning and Sch\u00fctze, 1999)",
"ref_id": "BIBREF11"
},
{
"start": 372,
"end": 396,
"text": "(Brants and Franz, 2006)",
"ref_id": "BIBREF3"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "We evaluate against several gold-standard thesauri and observe a slight decrease in overall performance when the bigram MWEs were included. This is unsurprising since the larger vocabulary and sparser contextual information for bigrams makes the task significantly harder. We also experimented with contexts extracted from WEB1T in an attempt to overcome the data sparseness problem. Inspection of the results for individual headwords revealed many cases where the synonyms returned were significantly better when bigram data was included.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Distributional similarity relies on the distributional hypothesis that similar terms appear in similar contexts (Harris, 1954) . Here we extend the SEXTANT parser (Grefenstette, 1994) to include multi-word terms and syntactic contexts. Curran (2004) experiments with different parsers for extracting contextual information, including SEXTANT, MINIPAR (Lin, 1994) , RASP (Briscoe and Carroll, 2002) , and CASS (Abney, 1996) . Lin (1998) used MINIPAR and Weeds (2003) used RASP for distributional similarity calculations. MINIPAR is the only parser to identify a range of MWEs that has been used for distributional similarity. Weeds (2003) and Curran (2004) evaluate measures for calculating distributional similarity. We follow (Curran, 2004) in using the weighted Jaccard measure with truncated t-test relation weighting for our experiments.",
"cite_spans": [
{
"start": 112,
"end": 126,
"text": "(Harris, 1954)",
"ref_id": "BIBREF8"
},
{
"start": 163,
"end": 183,
"text": "(Grefenstette, 1994)",
"ref_id": "BIBREF7"
},
{
"start": 236,
"end": 249,
"text": "Curran (2004)",
"ref_id": "BIBREF5"
},
{
"start": 351,
"end": 362,
"text": "(Lin, 1994)",
"ref_id": "BIBREF9"
},
{
"start": 370,
"end": 397,
"text": "(Briscoe and Carroll, 2002)",
"ref_id": "BIBREF4"
},
{
"start": 409,
"end": 422,
"text": "(Abney, 1996)",
"ref_id": "BIBREF0"
},
{
"start": 425,
"end": 435,
"text": "Lin (1998)",
"ref_id": "BIBREF10"
},
{
"start": 441,
"end": 465,
"text": "MINIPAR and Weeds (2003)",
"ref_id": null
},
{
"start": 625,
"end": 637,
"text": "Weeds (2003)",
"ref_id": "BIBREF14"
},
{
"start": 642,
"end": 655,
"text": "Curran (2004)",
"ref_id": "BIBREF5"
},
{
"start": 727,
"end": 741,
"text": "(Curran, 2004)",
"ref_id": "BIBREF5"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Background",
"sec_num": "2"
},
{
"text": "The initial step in creating a thesaurus for MWEs is to identify potential MWE headwords using collocation statistics. We used various statistical tests, e.g. the t-test and the log-likelihood test (Manning and Sch\u00fctze, 1999) , calculated over the Google WEB1T unigram and bigram counts. These counts, calculated over 1 trillion words of web text, gave the most reliable counts. However, highly ranked terms, e.g.",
"cite_spans": [
{
"start": 198,
"end": 225,
"text": "(Manning and Sch\u00fctze, 1999)",
"ref_id": "BIBREF11"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Detecting MWEs",
"sec_num": "3"
},
{
"text": "Contact Us and Site Map, demonstrate bias towards web-related terminology. This list of selected bigrams is used to detect bigrams within the BNC using a modified version of the Viterbi algorithm.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Detecting MWEs",
"sec_num": "3"
},
{
"text": "Grefenstette's (2004) (SEXTANT) parser was extended to extract contextual information for the list of selected bigrams extracted above. Adding these bigrams does not result in a substantial increase in the number of relations which implies that there is very little contextual information available about the bigram data. This has a significant impact on the difficulty of the task.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Context Extraction",
"sec_num": "4"
},
{
"text": "Experiments were also conducted whereby the contextual information was extracted from the WEB1T 3, 4 and 5-gram data for a list of known bigrams from the gold-standard thesauri. This data lacks the syntactic information provided by SEX-TANT but the counts are estimated over 10,000 times as much data. This should reduce the sparseness problem for the bigram headwords.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Context Extraction",
"sec_num": "4"
},
{
"text": "Following Curran (2004) , the extracted synonyms are compared directly against multiple goldstandard thesauri. We extend this evaluation to include multi-word headwords and synonyms. We randomly selected 300 unigram and 300 bigram headwords from the MAQCUARIE (Bernard, 1990), MOBY (Ward, 1996) , and ROGET'S (1911) thesauri, and WORDNET (Fellbaum, 1998) .",
"cite_spans": [
{
"start": 10,
"end": 23,
"text": "Curran (2004)",
"ref_id": "BIBREF5"
},
{
"start": 282,
"end": 294,
"text": "(Ward, 1996)",
"ref_id": "BIBREF13"
},
{
"start": 301,
"end": 329,
"text": "ROGET'S (1911) thesauri, and",
"ref_id": null
},
{
"start": 330,
"end": 354,
"text": "WORDNET (Fellbaum, 1998)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Synonym Extraction",
"sec_num": "5"
},
{
"text": "We calculated the number of direct matches against the gold standard (DIRECT) and the inverse rank (INVR), the sum of the reciprocal ranks of matches. The results for the unigram headword experiments are summarised in Table 1. Both INVR and DIRECT demonstrate that performance decreases when MWEs are included. However, performance did increase significantly for some terms when MWEs were added. For example, tool improved from 0.270 to 0.568 INVR. The results for rate, shown in Table 2 , also improved.",
"cite_spans": [],
"ref_spans": [
{
"start": 218,
"end": 226,
"text": "Table 1.",
"ref_id": "TABREF1"
},
{
"start": 480,
"end": 487,
"text": "Table 2",
"ref_id": "TABREF2"
}
],
"eq_spans": [],
"section": "Synonym Extraction",
"sec_num": "5"
},
{
"text": "The next set of experiments extracted synonyms for 300 bigram headwords drawn from the MAC-QUARIE thesaurus. The best results for bigram headwords was achieved when unigram and bigram data (Baldwin and Villavicencio, 2002) was included. Table 3 shows the top 5 synonyms (as ranked by the Jaccard measure) for atomic bomb and dining table.",
"cite_spans": [
{
"start": 189,
"end": 222,
"text": "(Baldwin and Villavicencio, 2002)",
"ref_id": "BIBREF1"
}
],
"ref_spans": [
{
"start": 237,
"end": 244,
"text": "Table 3",
"ref_id": "TABREF3"
}
],
"eq_spans": [],
"section": "Synonym Extraction",
"sec_num": "5"
},
{
"text": "We have integrated the identification of simple multi-word expressions (MWEs) with a state-of-theart distributional similarity system. We evaluated extracted synonyms for both unigram and bigram headwords against a gold standard consisting of the union of multiple thesauri. The main difficulties are the sparsity of distributional evidence for MWEs and their low coverage in the gold standard. These preliminary experiments show the potential of distributional similarity for extracting lexical-semantic resources for both unigrams and MWEs.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion",
"sec_num": "6"
},
{
"text": "Proceedings of the Australasian Language TechnologyWorkshop 2007, pages 146-148",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
}
],
"back_matter": [],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "Partial parsing via finite-state cascades",
"authors": [
{
"first": "Steven",
"middle": [],
"last": "Abney",
"suffix": ""
}
],
"year": 1996,
"venue": "Journal of Natural Language Engineering",
"volume": "2",
"issue": "4",
"pages": "337--344",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Steven Abney. 1996. Partial parsing via finite-state cascades. Journal of Natural Language Engineering, 2(4):337-344.",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "Extracting the unextractable: A case study on verbparticles",
"authors": [
{
"first": "Timothy",
"middle": [],
"last": "Baldwin",
"suffix": ""
},
{
"first": "Aline",
"middle": [],
"last": "Villavicencio",
"suffix": ""
}
],
"year": 2002,
"venue": "Proceedings of the Sixth Conference on Computational Natural Language Learning",
"volume": "",
"issue": "",
"pages": "98--104",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Timothy Baldwin and Aline Villavicencio. 2002. Ex- tracting the unextractable: A case study on verb- particles. In Proceedings of the Sixth Conference on Computational Natural Language Learning (CoNLL 2002), pages 98-104, Taipei, Taiwan.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "The Macquarie Encyclopedic Thesaurus. The Macquarie Library",
"authors": [
{
"first": "R",
"middle": [
"L"
],
"last": "John",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Bernard",
"suffix": ""
}
],
"year": 1990,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "John R.L. Bernard, editor. 1990. The Macquarie Ency- clopedic Thesaurus. The Macquarie Library, Sydney, Australia.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "Web 1T 5-gram corpus version 1",
"authors": [
{
"first": "Thorsten",
"middle": [],
"last": "Brants",
"suffix": ""
},
{
"first": "Alex",
"middle": [],
"last": "Franz",
"suffix": ""
}
],
"year": 2006,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Thorsten Brants and Alex Franz. 2006. Web 1T 5-gram corpus version 1. Technical report, Google Inc.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "Robust accurate statistical annotation of general text",
"authors": [
{
"first": "Ted",
"middle": [],
"last": "Briscoe",
"suffix": ""
},
{
"first": "John",
"middle": [],
"last": "Carroll",
"suffix": ""
}
],
"year": 2002,
"venue": "Proceedings of the Third International Conference on Language Resources and Evaluation",
"volume": "",
"issue": "",
"pages": "29--31",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Ted Briscoe and John Carroll. 2002. Robust accurate sta- tistical annotation of general text. In Proceedings of the Third International Conference on Language Re- sources and Evaluation, pages 1499-1504, Las Palmas de Gran Canaria, 29-31 May.",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "From Distributional to Semantic Similarity",
"authors": [
{
"first": "James",
"middle": [
"R"
],
"last": "Curran",
"suffix": ""
}
],
"year": 2004,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "James R. Curran. 2004. From Distributional to Semantic Similarity. Ph.D. thesis, University of Edinburgh.",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "Wordnet: An Electronic Lexical Database",
"authors": [],
"year": 1998,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Chrisriane Fellbaum, editor. 1998. Wordnet: An Elec- tronic Lexical Database. MIT Press, Cambridge.",
"links": null
},
"BIBREF7": {
"ref_id": "b7",
"title": "Explorations in Automatic Thesaurus Discovery",
"authors": [
{
"first": "Gregory",
"middle": [],
"last": "Grefenstette",
"suffix": ""
}
],
"year": 1994,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Gregory Grefenstette. 1994. Explorations in Automatic Thesaurus Discovery. Kluwer Academic Publishers, USA.",
"links": null
},
"BIBREF8": {
"ref_id": "b8",
"title": "Distributional structure. Word",
"authors": [
{
"first": "Zellig",
"middle": [],
"last": "Harris",
"suffix": ""
}
],
"year": 1954,
"venue": "",
"volume": "10",
"issue": "",
"pages": "146--162",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Zellig Harris. 1954. Distributional structure. Word, 10(2/3):146-162.",
"links": null
},
"BIBREF9": {
"ref_id": "b9",
"title": "Principar -an efficient, broadcoverage, principle-based parser",
"authors": [
{
"first": "Dekang",
"middle": [],
"last": "Lin",
"suffix": ""
}
],
"year": 1994,
"venue": "Proceedings of COLING-94",
"volume": "",
"issue": "",
"pages": "482--488",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Dekang Lin. 1994. Principar -an efficient, broad- coverage, principle-based parser. In Proceedings of COLING-94, pages 482-488, Kyoto, Japan.",
"links": null
},
"BIBREF10": {
"ref_id": "b10",
"title": "Automatic retrieval and clustering of similar words",
"authors": [
{
"first": "Dekang",
"middle": [],
"last": "Lin",
"suffix": ""
}
],
"year": 1998,
"venue": "Proceedings of the 17th International Conference on Computational Linguistics",
"volume": "",
"issue": "",
"pages": "768--774",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Dekang Lin. 1998. Automatic retrieval and clustering of similar words. In Proceedings of the 17th Interna- tional Conference on Computational Linguistics, vol- ume 2, pages 768-774, Montreal, Quebec, Canada.",
"links": null
},
"BIBREF11": {
"ref_id": "b11",
"title": "Foundations of Statistical Language Processing",
"authors": [
{
"first": "D",
"middle": [],
"last": "Christopher",
"suffix": ""
},
{
"first": "Hinrich",
"middle": [],
"last": "Manning",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Sch\u00fctze",
"suffix": ""
}
],
"year": 1999,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Christopher D. Manning and Hinrich Sch\u00fctze. 1999. Foundations of Statistical Language Processing. MIT Press, Cambridge, MA USA.",
"links": null
},
"BIBREF12": {
"ref_id": "b12",
"title": "Thesaurus of English Words and Phrases. Longmans, Green and Company",
"authors": [
{
"first": "Peter",
"middle": [],
"last": "Mark",
"suffix": ""
},
{
"first": "Roget",
"middle": [],
"last": "",
"suffix": ""
}
],
"year": 1911,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Peter Mark Roget. 1911. Thesaurus of English Words and Phrases. Longmans, Green and Company, Lon- don, UK.",
"links": null
},
"BIBREF13": {
"ref_id": "b13",
"title": "Moby thesaurus",
"authors": [
{
"first": "Grady",
"middle": [],
"last": "Ward",
"suffix": ""
}
],
"year": 1996,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Grady Ward. 1996. Moby thesaurus.",
"links": null
},
"BIBREF14": {
"ref_id": "b14",
"title": "Measures and Applications of Lexical Distributional Similarity",
"authors": [
{
"first": "Julie",
"middle": [
"E"
],
"last": "Weeds",
"suffix": ""
}
],
"year": 2003,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Julie E. Weeds. 2003. Measures and Applications of Lex- ical Distributional Similarity. Ph.D. thesis, University of Sussex.",
"links": null
}
},
"ref_entries": {
"TABREF1": {
"type_str": "table",
"text": "",
"html": null,
"num": null,
"content": "<table><tr><td colspan=\"3\">: Results for unigram headwords</td></tr><tr><td>UNI</td><td>UNI+ BI</td><td>UNI+ BI+ VPC</td></tr><tr><td>level</td><td>level</td><td>level</td></tr><tr><td>price</td><td>price</td><td>price</td></tr><tr><td>cost</td><td colspan=\"2\">amount cost</td></tr><tr><td colspan=\"2\">income cost</td><td>amount</td></tr><tr><td>growth</td><td>speed</td><td>average</td></tr></table>"
},
"TABREF2": {
"type_str": "table",
"text": "Sample synonyms for rate",
"html": null,
"num": null,
"content": "<table><tr><td>ATOMIC BOMB</td><td>DINING TABLE</td></tr><tr><td>nuclear bomb</td><td>coffee table</td></tr><tr><td>atom bomb</td><td>dining room</td></tr><tr><td colspan=\"2\">nuclear explosion cocktail table</td></tr><tr><td>atomic explosion</td><td>dining chair</td></tr><tr><td>nuclear weapon</td><td>bedroom furniture</td></tr></table>"
},
"TABREF3": {
"type_str": "table",
"text": "Sample bigram synonyms was extracted from WEB1T and the VPC resource",
"html": null,
"num": null,
"content": "<table/>"
}
}
}
}