| { |
| "paper_id": "U09-1015", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T03:08:09.915600Z" |
| }, |
| "title": "Improved Text Categorisation for Wikipedia Named Entities", |
| "authors": [ |
| { |
| "first": "Sam", |
| "middle": [], |
| "last": "Tardif", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "University of Sydney NSW 2006", |
| "location": { |
| "country": "Australia" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "James", |
| "middle": [ |
| "R" |
| ], |
| "last": "Curran", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "University of Sydney NSW 2006", |
| "location": { |
| "country": "Australia" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Tara", |
| "middle": [], |
| "last": "Murphy", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "University of Sydney NSW 2006", |
| "location": { |
| "country": "Australia" |
| } |
| }, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "The accuracy of named entity recognition systems relies heavily upon the volume and quality of available training data. Improving the process of automatically producing such training data is an important task, as manual acquisition is both time consuming and expensive. We explore the use of a variety of machine learning algorithms for categorising Wikipedia articles, an initial step in producing the named entity training data. We were able to achieve a categorisation accuracy of 95% F-score over six coarse categories, an improvement of up to 5% F-score over previous methods.", |
| "pdf_parse": { |
| "paper_id": "U09-1015", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "The accuracy of named entity recognition systems relies heavily upon the volume and quality of available training data. Improving the process of automatically producing such training data is an important task, as manual acquisition is both time consuming and expensive. We explore the use of a variety of machine learning algorithms for categorising Wikipedia articles, an initial step in producing the named entity training data. We were able to achieve a categorisation accuracy of 95% F-score over six coarse categories, an improvement of up to 5% F-score over previous methods.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Named Entity Recognition (NER) is the task of identifying proper nouns, such as location, organisation and personal names, in text. It emerged as a distinct type of information extraction during the sixth Message Understanding Conference (MUC) evaluation in 1995, and was further defined and explored in the CONLL NER evaluations of 2002 and 2003. A set of four broad categories became the standard scheme for marking named entities (NEs) in text: person (PER), organisation (ORG), location (LOC), and miscellaneous (MISC). This scheme remains the most common, despite the development of more complex hierarchical category schemes (e.g. Brunstein (2002) ; Sekine et al. (2002) ). Domainspecific category schemes have also been developed in many areas, such as astroinformatics (Murphy et al., 2006) , bioinformatics (Kim et al., 2003) and the travel industry (Vijayakrishna and Sobha, 2008) . We also extend the broad scheme with a DAB category for Wikipedia \"disambiguation\" pagespages used to group articles with identical titles.", |
| "cite_spans": [ |
| { |
| "start": 333, |
| "end": 347, |
| "text": "2002 and 2003.", |
| "ref_id": null |
| }, |
| { |
| "start": 637, |
| "end": 653, |
| "text": "Brunstein (2002)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 656, |
| "end": 676, |
| "text": "Sekine et al. (2002)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 777, |
| "end": 798, |
| "text": "(Murphy et al., 2006)", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 816, |
| "end": 834, |
| "text": "(Kim et al., 2003)", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 859, |
| "end": 890, |
| "text": "(Vijayakrishna and Sobha, 2008)", |
| "ref_id": "BIBREF14" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "NER systems that categorise NEs under these schemes require a large amount of highly accurate training data to perform well at the task. Expert annotation is time consuming and expensive, so there is an imperative to generate this data automatically. Wikipedia is emerging as a significant resource due to its immense size and rich structural information, such as its link structure. Nothman et al. (2009) introduced a novel approach to exploiting Wikipedia's internal structure to produce training data for NER systems. Their process involved an initial step of categorising all Wikipedia articles using a simple heuristic-based bootstrapping algorithm. Potential NEs were then identified as the words in an article's text that served as links to other Wikipedia articles. To label a NE they then used the category assigned to the article that it linked to.", |
| "cite_spans": [ |
| { |
| "start": 384, |
| "end": 405, |
| "text": "Nothman et al. (2009)", |
| "ref_id": "BIBREF10" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "We have explored the use of Na\u00efve Bayes (NB) and support vector machines (SVMs) as replacements for the text categorisation approach taken by Nothman. This involved the conversion of heuristics used by Nothman into features as well as the incorporation of a number of new features. We demonstrate the superiority of our approach, providing a comparison of the individual text categorisation step to both Nothman's system and other previous research. Our state-of-the-art text categorisation system for Wikipedia achieved an improvement of up to 5% F -score over previous approaches.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Accurate classifications for Wikipedia articles are useful for a number of natural language processing (NLP) tasks, such as question answering and NER. To produce article classifications for generating NER training data, Nothman et al. (2009) used a heuristicbased text categorisation system. This involved extracting the first head noun after the copula, head nouns from an article's categories, and incoming link information. They reported an F -score of 89% when evaluating on a set of 1,300 hand-labelled articles. Dakka and Cucerzan (2008) explored the use of NB and SVM classifiers for categorising Wikipedia. They expanded each article's bag-of-words representation with disambiguated surface forms, as well as terms extracted from its first paragraph, abstract, and any tables present. They also extracted a small amount of context surrounding links to other Wikipedia articles. Dakka and Cucerzan (2008) expanded their set of 800 hand-labelled articles using a semisupervised approach, extracting training samples from Wikipedia \"List\" pages -pages that group other articles by type. For each \"List\" page containing a link to an article from the hand-labelled set they used the hand-labelled article's category to classify other articles on the list. They neglected to report how many training instances this left them with, but noted that they maintained the original class distribution of the hand-labelled data. They achieved an F -score of 89.7% with an SVM classifier and the category set PER, LOC, ORG, MISC and COM (for common nouns) when classifying their full article set.", |
| "cite_spans": [ |
| { |
| "start": 221, |
| "end": 242, |
| "text": "Nothman et al. (2009)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 519, |
| "end": 544, |
| "text": "Dakka and Cucerzan (2008)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 887, |
| "end": 912, |
| "text": "Dakka and Cucerzan (2008)", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background", |
| "sec_num": "2" |
| }, |
| { |
| "text": "We experimented with a combination of the classification techniques used by Dakka and Cucerzan (2008) and the feature extraction methods used by Nothman et al. (2009) and others (Ponzetto and Strube, 2007; Hu et al., 2008; Biadsy et al., 2008) , focusing on the extraction of features from Wikipedia's rich metadata.", |
| "cite_spans": [ |
| { |
| "start": 76, |
| "end": 101, |
| "text": "Dakka and Cucerzan (2008)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 145, |
| "end": 166, |
| "text": "Nothman et al. (2009)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 178, |
| "end": 205, |
| "text": "(Ponzetto and Strube, 2007;", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 206, |
| "end": 222, |
| "text": "Hu et al., 2008;", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 223, |
| "end": 243, |
| "text": "Biadsy et al., 2008)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Background", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Our annotation and experiments were all run on a March 2009 dump of Wikipedia. (Kiss and Strunk, 2006) and the NLTK (Loper and Bird, 2002) were used to tokenise the corpus.", |
| "cite_spans": [ |
| { |
| "start": 79, |
| "end": 102, |
| "text": "(Kiss and Strunk, 2006)", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 116, |
| "end": 138, |
| "text": "(Loper and Bird, 2002)", |
| "ref_id": "BIBREF8" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Annotation was performed under a slightly modified BBN category hierarchy (Brunstein, 2002) . During annotation we discovered the need for a number of additional categories due to the large number of articles Wikipedia contains relating to popular culture, for example the new categories Organisation \u2192 Band and M isc \u2192 W ork of Art \u2192 T V Series were quite common. We map these categories back to the \"Other\" subcategory of their parent category to allow accurate comparison with the original BBN scheme. Table 1 lists some of our new categories and gives an example for each.", |
| "cite_spans": [ |
| { |
| "start": 74, |
| "end": 91, |
| "text": "(Brunstein, 2002)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 505, |
| "end": 512, |
| "text": "Table 1", |
| "ref_id": "TABREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Annotation scheme", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "We also discovered a number of ambiguities in the original BBN scheme. A number of Wikipedia articles were border cases in the BBN scheme -they related to a number of categories, but did not fit perfectly into any single one. The category M isc \u2192 F ranchise is an example of an additional category to label articles such as \"Star Wars\" and \"Final Fantasy\". We also noticed some unresolvable overlaps in categories, such as Location \u2192 Location \u2192 Island and Location \u2192 GPE \u2192 State for articles such as \"Tasmania\" and \"Hawaii\".", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Annotation scheme", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "A list of Wikipedia articles was selected for annotation based on several criteria. Given the large number of stub articles that exist within Wikipedia and the poor representation of categories that selecting random articles would achieve, our list of articles was primarily based on their popularity as detailed by Ringland et al. (2009) . We took into consideration the number of different language versions of Wikipedia that the article existed in to try and maximise the usefulness of our annotated data for further multi-lingual NLP tasks. We took a list of the most popular articles from August 2008 and checked for an article's existence on that list. We also considered the number of incoming links an article attracted. Based on these three criteria we produced a list of 2,311 articles for annotation.", |
| "cite_spans": [ |
| { |
| "start": 316, |
| "end": 338, |
| "text": "Ringland et al. (2009)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Manual annotation", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Our resulting set of articles was of much higher quality than one that a random article selection process would produce. Random article selection fails to achieve good coverage of some important article categories, such as Location \u2192 GP E \u2192 Country which annotators are likely to never come across using a random selection method. Random selection also yields a high number of stub articles with fewer features for a machine learner to learn from.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Manual annotation", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Our final set of Wikipedia articles was doubleannotated with an inter-annotator agreement of 99.7% using the fine-grained category scheme, and an agreement of 99.87% on the broad NER categories. The remaining classification discrepancies were due to fundamental conflicts in the category hierarchy that could not be resolved. This set of handlabelled articles will be released after publication.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Manual annotation", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Our baseline system used a simple bag-of-words including tokens from the entire article body and the article title. This did not include tokens that appear in templates used in the generation of an article.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Features for text categorisation", |
| "sec_num": "4" |
| }, |
| { |
| "text": "We then experimented with a number of different feature extraction methods, focusing primarily on the document structure for identifying useful features. Tokens in the first paragraph were identified by Dakka and Cucerzan (2008) as useful features for a machine learner, an idea stemming from the fact that most human annotators will recognise an article's category after reading just the first paragraph. We extended this idea by also marking the first sentence and title tokens as separate from other tokens, as we found that often the first sentence was all that was required for a human annotator to classify an article. We ran experiments limiting the feature space to these smaller portions of the document.", |
| "cite_spans": [ |
| { |
| "start": 203, |
| "end": 228, |
| "text": "Dakka and Cucerzan (2008)", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Features for text categorisation", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Wikipedia articles often have a large amount of metadata that helps in identifying an article's category, in particular Wikipedia categories and templates. Wikipedia categories are informal user defined and applied categories, forming a \"folksonomy\" rather than a strict taxonomy suitable for classification tasks, but the terms in the category names are usually strong indicators of an article's class. We extracted the list of categories applied to each article, tokenised the category names and added each token to the bag-of-words representation of the article.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Features for text categorisation", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Using the same reasoning we also extracted a list of each article's templates, tokenised their names, and expanded the article's bag-of-words representation with these tokens. Furthermore, we expanded the templates \"Infobox\", \"Sidebar\" and \"Taxobox\" to extract tokens from their content. These templates often contain a condensed set of important facts relating to the article, and so are powerful additions to the bag-of-words representation of an article. Category, template and infobox features were marked with prefixes to distinguish them from each other and from features extracted from the article body.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Features for text categorisation", |
| "sec_num": "4" |
| }, |
| { |
| "text": "We reduced our raw set of features using a stop list of frequent terms, and removing terms with frequency less than 20 in a set of 1,800,800 articles taken from a separate Wikipedia dump. The assumption is that the majority of low frequency tokens will be typographical errors, or otherwise statistically unreliable data.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Features for text categorisation", |
| "sec_num": "4" |
| }, |
| { |
| "text": "We compared our two classifiers against the heuristic-based system described by Nothman et al. (2009) and the classifiers described by Dakka and Cucerzan (2008) . We also tested a baseline system that used a bag-of-words representation of Wikipedia articles with rich metadata excluded. All SVM experiments were run using LIB-SVM (Chang and Lin, 2001 ) using a linear kernel with parameter C = 2. For NB experiments we used the NLTK. The text categorisation system developed by Nothman et al. (2009) was provided to us by the authors, and we evaluated it using our hand-labelled training data. Direct comparison with this system was difficult, as it has the ability to mark an article as \"unknown\" or \"conflict\" and defer classification. Given that these classifications cannot be considered correct we marked them as classification errors.", |
| "cite_spans": [ |
| { |
| "start": 80, |
| "end": 101, |
| "text": "Nothman et al. (2009)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 135, |
| "end": 160, |
| "text": "Dakka and Cucerzan (2008)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 330, |
| "end": 350, |
| "text": "(Chang and Lin, 2001", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 478, |
| "end": 499, |
| "text": "Nothman et al. (2009)", |
| "ref_id": "BIBREF10" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "5" |
| }, |
| { |
| "text": "There were also a number of complications when comparing our system with the system described by Dakka and Cucerzan (2008) : they used a different, and substantially smaller, hand-labelled data set; they did not specify how they handled disambiguation pages; they provided no results for experiments using only hand-labelled data, instead incorporating training data produced via their semi-automated approach into the final results; and they neglected to report the final size of the training data produced by their semi-automated annotation. However, these two systems provided the closest benchmarks for comparison.", |
| "cite_spans": [ |
| { |
| "start": 97, |
| "end": 122, |
| "text": "Dakka and Cucerzan (2008)", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "5" |
| }, |
| { |
| "text": "We found that across all experiments the NB classifier performed best when using a bag-of-words representation incorporating the first sentence of an article only, along with tokens extracted from categories, templates and infoboxes. Conversely, the SVM classifier performed best using a bag-of-words representation incorporating the entire body of an article, along with category, template and infobox tokens. All experiment results listed were run with these respective configurations. We evaluated our system on two coarse-grained sets of data: the first containing all articles from our hand-labelled set, and the second containing only those articles that described NEs. Table 2 lists results from the top scoring configurations for both the NB and SVM classifiers. The SVM classifier performed significantly better than the NB classifier.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 676, |
| "end": 683, |
| "text": "Table 2", |
| "ref_id": "TABREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Limiting the categorisation scheme to NE-only classes improved the classification accuracy for both classifiers, as the difficult NON class was excluded. With this exclusion the NB classifier became much more competitive with the SVM classifier. Table 3 is a comparison of precision, recall and F -scores between our baseline and final systems, and the systems produced by Nothman et al. (2009) and Dakka and Cucerzan (2008) . The difference between results from Nothman's system, our baseline and our full feature classifier were all found to be statistically significant at the p < 0.05 level. We performed this significance test using a stratified sam- pling approach outlined by Chinchor (1992) .", |
| "cite_spans": [ |
| { |
| "start": 373, |
| "end": 394, |
| "text": "Nothman et al. (2009)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 399, |
| "end": 424, |
| "text": "Dakka and Cucerzan (2008)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 683, |
| "end": 698, |
| "text": "Chinchor (1992)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 246, |
| "end": 253, |
| "text": "Table 3", |
| "ref_id": "TABREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "5" |
| }, |
| { |
| "text": "We exploited Wikipedia's rich document structure and content, such as categories, templates and infoboxes, to classify its articles under a categorisation scheme using NB and SVM machine learners.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "6" |
| }, |
| { |
| "text": "Our system produced state-of-the-art results, achieving an F -score of 95%, an improvement of up to 5% over previous approaches. These high quality classifications are useful for a number of NLP tasks, in particular named entity recognition.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "6" |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "We would like to thank the Language Technology Research Group and the anonymous reviewers for their helpful feedback. This work was partially supported by the Capital Markets Cooperative Research Centre Limited.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgements", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "An unsupervised approach to biography production using wikipedia", |
| "authors": [ |
| { |
| "first": "Fadi", |
| "middle": [], |
| "last": "Biadsy", |
| "suffix": "" |
| }, |
| { |
| "first": "Julia", |
| "middle": [], |
| "last": "Hirschberg", |
| "suffix": "" |
| }, |
| { |
| "first": "Elena", |
| "middle": [], |
| "last": "Filatova", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proceedings of ACL-08: HLT", |
| "volume": "", |
| "issue": "", |
| "pages": "807--815", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Fadi Biadsy, Julia Hirschberg, and Elena Filatova. An unsupervised approach to biography produc- tion using wikipedia. In Proceedings of ACL- 08: HLT, pages 807-815, Columbus, Ohio, June 2008. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Annotation guidelines for answer types. LDC2005T33, Linguistic Data Consortium", |
| "authors": [ |
| { |
| "first": "Ada", |
| "middle": [], |
| "last": "Brunstein", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ada Brunstein. Annotation guidelines for answer types. LDC2005T33, Linguistic Data Consor- tium, Philadelphia, 2002.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "LIBSVM: a library for support vector machines", |
| "authors": [ |
| { |
| "first": "Chih-Chung", |
| "middle": [], |
| "last": "Chang", |
| "suffix": "" |
| }, |
| { |
| "first": "Chih-Jen", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Chih-Chung Chang and Chih-Jen Lin. LIBSVM: a library for support vector machines, 2001.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Statistical significance of muc-6 results", |
| "authors": [ |
| { |
| "first": "N", |
| "middle": [], |
| "last": "Chinchor", |
| "suffix": "" |
| } |
| ], |
| "year": 1992, |
| "venue": "Proceedings, Fourth Message Understanding Conference (MUC-4)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "N Chinchor. Statistical significance of muc-6 results. In Proceedings, Fourth Message Understanding Conference (MUC-4), 1992.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Augmenting wikipedia with named entity tags", |
| "authors": [ |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Dakka", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Cucerzan", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proceedings of IJC-NLP 2008", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "W Dakka and S Cucerzan. Augmenting wikipedia with named entity tags. In Proceedings of IJC- NLP 2008, 2008.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Enhancing text clustering by leveraging wikipedia semantics", |
| "authors": [ |
| { |
| "first": "Jian", |
| "middle": [], |
| "last": "Hu", |
| "suffix": "" |
| }, |
| { |
| "first": "Lujun", |
| "middle": [], |
| "last": "Fang", |
| "suffix": "" |
| }, |
| { |
| "first": "Yang", |
| "middle": [], |
| "last": "Cao", |
| "suffix": "" |
| }, |
| { |
| "first": "Hua-Jun", |
| "middle": [], |
| "last": "Zeng", |
| "suffix": "" |
| }, |
| { |
| "first": "Hua", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Qiang", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "" |
| }, |
| { |
| "first": "Zheng", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "SIGIR '08: Proceedings of the 31st annual international ACM SIGIR conference on Research and development in information retrieval", |
| "volume": "", |
| "issue": "", |
| "pages": "179--186", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jian Hu, Lujun Fang, Yang Cao, Hua-Jun Zeng, Hua Li, Qiang Yang, and Zheng Chen. Enhancing text clustering by leveraging wikipedia semantics. In SIGIR '08: Proceedings of the 31st annual inter- national ACM SIGIR conference on Research and development in information retrieval, pages 179- 186, New York, NY, USA, 2008. ACM. ISBN 978-1-60558-164-4.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "GENIA corpus-a semantically annotated corpus for bio-textmining", |
| "authors": [ |
| { |
| "first": "Jin-Dong", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| }, |
| { |
| "first": "Tomoko", |
| "middle": [], |
| "last": "Ohta", |
| "suffix": "" |
| }, |
| { |
| "first": "Yuka", |
| "middle": [], |
| "last": "Tateisi", |
| "suffix": "" |
| }, |
| { |
| "first": "Jun'ichi", |
| "middle": [], |
| "last": "Tsujii", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Bioinformatics", |
| "volume": "19", |
| "issue": "", |
| "pages": "180--182", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jin-Dong Kim, Tomoko Ohta, Yuka Tateisi, and Jun'ichi Tsujii. GENIA corpus-a semantically annotated corpus for bio-textmining. Bioinfor- matics, 19(suppl. 1):i180-i182, 2003.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Unsupervised multilingual sentence boundary detection", |
| "authors": [ |
| { |
| "first": "Tibor", |
| "middle": [], |
| "last": "Kiss", |
| "suffix": "" |
| }, |
| { |
| "first": "Jan", |
| "middle": [], |
| "last": "Strunk", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Computational Linguistics", |
| "volume": "32", |
| "issue": "", |
| "pages": "485--525", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tibor Kiss and Jan Strunk. Unsupervised multilin- gual sentence boundary detection. Computational Linguistics, 32:485-525, 2006.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "NLTK: The natural language toolkit", |
| "authors": [ |
| { |
| "first": "Edward", |
| "middle": [], |
| "last": "Loper", |
| "suffix": "" |
| }, |
| { |
| "first": "Steven", |
| "middle": [], |
| "last": "Bird", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "Proceedings of the Workshop on Effective Tools and Methodologies for Teaching Natural Language Processing and Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "63--70", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Edward Loper and Steven Bird. NLTK: The natural language toolkit. In Proceedings of the Workshop on Effective Tools and Methodologies for Teach- ing Natural Language Processing and Compu- tational Linguistics, pages 63-70, Philadelphia, July 2002.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Curran. Named entity recognition for astronomy literature", |
| "authors": [ |
| { |
| "first": "Tara", |
| "middle": [], |
| "last": "Murphy", |
| "suffix": "" |
| }, |
| { |
| "first": "Tara", |
| "middle": [], |
| "last": "Mcintosh", |
| "suffix": "" |
| }, |
| { |
| "first": "James", |
| "middle": [ |
| "R" |
| ], |
| "last": "", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of the Australian Language Technology Workshop", |
| "volume": "", |
| "issue": "", |
| "pages": "59--66", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tara Murphy, Tara McIntosh, and James R. Cur- ran. Named entity recognition for astronomy lit- erature. In Proceedings of the Australian Lan- guage Technology Workshop, pages 59-66, Syd- ney, Australia, 2006.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Analysing Wikipedia and gold-standard corpora for NER training", |
| "authors": [ |
| { |
| "first": "Joel", |
| "middle": [], |
| "last": "Nothman", |
| "suffix": "" |
| }, |
| { |
| "first": "Tara", |
| "middle": [], |
| "last": "Murphy", |
| "suffix": "" |
| }, |
| { |
| "first": "James", |
| "middle": [ |
| "R" |
| ], |
| "last": "Curran", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Proceedings of the 12th Conference of the European Chapter of the ACL (EACL 2009)", |
| "volume": "", |
| "issue": "", |
| "pages": "612--620", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Joel Nothman, Tara Murphy, and James R. Curran. Analysing Wikipedia and gold-standard corpora for NER training. In Proceedings of the 12th Conference of the European Chapter of the ACL (EACL 2009), pages 612-620, Athens, Greece, March 2009. Association for Computational Lin- guistics.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Deriving a large scale taxonomy from wikipedia", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "S P Ponzetto", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Strube", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proceedings of the 22nd National Conference on Artificial Intelligence", |
| "volume": "", |
| "issue": "", |
| "pages": "1440--1445", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "S P Ponzetto and M Strube. Deriving a large scale taxonomy from wikipedia. In Proceedings of the 22nd National Conference on Artificial Intelli- gence, pages 1440-1445, 2007.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Classifying articles in english and german wikipedias", |
| "authors": [ |
| { |
| "first": "Nicky", |
| "middle": [], |
| "last": "Ringland", |
| "suffix": "" |
| }, |
| { |
| "first": "James", |
| "middle": [ |
| "R" |
| ], |
| "last": "Curran", |
| "suffix": "" |
| }, |
| { |
| "first": "Tara", |
| "middle": [], |
| "last": "Murphy", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nicky Ringland, James R. Curran, and Tara Mur- phy. Classifying articles in english and german wikipedias. In Submitted to ALTA, 2009.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Extended named entity hierarchy", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Sekine", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Sudo", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Nobata", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "Proceedings of the LREC-2002", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "S Sekine, K Sudo, and C Nobata. Extended named entity hierarchy. In Proceedings of the LREC- 2002, 2002.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Domain focused named entity recognizer for tamil using conditional random fields", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Vijayakrishna", |
| "suffix": "" |
| }, |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Sobha", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proceedings of the IJCNLP-08 Workshop on Named Entity Recognition for South and South East Asian Languages", |
| "volume": "", |
| "issue": "", |
| "pages": "59--66", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "R. Vijayakrishna and L. Sobha. Domain focused named entity recognizer for tamil using condi- tional random fields. In Proceedings of the IJCNLP-08 Workshop on Named Entity Recogni- tion for South and South East Asian Languages, pages 59-66, Hyderabad, India, January 2008.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "TABREF0": { |
| "type_str": "table", |
| "text": "Extensions to the BBN categories with examples was used to parse the Mediawiki markup and perform tasks such as expanding Wikipedia templates and extracting article categories and links. Punkt", |
| "num": null, |
| "content": "<table><tr><td>The mwlib 1 library</td></tr><tr><td>1 http://code.pediapress.com</td></tr></table>", |
| "html": null |
| }, |
| "TABREF2": { |
| "type_str": "table", |
| "text": "NB and SVM results on coarse-grained problems.", |
| "num": null, |
| "content": "<table><tr><td>Classifier</td><td>F</td></tr><tr><td>Nothman</td><td>91</td></tr><tr><td>Dakka</td><td>90</td></tr><tr><td>BASELINE</td><td>94</td></tr><tr><td>BEST</td><td>95</td></tr></table>", |
| "html": null |
| }, |
| "TABREF3": { |
| "type_str": "table", |
| "text": "Comparison with previous systems.", |
| "num": null, |
| "content": "<table/>", |
| "html": null |
| } |
| } |
| } |
| } |