| from datasets import BuilderConfig, Version, GeneratorBasedBuilder, DatasetInfo, Features, Value, \ |
| Sequence, ClassLabel, DownloadManager, SplitGenerator, Split |
| import datasets |
| import os |
| import textwrap |
| import csv |
| from ast import literal_eval |
|
|
|
|
| _DESCRIPTION = """ |
| The recognition and classification of proper nouns and names in plain text is of key importance in Natural Language |
| Processing (NLP) as it has a beneficial effect on the performance of various types of applications, including |
| Information Extraction, Machine Translation, Syntactic Parsing/Chunking, etc.""" |
| _CITATION = """""" |
| _FEATURES = Features( |
| { |
| "id": Value("int32"), |
| "tokens": Sequence(Value("string")), |
| "ner": Sequence( |
| ClassLabel( |
| names=[ |
| "O", |
| "B-PER", |
| "I-PER", |
| "B-ORG", |
| "I-ORG", |
| "B-LOC", |
| "I-LOC", |
| "B-MISC", |
| "I-MISC", |
| ] |
| ) |
| ), |
| "document_id": Value("int32"), |
| "sentence_id": Value("int32") |
| } |
| ) |
|
|
|
|
| class SzegedNERConfig(BuilderConfig): |
| """BuilderConfig for SzegedNER.""" |
|
|
| def __init__( |
| self, |
| features, |
| label_column, |
| data_dir, |
| citation, |
| url, |
| process_label=lambda x: x, |
| **kwargs, |
| ): |
| super(SzegedNERConfig, self).__init__(version=Version("1.0.0", ""), **kwargs) |
| self.features = features |
| self.label_column = label_column |
| self.data_dir = data_dir |
| self.citation = citation |
| self.url = url |
| self.process_label = process_label |
|
|
|
|
| class SzegedNER(GeneratorBasedBuilder): |
| """SzegedNER datasets.""" |
|
|
| BUILDER_CONFIGS = [ |
| SzegedNERConfig( |
| name="business", |
| description=textwrap.dedent( |
| """\ |
| The Named Entity Corpus for Hungarian is a subcorpus of the Szeged Treebank, which contains full syntactic |
| annotations done manually by linguist experts. A significant part of these texts has been annotated with |
| Named Entity class labels in line with the annotation standards used on the CoNLL-2003 shared task.""" |
| ), |
| features=_FEATURES, |
| label_column="ner_tags", |
| data_dir="data/business/", |
| citation=textwrap.dedent(_CITATION), |
| url="https://rgai.inf.u-szeged.hu/node/130" |
| ), |
| SzegedNERConfig( |
| name="criminal", |
| description=textwrap.dedent( |
| """\ |
| The Hungarian National Corpus and its Heti Világgazdaság (HVG) subcorpus provided the basis for corpus text |
| selection: articles related to the topic of financially liable offences were selected and annotated for the |
| categories person, organization, location and miscellaneous. There are two annotated versions of the corpus. |
| When preparing the tag-for-meaning annotation, our linguists took into consideration the context in which |
| the Named Entity under investigation occurred, thus, it was not the primary sense of the Named Entity that |
| determined the tag (e.g. Manchester=LOC) but its contextual reference (e.g. Manchester won the Premier |
| League=ORG). As for tag-for-tag annotation, these cases were not differentiated: tags were always given on |
| the basis of the primary sense.""" |
| ), |
| features=_FEATURES, |
| label_column="ner_tags", |
| data_dir="data/criminal/", |
| citation=textwrap.dedent(_CITATION), |
| url="https://rgai.inf.u-szeged.hu/node/130" |
| ) |
| ] |
|
|
| def _info(self): |
| return DatasetInfo( |
| description=self.config.description, |
| features=self.config.features, |
| homepage=self.config.url, |
| citation=self.config.citation, |
| ) |
|
|
| def _split_generators(self, dl_manager: DownloadManager): |
| url = f"{self.base_path}{self.config.data_dir}" |
|
|
| path = dl_manager.download({key: f"{url}{key}.csv" for key in ["train", "validation", "test"]}) |
| return [ |
| SplitGenerator( |
| name=Split.TRAIN, |
| gen_kwargs={"split_key": "train", "data_file": path['train']}, |
| ), |
| SplitGenerator( |
| name=Split.VALIDATION, |
| gen_kwargs={"split_key": "validation", "data_file": path['validation']}, |
| ), |
| SplitGenerator( |
| name=Split.TEST, |
| gen_kwargs={"split_key": "test", "data_file": path['test']}, |
| ) |
| ] |
|
|
| def _generate_examples(self, data_file, split_key, **kwargs): |
| with open(data_file, encoding="utf8") as f: |
| reader = csv.DictReader(f, delimiter=",", quoting=csv.QUOTE_MINIMAL) |
| for n, row in enumerate(reader): |
| labels = literal_eval(row['ner']) |
| tokens = literal_eval(row['tokens']) |
| if len(labels) != len(tokens): |
| raise ValueError("Number of tokens and labels does not match") |
| yield n, { |
| "id": int(row['id']), |
| "tokens": tokens, |
| "ner": labels, |
| "document_id": int(row['document_id']), |
| "sentence_id": int(row['sentence_id']) |
| } |
|
|