--- license: cc-by-4.0 language: - de size_categories: - n<1K --- # CO-Fun: Tokenized Sentences This datasets hosts a sentence-tokenized version of the [CO-Fun: A German Dataset on Company Outsourcing in Fund Prospectuses for Named Entity Recognition and Relation Extraction](https://arxiv.org/abs/2403.15322) dataset. ## Creation The following script can be used to reproduce the creation of the dataset: ```python import flair import json from flair.datasets.sequence_labeling import ColumnCorpus from flair.file_utils import cached_path from pathlib import Path from typing import Optional, Union class NER_CO_FUNER(ColumnCorpus): def __init__( self, base_path: Optional[Union[str, Path]] = None, in_memory: bool = True, **corpusargs, ) -> None: base_path = flair.cache_root / "datasets" if not base_path else Path(base_path) dataset_name = self.__class__.__name__.lower() data_folder = base_path / dataset_name data_path = flair.cache_root / "datasets" / dataset_name columns = {0: "text", 2: "ner"} hf_download_path = "https://huggingface.co/datasets/stefan-it/co-funer/resolve/main" for split in ["train", "dev", "test"]: cached_path(f"{hf_download_path}/{split}.tsv", data_path) super().__init__( data_folder, columns, in_memory=in_memory, comment_symbol=None, **corpusargs, ) corpus = NER_CO_FUNER() with open("./train.jsonl", "wt") as f_out: for sentence in corpus.train: current_example = { "text": sentence.to_tokenized_string() } f_out.write(json.dumps(current_example) + "\n") ``` The extracted dataset has 758 sentences.