| import glob |
| import logging |
| from dataclasses import dataclass |
| from os import listdir, path |
| from typing import Dict, List, Optional |
|
|
| import datasets |
| from datasets import BuilderConfig, DatasetInfo, Features, Sequence, SplitGenerator, Value |
|
|
| logger = logging.getLogger(__name__) |
|
|
|
|
| @dataclass |
| class BratConfig(BuilderConfig): |
| """BuilderConfig for BRAT.""" |
|
|
| url: str = None |
| description: Optional[str] = None |
| citation: Optional[str] = None |
| homepage: Optional[str] = None |
|
|
| subdirectory_mapping: Optional[Dict[str, str]] = None |
| file_name_blacklist: Optional[List[str]] = None |
| ann_file_extension: str = "ann" |
| txt_file_extension: str = "txt" |
|
|
|
|
| class Brat(datasets.GeneratorBasedBuilder): |
| BUILDER_CONFIG_CLASS = BratConfig |
|
|
| def _info(self): |
| return DatasetInfo( |
| description=self.config.description, |
| citation=self.config.citation, |
| homepage=self.config.homepage, |
| features=Features( |
| { |
| "context": Value("string"), |
| "file_name": Value("string"), |
| "spans": Sequence( |
| { |
| "id": Value("string"), |
| "type": Value("string"), |
| "locations": Sequence( |
| { |
| "start": Value("int32"), |
| "end": Value("int32"), |
| } |
| ), |
| "text": Value("string"), |
| } |
| ), |
| "relations": Sequence( |
| { |
| "id": Value("string"), |
| "type": Value("string"), |
| "arguments": Sequence( |
| {"type": Value("string"), "target": Value("string")} |
| ), |
| } |
| ), |
| "equivalence_relations": Sequence( |
| { |
| "type": Value("string"), |
| "targets": Sequence(Value("string")), |
| } |
| ), |
| "events": Sequence( |
| { |
| "id": Value("string"), |
| "type": Value("string"), |
| "trigger": Value("string"), |
| "arguments": Sequence( |
| {"type": Value("string"), "target": Value("string")} |
| ), |
| } |
| ), |
| "attributions": Sequence( |
| { |
| "id": Value("string"), |
| "type": Value("string"), |
| "target": Value("string"), |
| "value": Value("string"), |
| } |
| ), |
| "normalizations": Sequence( |
| { |
| "id": Value("string"), |
| "type": Value("string"), |
| "target": Value("string"), |
| "resource_id": Value("string"), |
| "entity_id": Value("string"), |
| } |
| ), |
| "notes": Sequence( |
| { |
| "id": Value("string"), |
| "type": Value("string"), |
| "target": Value("string"), |
| "note": Value("string"), |
| } |
| ), |
| } |
| ), |
| ) |
|
|
| @staticmethod |
| def _get_location(location_string): |
| parts = location_string.split(" ") |
| assert ( |
| len(parts) == 2 |
| ), f"Wrong number of entries in location string. Expected 2, but found: {parts}" |
| return {"start": int(parts[0]), "end": int(parts[1])} |
|
|
| @staticmethod |
| def _get_span_annotation(annotation_line): |
| """ |
| example input: |
| T1 Organization 0 4 Sony |
| """ |
|
|
| _id, remaining, text = annotation_line.split("\t", maxsplit=2) |
| _type, locations = remaining.split(" ", maxsplit=1) |
| return { |
| "id": _id, |
| "text": text, |
| "type": _type, |
| "locations": [Brat._get_location(loc) for loc in locations.split(";")], |
| } |
|
|
| @staticmethod |
| def _get_event_annotation(annotation_line): |
| """ |
| example input: |
| E1 MERGE-ORG:T2 Org1:T1 Org2:T3 |
| """ |
| _id, remaining = annotation_line.strip().split("\t") |
| args = [dict(zip(["type", "target"], a.split(":"))) for a in remaining.split(" ")] |
| return { |
| "id": _id, |
| "type": args[0]["type"], |
| "trigger": args[0]["target"], |
| "arguments": args[1:], |
| } |
|
|
| @staticmethod |
| def _get_relation_annotation(annotation_line): |
| """ |
| example input: |
| R1 Origin Arg1:T3 Arg2:T4 |
| """ |
|
|
| _id, remaining = annotation_line.strip().split("\t") |
| _type, remaining = remaining.split(" ", maxsplit=1) |
| args = [dict(zip(["type", "target"], a.split(":"))) for a in remaining.split(" ")] |
| return {"id": _id, "type": _type, "arguments": args} |
|
|
| @staticmethod |
| def _get_equivalence_relation_annotation(annotation_line): |
| """ |
| example input: |
| * Equiv T1 T2 T3 |
| """ |
| _, remaining = annotation_line.strip().split("\t") |
| parts = remaining.split(" ") |
| return {"type": parts[0], "targets": parts[1:]} |
|
|
| @staticmethod |
| def _get_attribute_annotation(annotation_line): |
| """ |
| example input (binary: implicit value is True, if present, False otherwise): |
| A1 Negation E1 |
| example input (multi-value: explicit value) |
| A2 Confidence E2 L1 |
| """ |
|
|
| _id, remaining = annotation_line.strip().split("\t") |
| parts = remaining.split(" ") |
| |
| if len(parts) == 2: |
| parts.append("true") |
| return { |
| "id": _id, |
| "type": parts[0], |
| "target": parts[1], |
| "value": parts[2], |
| } |
|
|
| @staticmethod |
| def _get_normalization_annotation(annotation_line): |
| """ |
| example input: |
| N1 Reference T1 Wikipedia:534366 Barack Obama |
| """ |
| _id, remaining, text = annotation_line.split("\t", maxsplit=2) |
| _type, target, ref = remaining.split(" ") |
| res_id, ent_id = ref.split(":") |
| return { |
| "id": _id, |
| "type": _type, |
| "target": target, |
| "resource_id": res_id, |
| "entity_id": ent_id, |
| } |
|
|
| @staticmethod |
| def _get_note_annotation(annotation_line): |
| """ |
| example input: |
| #1 AnnotatorNotes T1 this annotation is suspect |
| """ |
| _id, remaining, note = annotation_line.split("\t", maxsplit=2) |
| _type, target = remaining.split(" ") |
| return { |
| "id": _id, |
| "type": _type, |
| "target": target, |
| "note": note, |
| } |
|
|
| @staticmethod |
| def _read_annotation_file(filename): |
| """ |
| reads a BRAT v1.3 annotations file (see https://brat.nlplab.org/standoff.html) |
| """ |
|
|
| res = { |
| "spans": [], |
| "events": [], |
| "relations": [], |
| "equivalence_relations": [], |
| "attributions": [], |
| "normalizations": [], |
| "notes": [], |
| } |
|
|
| with open(filename) as file: |
| for i, line in enumerate(file): |
| if len(line.strip()) == 0: |
| continue |
| ann_type = line[0] |
|
|
| |
| if line.endswith("\n"): |
| line = line[:-1] |
|
|
| if ann_type == "T": |
| res["spans"].append(Brat._get_span_annotation(line)) |
| elif ann_type == "E": |
| res["events"].append(Brat._get_event_annotation(line)) |
| elif ann_type == "R": |
| res["relations"].append(Brat._get_relation_annotation(line)) |
| elif ann_type == "*": |
| res["equivalence_relations"].append( |
| Brat._get_equivalence_relation_annotation(line) |
| ) |
| elif ann_type in ["A", "M"]: |
| res["attributions"].append(Brat._get_attribute_annotation(line)) |
| elif ann_type == "N": |
| res["normalizations"].append(Brat._get_normalization_annotation(line)) |
| elif ann_type == "#": |
| res["notes"].append(Brat._get_note_annotation(line)) |
| else: |
| raise ValueError( |
| f'unknown BRAT annotation id type: "{line}" (from file {filename} @line {i}). ' |
| f"Annotation ids have to start with T (spans), E (events), R (relations), " |
| f"A (attributions), or N (normalizations). See " |
| f"https://brat.nlplab.org/standoff.html for the BRAT annotation file " |
| f"specification." |
| ) |
| return res |
|
|
| def _generate_examples(self, files=None, directory=None): |
| """Read context (.txt) and annotation (.ann) files.""" |
| if files is None: |
| assert ( |
| directory is not None |
| ), "If files is None, directory has to be provided, but it is also None." |
| _files = glob.glob(f"{directory}/*.{self.config.ann_file_extension}") |
| files = sorted(path.splitext(fn)[0] for fn in _files) |
|
|
| for filename in files: |
| basename = path.basename(filename) |
| if ( |
| self.config.file_name_blacklist is not None |
| and basename in self.config.file_name_blacklist |
| ): |
| logger.info(f"skip annotation file: {basename} (blacklisted)") |
| continue |
|
|
| ann_fn = f"{filename}.{self.config.ann_file_extension}" |
| brat_annotations = Brat._read_annotation_file(ann_fn) |
|
|
| txt_fn = f"{filename}.{self.config.txt_file_extension}" |
| txt_content = open(txt_fn).read() |
| brat_annotations["context"] = txt_content |
| brat_annotations["file_name"] = basename |
|
|
| yield basename, brat_annotations |
|
|
| def _split_generators(self, dl_manager): |
| """Returns SplitGenerators.""" |
|
|
| subdirectory_mapping = self.config.subdirectory_mapping |
|
|
| |
| |
| assert self.config.url is not None, "data url not specified" |
|
|
| |
| if path.exists(self.config.url) and path.isdir(self.config.url): |
| data_dir = self.config.url |
| |
| else: |
| data_dir = dl_manager.download_and_extract(self.config.url) |
| logging.info(f"load from data dir: {data_dir}") |
|
|
| |
| if subdirectory_mapping is None: |
| |
| subdirs = [f for f in listdir(data_dir) if path.isdir(path.join(data_dir, f))] |
| if len(subdirs) > 0: |
| subdirectory_mapping = {subdir: subdir for subdir in subdirs} |
| else: |
| |
| subdirectory_mapping = {"": "train"} |
|
|
| return [ |
| SplitGenerator( |
| name=split, |
| |
| gen_kwargs={ |
| "directory": path.join(data_dir, subdir), |
| }, |
| ) |
| for subdir, split in subdirectory_mapping.items() |
| ] |