| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
|
|
| import itertools |
| import os |
| import uuid |
| import xml.etree.ElementTree as ET |
| from typing import List |
|
|
| import datasets |
| from numpy import int32 |
|
|
| from .bigbiohub import kb_features |
| from .bigbiohub import BigBioConfig |
| from .bigbiohub import Tasks |
|
|
| _LANGUAGES = ['English'] |
| _PUBMED = True |
| _LOCAL = False |
| _CITATION = """\ |
| @ARTICLE{Furlong2008, |
| author = {Laura I Furlong and Holger Dach and Martin Hofmann-Apitius and Ferran Sanz}, |
| title = {OSIRISv1.2: a named entity recognition system for sequence variants |
| of genes in biomedical literature.}, |
| journal = {BMC Bioinformatics}, |
| year = {2008}, |
| volume = {9}, |
| pages = {84}, |
| doi = {10.1186/1471-2105-9-84}, |
| pii = {1471-2105-9-84}, |
| pmid = {18251998}, |
| timestamp = {2013.01.15}, |
| url = {http://dx.doi.org/10.1186/1471-2105-9-84} |
| } |
| """ |
|
|
| _DATASETNAME = "osiris" |
| _DISPLAYNAME = "OSIRIS" |
|
|
| _DESCRIPTION = """\ |
| The OSIRIS corpus is a set of MEDLINE abstracts manually annotated |
| with human variation mentions. The corpus is distributed under the terms |
| of the Creative Commons Attribution License |
| Creative Commons Attribution 3.0 Unported License, |
| which permits unrestricted use, distribution, and reproduction in any medium, |
| provided the original work is properly cited (Furlong et al, BMC Bioinformatics 2008, 9:84). |
| """ |
|
|
| _HOMEPAGE = "https://sites.google.com/site/laurafurlongweb/databases-and-tools/corpora/" |
|
|
|
|
| _LICENSE = 'Creative Commons Attribution 3.0 Unported' |
|
|
| _URLS = { |
| _DATASETNAME: [ |
| "https://github.com/rockt/SETH/blob/master/resources/OSIRIS/corpus.xml?raw=true " |
| ] |
| } |
|
|
| _SUPPORTED_TASKS = [Tasks.NAMED_ENTITY_RECOGNITION, Tasks.NAMED_ENTITY_DISAMBIGUATION] |
|
|
|
|
| _SOURCE_VERSION = "1.2.0" |
|
|
| _BIGBIO_VERSION = "1.0.0" |
|
|
|
|
| class Osiris(datasets.GeneratorBasedBuilder): |
| """ |
| The OSIRIS corpus is a set of MEDLINE abstracts manually annotated |
| with human variation mentions. |
| """ |
|
|
| SOURCE_VERSION = datasets.Version(_SOURCE_VERSION) |
| BIGBIO_VERSION = datasets.Version(_BIGBIO_VERSION) |
|
|
| |
| |
| |
|
|
| |
| |
| |
| |
|
|
| BUILDER_CONFIGS = [ |
| BigBioConfig( |
| name="osiris_source", |
| version=SOURCE_VERSION, |
| description="osiris source schema", |
| schema="source", |
| subset_id="osiris", |
| ), |
| BigBioConfig( |
| name="osiris_bigbio_kb", |
| version=BIGBIO_VERSION, |
| description="osiris BigBio schema", |
| schema="bigbio_kb", |
| subset_id="osiris", |
| ), |
| ] |
|
|
| DEFAULT_CONFIG_NAME = "osiris_source" |
|
|
| def _info(self) -> datasets.DatasetInfo: |
|
|
| if self.config.schema == "source": |
|
|
| features = datasets.Features( |
| { |
| "Pmid": datasets.Value("string"), |
| "Title": datasets.Value("string"), |
| "Abstract": datasets.Value("string"), |
| "genes": [ |
| { |
| "g_id": datasets.Value("string"), |
| "g_lex": datasets.Value("string"), |
| "offsets": [[datasets.Value("int32")]], |
| } |
| ], |
| "variants": [ |
| { |
| "v_id": datasets.Value("string"), |
| "v_lex": datasets.Value("string"), |
| "v_norm": datasets.Value("string"), |
| "offsets": [[datasets.Value("int32")]], |
| } |
| ], |
| } |
| ) |
|
|
| elif self.config.schema == "bigbio_kb": |
| features = kb_features |
|
|
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features=features, |
| homepage=_HOMEPAGE, |
| license=str(_LICENSE), |
| citation=_CITATION, |
| ) |
|
|
| def _split_generators(self, dl_manager) -> List[datasets.SplitGenerator]: |
|
|
| urls = _URLS[_DATASETNAME] |
| data_dir = dl_manager.download(urls) |
|
|
| return [ |
| datasets.SplitGenerator( |
| name=datasets.Split.TRAIN, |
| |
| gen_kwargs={ |
| "filepath": os.path.join(data_dir[0]), |
| "split": "data", |
| }, |
| ) |
| ] |
|
|
| def _get_offsets(self, parent: ET.Element, child: ET.Element) -> List[int32]: |
| """ |
| Retrieves character offsets for child from parent. |
| """ |
| parent_text = " ".join( |
| [ |
| " ".join([t for t in c.itertext()]) |
| for c in list(parent) |
| if c.tag != "Pmid" |
| ] |
| ) |
| child_text = " ".join([t for t in child.itertext()]) |
| start = parent_text.index(child_text) |
| end = start + len(child_text) |
| return [start, end] |
|
|
| def _get_dict(self, elem: ET.Element) -> dict: |
| """ |
| Retrieves dict from XML element. |
| """ |
| elem_d = dict() |
| for child in elem: |
| elem_d[child.tag] = {} |
| elem_d[child.tag]["text"] = " ".join([t for t in child.itertext()]) |
|
|
| if child.tag != "Pmid": |
| elem_d[child.tag]["offsets"] = self._get_offsets(elem, child) |
|
|
| for c in child: |
| elem_d[c.tag] = [] |
|
|
| for c in child: |
| c_dict = c.attrib |
| c_dict["offsets"] = self._get_offsets(elem, c) |
| elem_d[c.tag].append(c.attrib) |
|
|
| return elem_d |
|
|
| def _handle_missing_variants(self, row: dict) -> dict: |
| """ |
| If variant is not present in the row this function adds one variant |
| with no data (to make looping though items possible) and returns the new row. |
| These mocked variants will be romoved after parsing. |
| Otherwise returns unchanged row. |
| """ |
|
|
| if row.get("variant", 0) == 0: |
| row["variant"] = [ |
| {"v_id": "", "v_lex": "", "v_norm": "", "offsets": [0, 0]} |
| ] |
| return row |
|
|
| def _get_entities(self, row: dict) -> List[dict]: |
| """ |
| Retrieves two lists of dicts for genes and variants. |
| After that, chains both together. |
| """ |
| genes = [ |
| { |
| "id": str(uuid.uuid4()), |
| "offsets": [gene["offsets"]], |
| "text": [gene["g_lex"]], |
| "type": "gene", |
| "normalized": [{"db_name": "NCBI Gene", "db_id": gene["g_id"]}], |
| } |
| for gene in row["gene"] |
| ] |
|
|
| variants = [ |
| { |
| "id": str(uuid.uuid4()), |
| "offsets": [variant["offsets"]], |
| "text": [variant["v_lex"]], |
| "type": "variant", |
| "normalized": [ |
| { |
| "db_name": "HGVS-like" if variant["v_id"] == "No" else "dbSNP", |
| "db_id": variant["v_norm"] |
| if variant["v_id"] == "No" |
| else variant["v_id"], |
| } |
| ], |
| } |
| for variant in row["variant"] |
| if variant["v_id"] != "" |
| ] |
| return list(itertools.chain(genes, variants)) |
|
|
| def _generate_examples(self, filepath, split): |
|
|
| root = ET.parse(filepath).getroot() |
| uid = 0 |
| if self.config.schema == "source": |
| for elem in list(root): |
| row = self._get_dict(elem) |
|
|
| |
| row = self._handle_missing_variants(row) |
| uid += 1 |
| yield uid, { |
| "Pmid": row["Pmid"]["text"], |
| "Title": { |
| "offsets": [row["Title"]["offsets"]], |
| "text": row["Title"]["text"], |
| }, |
| "Abstract": { |
| "offsets": [row["Abstract"]["offsets"]], |
| "text": row["Abstract"]["text"], |
| }, |
| "genes": [ |
| { |
| "g_id": gene["g_id"], |
| "g_lex": gene["g_lex"], |
| "offsets": [gene["offsets"]], |
| } |
| for gene in row["gene"] |
| ], |
| "variants": [ |
| { |
| "v_id": variant["v_id"], |
| "v_lex": variant["v_lex"], |
| "v_norm": variant["v_norm"], |
| "offsets": [variant["offsets"]], |
| } |
| for variant in row["variant"] |
| ], |
| } |
|
|
| elif self.config.schema == "bigbio_kb": |
|
|
| for elem in list(root): |
| row = self._get_dict(elem) |
|
|
| |
| row = self._handle_missing_variants(row) |
| uid += 1 |
| yield uid, { |
| "id": str(uid), |
| "document_id": row["Pmid"]["text"], |
| "passages": [ |
| { |
| "id": str(uuid.uuid4()), |
| "type": "title", |
| "text": [row["Title"]["text"]], |
| "offsets": [row["Title"]["offsets"]], |
| }, |
| { |
| "id": str(uuid.uuid4()), |
| "type": "abstract", |
| "text": [row["Abstract"]["text"]], |
| "offsets": [row["Abstract"]["offsets"]], |
| }, |
| ], |
| "entities": self._get_entities(row), |
| "relations": [], |
| "events": [], |
| "coreferences": [], |
| } |
|
|