Datasets:

Languages:
English
ArXiv:
License:
AnonymousARR42 gabrielaltay commited on
Commit
15b0bb5
·
verified ·
0 Parent(s):

Duplicate from bigbio/medmentions

Browse files

Co-authored-by: Gabriel Altay <gabrielaltay@users.noreply.huggingface.co>

Files changed (4) hide show
  1. .gitattributes +54 -0
  2. README.md +63 -0
  3. bigbiohub.py +592 -0
  4. medmentions.py +307 -0
.gitattributes ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.lz4 filter=lfs diff=lfs merge=lfs -text
12
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
13
+ *.model filter=lfs diff=lfs merge=lfs -text
14
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
15
+ *.npy filter=lfs diff=lfs merge=lfs -text
16
+ *.npz filter=lfs diff=lfs merge=lfs -text
17
+ *.onnx filter=lfs diff=lfs merge=lfs -text
18
+ *.ot filter=lfs diff=lfs merge=lfs -text
19
+ *.parquet filter=lfs diff=lfs merge=lfs -text
20
+ *.pb filter=lfs diff=lfs merge=lfs -text
21
+ *.pickle filter=lfs diff=lfs merge=lfs -text
22
+ *.pkl filter=lfs diff=lfs merge=lfs -text
23
+ *.pt filter=lfs diff=lfs merge=lfs -text
24
+ *.pth filter=lfs diff=lfs merge=lfs -text
25
+ *.rar filter=lfs diff=lfs merge=lfs -text
26
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
27
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ # Audio files - uncompressed
37
+ *.pcm filter=lfs diff=lfs merge=lfs -text
38
+ *.sam filter=lfs diff=lfs merge=lfs -text
39
+ *.raw filter=lfs diff=lfs merge=lfs -text
40
+ # Audio files - compressed
41
+ *.aac filter=lfs diff=lfs merge=lfs -text
42
+ *.flac filter=lfs diff=lfs merge=lfs -text
43
+ *.mp3 filter=lfs diff=lfs merge=lfs -text
44
+ *.ogg filter=lfs diff=lfs merge=lfs -text
45
+ *.wav filter=lfs diff=lfs merge=lfs -text
46
+ # Image files - uncompressed
47
+ *.bmp filter=lfs diff=lfs merge=lfs -text
48
+ *.gif filter=lfs diff=lfs merge=lfs -text
49
+ *.png filter=lfs diff=lfs merge=lfs -text
50
+ *.tiff filter=lfs diff=lfs merge=lfs -text
51
+ # Image files - compressed
52
+ *.jpg filter=lfs diff=lfs merge=lfs -text
53
+ *.jpeg filter=lfs diff=lfs merge=lfs -text
54
+ *.webp filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ language:
4
+ - en
5
+ bigbio_language:
6
+ - English
7
+ license: cc0-1.0
8
+ multilinguality: monolingual
9
+ bigbio_license_shortname: CC0_1p0
10
+ pretty_name: MedMentions
11
+ homepage: https://github.com/chanzuckerberg/MedMentions
12
+ bigbio_pubmed: True
13
+ bigbio_public: True
14
+ bigbio_tasks:
15
+ - NAMED_ENTITY_DISAMBIGUATION
16
+ - NAMED_ENTITY_RECOGNITION
17
+ ---
18
+
19
+
20
+ # Dataset Card for MedMentions
21
+
22
+ ## Dataset Description
23
+
24
+ - **Homepage:** https://github.com/chanzuckerberg/MedMentions
25
+ - **Pubmed:** True
26
+ - **Public:** True
27
+ - **Tasks:** NED,NER
28
+
29
+
30
+ MedMentions is a new manually annotated resource for the recognition of biomedical concepts.
31
+ What distinguishes MedMentions from other annotated biomedical corpora is its size (over 4,000
32
+ abstracts and over 350,000 linked mentions), as well as the size of the concept ontology (over
33
+ 3 million concepts from UMLS 2017) and its broad coverage of biomedical disciplines.
34
+
35
+ Corpus: The MedMentions corpus consists of 4,392 papers (Titles and Abstracts) randomly selected
36
+ from among papers released on PubMed in 2016, that were in the biomedical field, published in
37
+ the English language, and had both a Title and an Abstract.
38
+
39
+ Annotators: We recruited a team of professional annotators with rich experience in biomedical
40
+ content curation to exhaustively annotate all UMLS® (2017AA full version) entity mentions in
41
+ these papers.
42
+
43
+ Annotation quality: We did not collect stringent IAA (Inter-annotator agreement) data. To gain
44
+ insight on the annotation quality of MedMentions, we randomly selected eight papers from the
45
+ annotated corpus, containing a total of 469 concepts. Two biologists ('Reviewer') who did not
46
+ participate in the annotation task then each reviewed four papers. The agreement between
47
+ Reviewers and Annotators, an estimate of the Precision of the annotations, was 97.3%.
48
+
49
+
50
+
51
+ ## Citation Information
52
+
53
+ ```
54
+ @misc{mohan2019medmentions,
55
+ title={MedMentions: A Large Biomedical Corpus Annotated with UMLS Concepts},
56
+ author={Sunil Mohan and Donghui Li},
57
+ year={2019},
58
+ eprint={1902.09476},
59
+ archivePrefix={arXiv},
60
+ primaryClass={cs.CL}
61
+ }
62
+
63
+ ```
bigbiohub.py ADDED
@@ -0,0 +1,592 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from collections import defaultdict
2
+ from dataclasses import dataclass
3
+ from enum import Enum
4
+ import logging
5
+ from pathlib import Path
6
+ from types import SimpleNamespace
7
+ from typing import TYPE_CHECKING, Dict, Iterable, List, Tuple
8
+
9
+ import datasets
10
+
11
+ if TYPE_CHECKING:
12
+ import bioc
13
+
14
+ logger = logging.getLogger(__name__)
15
+
16
+
17
+ BigBioValues = SimpleNamespace(NULL="<BB_NULL_STR>")
18
+
19
+
20
+ @dataclass
21
+ class BigBioConfig(datasets.BuilderConfig):
22
+ """BuilderConfig for BigBio."""
23
+
24
+ name: str = None
25
+ version: datasets.Version = None
26
+ description: str = None
27
+ schema: str = None
28
+ subset_id: str = None
29
+
30
+
31
+ class Tasks(Enum):
32
+ NAMED_ENTITY_RECOGNITION = "NER"
33
+ NAMED_ENTITY_DISAMBIGUATION = "NED"
34
+ EVENT_EXTRACTION = "EE"
35
+ RELATION_EXTRACTION = "RE"
36
+ COREFERENCE_RESOLUTION = "COREF"
37
+ QUESTION_ANSWERING = "QA"
38
+ TEXTUAL_ENTAILMENT = "TE"
39
+ SEMANTIC_SIMILARITY = "STS"
40
+ TEXT_PAIRS_CLASSIFICATION = "TXT2CLASS"
41
+ PARAPHRASING = "PARA"
42
+ TRANSLATION = "TRANSL"
43
+ SUMMARIZATION = "SUM"
44
+ TEXT_CLASSIFICATION = "TXTCLASS"
45
+
46
+
47
+ entailment_features = datasets.Features(
48
+ {
49
+ "id": datasets.Value("string"),
50
+ "premise": datasets.Value("string"),
51
+ "hypothesis": datasets.Value("string"),
52
+ "label": datasets.Value("string"),
53
+ }
54
+ )
55
+
56
+ pairs_features = datasets.Features(
57
+ {
58
+ "id": datasets.Value("string"),
59
+ "document_id": datasets.Value("string"),
60
+ "text_1": datasets.Value("string"),
61
+ "text_2": datasets.Value("string"),
62
+ "label": datasets.Value("string"),
63
+ }
64
+ )
65
+
66
+ qa_features = datasets.Features(
67
+ {
68
+ "id": datasets.Value("string"),
69
+ "question_id": datasets.Value("string"),
70
+ "document_id": datasets.Value("string"),
71
+ "question": datasets.Value("string"),
72
+ "type": datasets.Value("string"),
73
+ "choices": [datasets.Value("string")],
74
+ "context": datasets.Value("string"),
75
+ "answer": datasets.Sequence(datasets.Value("string")),
76
+ }
77
+ )
78
+
79
+ text_features = datasets.Features(
80
+ {
81
+ "id": datasets.Value("string"),
82
+ "document_id": datasets.Value("string"),
83
+ "text": datasets.Value("string"),
84
+ "labels": [datasets.Value("string")],
85
+ }
86
+ )
87
+
88
+ text2text_features = datasets.Features(
89
+ {
90
+ "id": datasets.Value("string"),
91
+ "document_id": datasets.Value("string"),
92
+ "text_1": datasets.Value("string"),
93
+ "text_2": datasets.Value("string"),
94
+ "text_1_name": datasets.Value("string"),
95
+ "text_2_name": datasets.Value("string"),
96
+ }
97
+ )
98
+
99
+ kb_features = datasets.Features(
100
+ {
101
+ "id": datasets.Value("string"),
102
+ "document_id": datasets.Value("string"),
103
+ "passages": [
104
+ {
105
+ "id": datasets.Value("string"),
106
+ "type": datasets.Value("string"),
107
+ "text": datasets.Sequence(datasets.Value("string")),
108
+ "offsets": datasets.Sequence([datasets.Value("int32")]),
109
+ }
110
+ ],
111
+ "entities": [
112
+ {
113
+ "id": datasets.Value("string"),
114
+ "type": datasets.Value("string"),
115
+ "text": datasets.Sequence(datasets.Value("string")),
116
+ "offsets": datasets.Sequence([datasets.Value("int32")]),
117
+ "normalized": [
118
+ {
119
+ "db_name": datasets.Value("string"),
120
+ "db_id": datasets.Value("string"),
121
+ }
122
+ ],
123
+ }
124
+ ],
125
+ "events": [
126
+ {
127
+ "id": datasets.Value("string"),
128
+ "type": datasets.Value("string"),
129
+ # refers to the text_bound_annotation of the trigger
130
+ "trigger": {
131
+ "text": datasets.Sequence(datasets.Value("string")),
132
+ "offsets": datasets.Sequence([datasets.Value("int32")]),
133
+ },
134
+ "arguments": [
135
+ {
136
+ "role": datasets.Value("string"),
137
+ "ref_id": datasets.Value("string"),
138
+ }
139
+ ],
140
+ }
141
+ ],
142
+ "coreferences": [
143
+ {
144
+ "id": datasets.Value("string"),
145
+ "entity_ids": datasets.Sequence(datasets.Value("string")),
146
+ }
147
+ ],
148
+ "relations": [
149
+ {
150
+ "id": datasets.Value("string"),
151
+ "type": datasets.Value("string"),
152
+ "arg1_id": datasets.Value("string"),
153
+ "arg2_id": datasets.Value("string"),
154
+ "normalized": [
155
+ {
156
+ "db_name": datasets.Value("string"),
157
+ "db_id": datasets.Value("string"),
158
+ }
159
+ ],
160
+ }
161
+ ],
162
+ }
163
+ )
164
+
165
+
166
+ TASK_TO_SCHEMA = {
167
+ Tasks.NAMED_ENTITY_RECOGNITION.name: "KB",
168
+ Tasks.NAMED_ENTITY_DISAMBIGUATION.name: "KB",
169
+ Tasks.EVENT_EXTRACTION.name: "KB",
170
+ Tasks.RELATION_EXTRACTION.name: "KB",
171
+ Tasks.COREFERENCE_RESOLUTION.name: "KB",
172
+ Tasks.QUESTION_ANSWERING.name: "QA",
173
+ Tasks.TEXTUAL_ENTAILMENT.name: "TE",
174
+ Tasks.SEMANTIC_SIMILARITY.name: "PAIRS",
175
+ Tasks.TEXT_PAIRS_CLASSIFICATION.name: "PAIRS",
176
+ Tasks.PARAPHRASING.name: "T2T",
177
+ Tasks.TRANSLATION.name: "T2T",
178
+ Tasks.SUMMARIZATION.name: "T2T",
179
+ Tasks.TEXT_CLASSIFICATION.name: "TEXT",
180
+ }
181
+
182
+ SCHEMA_TO_TASKS = defaultdict(set)
183
+ for task, schema in TASK_TO_SCHEMA.items():
184
+ SCHEMA_TO_TASKS[schema].add(task)
185
+ SCHEMA_TO_TASKS = dict(SCHEMA_TO_TASKS)
186
+
187
+ VALID_TASKS = set(TASK_TO_SCHEMA.keys())
188
+ VALID_SCHEMAS = set(TASK_TO_SCHEMA.values())
189
+
190
+ SCHEMA_TO_FEATURES = {
191
+ "KB": kb_features,
192
+ "QA": qa_features,
193
+ "TE": entailment_features,
194
+ "T2T": text2text_features,
195
+ "TEXT": text_features,
196
+ "PAIRS": pairs_features,
197
+ }
198
+
199
+
200
+ def get_texts_and_offsets_from_bioc_ann(ann: "bioc.BioCAnnotation") -> Tuple:
201
+
202
+ offsets = [(loc.offset, loc.offset + loc.length) for loc in ann.locations]
203
+
204
+ text = ann.text
205
+
206
+ if len(offsets) > 1:
207
+ i = 0
208
+ texts = []
209
+ for start, end in offsets:
210
+ chunk_len = end - start
211
+ texts.append(text[i : chunk_len + i])
212
+ i += chunk_len
213
+ while i < len(text) and text[i] == " ":
214
+ i += 1
215
+ else:
216
+ texts = [text]
217
+
218
+ return offsets, texts
219
+
220
+
221
+ def remove_prefix(a: str, prefix: str) -> str:
222
+ if a.startswith(prefix):
223
+ a = a[len(prefix) :]
224
+ return a
225
+
226
+
227
+ def parse_brat_file(
228
+ txt_file: Path,
229
+ annotation_file_suffixes: List[str] = None,
230
+ parse_notes: bool = False,
231
+ ) -> Dict:
232
+ """
233
+ Parse a brat file into the schema defined below.
234
+ `txt_file` should be the path to the brat '.txt' file you want to parse, e.g. 'data/1234.txt'
235
+ Assumes that the annotations are contained in one or more of the corresponding '.a1', '.a2' or '.ann' files,
236
+ e.g. 'data/1234.ann' or 'data/1234.a1' and 'data/1234.a2'.
237
+ Will include annotator notes, when `parse_notes == True`.
238
+ brat_features = datasets.Features(
239
+ {
240
+ "id": datasets.Value("string"),
241
+ "document_id": datasets.Value("string"),
242
+ "text": datasets.Value("string"),
243
+ "text_bound_annotations": [ # T line in brat, e.g. type or event trigger
244
+ {
245
+ "offsets": datasets.Sequence([datasets.Value("int32")]),
246
+ "text": datasets.Sequence(datasets.Value("string")),
247
+ "type": datasets.Value("string"),
248
+ "id": datasets.Value("string"),
249
+ }
250
+ ],
251
+ "events": [ # E line in brat
252
+ {
253
+ "trigger": datasets.Value(
254
+ "string"
255
+ ), # refers to the text_bound_annotation of the trigger,
256
+ "id": datasets.Value("string"),
257
+ "type": datasets.Value("string"),
258
+ "arguments": datasets.Sequence(
259
+ {
260
+ "role": datasets.Value("string"),
261
+ "ref_id": datasets.Value("string"),
262
+ }
263
+ ),
264
+ }
265
+ ],
266
+ "relations": [ # R line in brat
267
+ {
268
+ "id": datasets.Value("string"),
269
+ "head": {
270
+ "ref_id": datasets.Value("string"),
271
+ "role": datasets.Value("string"),
272
+ },
273
+ "tail": {
274
+ "ref_id": datasets.Value("string"),
275
+ "role": datasets.Value("string"),
276
+ },
277
+ "type": datasets.Value("string"),
278
+ }
279
+ ],
280
+ "equivalences": [ # Equiv line in brat
281
+ {
282
+ "id": datasets.Value("string"),
283
+ "ref_ids": datasets.Sequence(datasets.Value("string")),
284
+ }
285
+ ],
286
+ "attributes": [ # M or A lines in brat
287
+ {
288
+ "id": datasets.Value("string"),
289
+ "type": datasets.Value("string"),
290
+ "ref_id": datasets.Value("string"),
291
+ "value": datasets.Value("string"),
292
+ }
293
+ ],
294
+ "normalizations": [ # N lines in brat
295
+ {
296
+ "id": datasets.Value("string"),
297
+ "type": datasets.Value("string"),
298
+ "ref_id": datasets.Value("string"),
299
+ "resource_name": datasets.Value(
300
+ "string"
301
+ ), # Name of the resource, e.g. "Wikipedia"
302
+ "cuid": datasets.Value(
303
+ "string"
304
+ ), # ID in the resource, e.g. 534366
305
+ "text": datasets.Value(
306
+ "string"
307
+ ), # Human readable description/name of the entity, e.g. "Barack Obama"
308
+ }
309
+ ],
310
+ ### OPTIONAL: Only included when `parse_notes == True`
311
+ "notes": [ # # lines in brat
312
+ {
313
+ "id": datasets.Value("string"),
314
+ "type": datasets.Value("string"),
315
+ "ref_id": datasets.Value("string"),
316
+ "text": datasets.Value("string"),
317
+ }
318
+ ],
319
+ },
320
+ )
321
+ """
322
+
323
+ example = {}
324
+ example["document_id"] = txt_file.with_suffix("").name
325
+ with txt_file.open() as f:
326
+ example["text"] = f.read()
327
+
328
+ # If no specific suffixes of the to-be-read annotation files are given - take standard suffixes
329
+ # for event extraction
330
+ if annotation_file_suffixes is None:
331
+ annotation_file_suffixes = [".a1", ".a2", ".ann"]
332
+
333
+ if len(annotation_file_suffixes) == 0:
334
+ raise AssertionError(
335
+ "At least one suffix for the to-be-read annotation files should be given!"
336
+ )
337
+
338
+ ann_lines = []
339
+ for suffix in annotation_file_suffixes:
340
+ annotation_file = txt_file.with_suffix(suffix)
341
+ try:
342
+ with annotation_file.open() as f:
343
+ ann_lines.extend(f.readlines())
344
+ except Exception:
345
+ continue
346
+
347
+ example["text_bound_annotations"] = []
348
+ example["events"] = []
349
+ example["relations"] = []
350
+ example["equivalences"] = []
351
+ example["attributes"] = []
352
+ example["normalizations"] = []
353
+
354
+ if parse_notes:
355
+ example["notes"] = []
356
+
357
+ for line in ann_lines:
358
+ line = line.strip()
359
+ if not line:
360
+ continue
361
+
362
+ if line.startswith("T"): # Text bound
363
+ ann = {}
364
+ fields = line.split("\t")
365
+
366
+ ann["id"] = fields[0]
367
+ ann["type"] = fields[1].split()[0]
368
+ ann["offsets"] = []
369
+ span_str = remove_prefix(fields[1], (ann["type"] + " "))
370
+ text = fields[2]
371
+ for span in span_str.split(";"):
372
+ start, end = span.split()
373
+ ann["offsets"].append([int(start), int(end)])
374
+
375
+ # Heuristically split text of discontiguous entities into chunks
376
+ ann["text"] = []
377
+ if len(ann["offsets"]) > 1:
378
+ i = 0
379
+ for start, end in ann["offsets"]:
380
+ chunk_len = end - start
381
+ ann["text"].append(text[i : chunk_len + i])
382
+ i += chunk_len
383
+ while i < len(text) and text[i] == " ":
384
+ i += 1
385
+ else:
386
+ ann["text"] = [text]
387
+
388
+ example["text_bound_annotations"].append(ann)
389
+
390
+ elif line.startswith("E"):
391
+ ann = {}
392
+ fields = line.split("\t")
393
+
394
+ ann["id"] = fields[0]
395
+
396
+ ann["type"], ann["trigger"] = fields[1].split()[0].split(":")
397
+
398
+ ann["arguments"] = []
399
+ for role_ref_id in fields[1].split()[1:]:
400
+ argument = {
401
+ "role": (role_ref_id.split(":"))[0],
402
+ "ref_id": (role_ref_id.split(":"))[1],
403
+ }
404
+ ann["arguments"].append(argument)
405
+
406
+ example["events"].append(ann)
407
+
408
+ elif line.startswith("R"):
409
+ ann = {}
410
+ fields = line.split("\t")
411
+
412
+ ann["id"] = fields[0]
413
+ ann["type"] = fields[1].split()[0]
414
+
415
+ ann["head"] = {
416
+ "role": fields[1].split()[1].split(":")[0],
417
+ "ref_id": fields[1].split()[1].split(":")[1],
418
+ }
419
+ ann["tail"] = {
420
+ "role": fields[1].split()[2].split(":")[0],
421
+ "ref_id": fields[1].split()[2].split(":")[1],
422
+ }
423
+
424
+ example["relations"].append(ann)
425
+
426
+ # '*' seems to be the legacy way to mark equivalences,
427
+ # but I couldn't find any info on the current way
428
+ # this might have to be adapted dependent on the brat version
429
+ # of the annotation
430
+ elif line.startswith("*"):
431
+ ann = {}
432
+ fields = line.split("\t")
433
+
434
+ ann["id"] = fields[0]
435
+ ann["ref_ids"] = fields[1].split()[1:]
436
+
437
+ example["equivalences"].append(ann)
438
+
439
+ elif line.startswith("A") or line.startswith("M"):
440
+ ann = {}
441
+ fields = line.split("\t")
442
+
443
+ ann["id"] = fields[0]
444
+
445
+ info = fields[1].split()
446
+ ann["type"] = info[0]
447
+ ann["ref_id"] = info[1]
448
+
449
+ if len(info) > 2:
450
+ ann["value"] = info[2]
451
+ else:
452
+ ann["value"] = ""
453
+
454
+ example["attributes"].append(ann)
455
+
456
+ elif line.startswith("N"):
457
+ ann = {}
458
+ fields = line.split("\t")
459
+
460
+ ann["id"] = fields[0]
461
+ ann["text"] = fields[2]
462
+
463
+ info = fields[1].split()
464
+
465
+ ann["type"] = info[0]
466
+ ann["ref_id"] = info[1]
467
+ ann["resource_name"] = info[2].split(":")[0]
468
+ ann["cuid"] = info[2].split(":")[1]
469
+ example["normalizations"].append(ann)
470
+
471
+ elif parse_notes and line.startswith("#"):
472
+ ann = {}
473
+ fields = line.split("\t")
474
+
475
+ ann["id"] = fields[0]
476
+ ann["text"] = fields[2] if len(fields) == 3 else BigBioValues.NULL
477
+
478
+ info = fields[1].split()
479
+
480
+ ann["type"] = info[0]
481
+ ann["ref_id"] = info[1]
482
+ example["notes"].append(ann)
483
+
484
+ return example
485
+
486
+
487
+ def brat_parse_to_bigbio_kb(brat_parse: Dict) -> Dict:
488
+ """
489
+ Transform a brat parse (conforming to the standard brat schema) obtained with
490
+ `parse_brat_file` into a dictionary conforming to the `bigbio-kb` schema (as defined in ../schemas/kb.py)
491
+ :param brat_parse:
492
+ """
493
+
494
+ unified_example = {}
495
+
496
+ # Prefix all ids with document id to ensure global uniqueness,
497
+ # because brat ids are only unique within their document
498
+ id_prefix = brat_parse["document_id"] + "_"
499
+
500
+ # identical
501
+ unified_example["document_id"] = brat_parse["document_id"]
502
+ unified_example["passages"] = [
503
+ {
504
+ "id": id_prefix + "_text",
505
+ "type": "abstract",
506
+ "text": [brat_parse["text"]],
507
+ "offsets": [[0, len(brat_parse["text"])]],
508
+ }
509
+ ]
510
+
511
+ # get normalizations
512
+ ref_id_to_normalizations = defaultdict(list)
513
+ for normalization in brat_parse["normalizations"]:
514
+ ref_id_to_normalizations[normalization["ref_id"]].append(
515
+ {
516
+ "db_name": normalization["resource_name"],
517
+ "db_id": normalization["cuid"],
518
+ }
519
+ )
520
+
521
+ # separate entities and event triggers
522
+ unified_example["events"] = []
523
+ non_event_ann = brat_parse["text_bound_annotations"].copy()
524
+ for event in brat_parse["events"]:
525
+ event = event.copy()
526
+ event["id"] = id_prefix + event["id"]
527
+ trigger = next(
528
+ tr
529
+ for tr in brat_parse["text_bound_annotations"]
530
+ if tr["id"] == event["trigger"]
531
+ )
532
+ if trigger in non_event_ann:
533
+ non_event_ann.remove(trigger)
534
+ event["trigger"] = {
535
+ "text": trigger["text"].copy(),
536
+ "offsets": trigger["offsets"].copy(),
537
+ }
538
+ for argument in event["arguments"]:
539
+ argument["ref_id"] = id_prefix + argument["ref_id"]
540
+
541
+ unified_example["events"].append(event)
542
+
543
+ unified_example["entities"] = []
544
+ anno_ids = [ref_id["id"] for ref_id in non_event_ann]
545
+ for ann in non_event_ann:
546
+ entity_ann = ann.copy()
547
+ entity_ann["id"] = id_prefix + entity_ann["id"]
548
+ entity_ann["normalized"] = ref_id_to_normalizations[ann["id"]]
549
+ unified_example["entities"].append(entity_ann)
550
+
551
+ # massage relations
552
+ unified_example["relations"] = []
553
+ skipped_relations = set()
554
+ for ann in brat_parse["relations"]:
555
+ if (
556
+ ann["head"]["ref_id"] not in anno_ids
557
+ or ann["tail"]["ref_id"] not in anno_ids
558
+ ):
559
+ skipped_relations.add(ann["id"])
560
+ continue
561
+ unified_example["relations"].append(
562
+ {
563
+ "arg1_id": id_prefix + ann["head"]["ref_id"],
564
+ "arg2_id": id_prefix + ann["tail"]["ref_id"],
565
+ "id": id_prefix + ann["id"],
566
+ "type": ann["type"],
567
+ "normalized": [],
568
+ }
569
+ )
570
+ if len(skipped_relations) > 0:
571
+ example_id = brat_parse["document_id"]
572
+ logger.info(
573
+ f"Example:{example_id}: The `bigbio_kb` schema allows `relations` only between entities."
574
+ f" Skip (for now): "
575
+ f"{list(skipped_relations)}"
576
+ )
577
+
578
+ # get coreferences
579
+ unified_example["coreferences"] = []
580
+ for i, ann in enumerate(brat_parse["equivalences"], start=1):
581
+ is_entity_cluster = True
582
+ for ref_id in ann["ref_ids"]:
583
+ if not ref_id.startswith("T"): # not textbound -> no entity
584
+ is_entity_cluster = False
585
+ elif ref_id not in anno_ids: # event trigger -> no entity
586
+ is_entity_cluster = False
587
+ if is_entity_cluster:
588
+ entity_ids = [id_prefix + i for i in ann["ref_ids"]]
589
+ unified_example["coreferences"].append(
590
+ {"id": id_prefix + str(i), "entity_ids": entity_ids}
591
+ )
592
+ return unified_example
medmentions.py ADDED
@@ -0,0 +1,307 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # coding=utf-8
2
+ # Copyright 2022 The HuggingFace Datasets Authors and Simon Ott, github: nomisto
3
+ #
4
+ # Licensed under the Apache License, Version 2.0 (the "License");
5
+ # you may not use this file except in compliance with the License.
6
+ # You may obtain a copy of the License at
7
+ #
8
+ # http://www.apache.org/licenses/LICENSE-2.0
9
+ #
10
+ # Unless required by applicable law or agreed to in writing, software
11
+ # distributed under the License is distributed on an "AS IS" BASIS,
12
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
+ # See the License for the specific language governing permissions and
14
+ # limitations under the License.
15
+
16
+ """
17
+ MedMentions is a new manually annotated resource for the recognition of biomedical concepts.
18
+ What distinguishes MedMentions from other annotated biomedical corpora is its size (over 4,000
19
+ abstracts and over 350,000 linked mentions), as well as the size of the concept ontology (over
20
+ 3 million concepts from UMLS 2017) and its broad coverage of biomedical disciplines.
21
+
22
+ Corpus: The MedMentions corpus consists of 4,392 papers (Titles and Abstracts) randomly selected
23
+ from among papers released on PubMed in 2016, that were in the biomedical field, published in
24
+ the English language, and had both a Title and an Abstract.
25
+
26
+ Annotators: We recruited a team of professional annotators with rich experience in biomedical
27
+ content curation to exhaustively annotate all UMLS® (2017AA full version) entity mentions in
28
+ these papers.
29
+
30
+ Annotation quality: We did not collect stringent IAA (Inter-annotator agreement) data. To gain
31
+ insight on the annotation quality of MedMentions, we randomly selected eight papers from the
32
+ annotated corpus, containing a total of 469 concepts. Two biologists ('Reviewer') who did not
33
+ participate in the annotation task then each reviewed four papers. The agreement between
34
+ Reviewers and Annotators, an estimate of the Precision of the annotations, was 97.3%.
35
+
36
+ For more information visit: https://github.com/chanzuckerberg/MedMentions
37
+ """
38
+
39
+ import itertools as it
40
+ from typing import List
41
+
42
+ import datasets
43
+
44
+ from .bigbiohub import kb_features
45
+ from .bigbiohub import BigBioConfig
46
+ from .bigbiohub import Tasks
47
+
48
+ _LANGUAGES = ['English']
49
+ _PUBMED = True
50
+ _LOCAL = False
51
+ _CITATION = """\
52
+ @misc{mohan2019medmentions,
53
+ title={MedMentions: A Large Biomedical Corpus Annotated with UMLS Concepts},
54
+ author={Sunil Mohan and Donghui Li},
55
+ year={2019},
56
+ eprint={1902.09476},
57
+ archivePrefix={arXiv},
58
+ primaryClass={cs.CL}
59
+ }
60
+ """
61
+
62
+ _DATASETNAME = "medmentions"
63
+ _DISPLAYNAME = "MedMentions"
64
+
65
+ _DESCRIPTION = """\
66
+ MedMentions is a new manually annotated resource for the recognition of biomedical concepts.
67
+ What distinguishes MedMentions from other annotated biomedical corpora is its size (over 4,000
68
+ abstracts and over 350,000 linked mentions), as well as the size of the concept ontology (over
69
+ 3 million concepts from UMLS 2017) and its broad coverage of biomedical disciplines.
70
+
71
+ Corpus: The MedMentions corpus consists of 4,392 papers (Titles and Abstracts) randomly selected
72
+ from among papers released on PubMed in 2016, that were in the biomedical field, published in
73
+ the English language, and had both a Title and an Abstract.
74
+
75
+ Annotators: We recruited a team of professional annotators with rich experience in biomedical
76
+ content curation to exhaustively annotate all UMLS® (2017AA full version) entity mentions in
77
+ these papers.
78
+
79
+ Annotation quality: We did not collect stringent IAA (Inter-annotator agreement) data. To gain
80
+ insight on the annotation quality of MedMentions, we randomly selected eight papers from the
81
+ annotated corpus, containing a total of 469 concepts. Two biologists ('Reviewer') who did not
82
+ participate in the annotation task then each reviewed four papers. The agreement between
83
+ Reviewers and Annotators, an estimate of the Precision of the annotations, was 97.3%.
84
+ """
85
+
86
+ _HOMEPAGE = "https://github.com/chanzuckerberg/MedMentions"
87
+
88
+ _LICENSE = 'Creative Commons Zero v1.0 Universal'
89
+
90
+ _URLS = {
91
+ "medmentions_full": [
92
+ "https://github.com/chanzuckerberg/MedMentions/raw/master/full/data/corpus_pubtator.txt.gz",
93
+ "https://github.com/chanzuckerberg/MedMentions/raw/master/full/data/corpus_pubtator_pmids_trng.txt",
94
+ "https://github.com/chanzuckerberg/MedMentions/raw/master/full/data/corpus_pubtator_pmids_dev.txt",
95
+ "https://github.com/chanzuckerberg/MedMentions/raw/master/full/data/corpus_pubtator_pmids_test.txt",
96
+ ],
97
+ "medmentions_st21pv": [
98
+ "https://github.com/chanzuckerberg/MedMentions/raw/master/st21pv/data/corpus_pubtator.txt.gz",
99
+ "https://github.com/chanzuckerberg/MedMentions/raw/master/full/data/corpus_pubtator_pmids_trng.txt",
100
+ "https://github.com/chanzuckerberg/MedMentions/raw/master/full/data/corpus_pubtator_pmids_dev.txt",
101
+ "https://github.com/chanzuckerberg/MedMentions/raw/master/full/data/corpus_pubtator_pmids_test.txt",
102
+ ],
103
+ }
104
+
105
+ _SUPPORTED_TASKS = [Tasks.NAMED_ENTITY_DISAMBIGUATION, Tasks.NAMED_ENTITY_RECOGNITION]
106
+
107
+ _SOURCE_VERSION = "1.0.0"
108
+
109
+ _BIGBIO_VERSION = "1.0.0"
110
+
111
+
112
+ class MedMentionsDataset(datasets.GeneratorBasedBuilder):
113
+ """MedMentions dataset for named-entity disambiguation (NED)"""
114
+
115
+ SOURCE_VERSION = datasets.Version(_SOURCE_VERSION)
116
+ BIGBIO_VERSION = datasets.Version(_BIGBIO_VERSION)
117
+
118
+ BUILDER_CONFIGS = [
119
+ BigBioConfig(
120
+ name="medmentions_full_source",
121
+ version=SOURCE_VERSION,
122
+ description="MedMentions Full source schema",
123
+ schema="source",
124
+ subset_id="medmentions_full",
125
+ ),
126
+ BigBioConfig(
127
+ name="medmentions_full_bigbio_kb",
128
+ version=BIGBIO_VERSION,
129
+ description="MedMentions Full BigBio schema",
130
+ schema="bigbio_kb",
131
+ subset_id="medmentions_full",
132
+ ),
133
+ BigBioConfig(
134
+ name="medmentions_st21pv_source",
135
+ version=SOURCE_VERSION,
136
+ description="MedMentions ST21pv source schema",
137
+ schema="source",
138
+ subset_id="medmentions_st21pv",
139
+ ),
140
+ BigBioConfig(
141
+ name="medmentions_st21pv_bigbio_kb",
142
+ version=BIGBIO_VERSION,
143
+ description="MedMentions ST21pv BigBio schema",
144
+ schema="bigbio_kb",
145
+ subset_id="medmentions_st21pv",
146
+ ),
147
+ ]
148
+
149
+ DEFAULT_CONFIG_NAME = "medmentions_full_source"
150
+
151
+ def _info(self) -> datasets.DatasetInfo:
152
+
153
+ if self.config.schema == "source":
154
+ features = datasets.Features(
155
+ {
156
+ "pmid": datasets.Value("string"),
157
+ "passages": [
158
+ {
159
+ "type": datasets.Value("string"),
160
+ "text": datasets.Sequence(datasets.Value("string")),
161
+ "offsets": datasets.Sequence([datasets.Value("int32")]),
162
+ }
163
+ ],
164
+ "entities": [
165
+ {
166
+ "text": datasets.Sequence(datasets.Value("string")),
167
+ "offsets": datasets.Sequence([datasets.Value("int32")]),
168
+ "concept_id": datasets.Value("string"),
169
+ "semantic_type_id": datasets.Sequence(
170
+ datasets.Value("string")
171
+ ),
172
+ }
173
+ ],
174
+ }
175
+ )
176
+
177
+ elif self.config.schema == "bigbio_kb":
178
+ features = kb_features
179
+
180
+ return datasets.DatasetInfo(
181
+ description=_DESCRIPTION,
182
+ features=features,
183
+ homepage=_HOMEPAGE,
184
+ license=str(_LICENSE),
185
+ citation=_CITATION,
186
+ )
187
+
188
+ def _split_generators(self, dl_manager) -> List[datasets.SplitGenerator]:
189
+
190
+ urls = _URLS[self.config.subset_id]
191
+ (
192
+ corpus_path,
193
+ pmids_train,
194
+ pmids_dev,
195
+ pmids_test,
196
+ ) = dl_manager.download_and_extract(urls)
197
+
198
+ return [
199
+ datasets.SplitGenerator(
200
+ name=datasets.Split.TRAIN,
201
+ gen_kwargs={"corpus_path": corpus_path, "pmids_path": pmids_train},
202
+ ),
203
+ datasets.SplitGenerator(
204
+ name=datasets.Split.TEST,
205
+ gen_kwargs={"corpus_path": corpus_path, "pmids_path": pmids_test},
206
+ ),
207
+ datasets.SplitGenerator(
208
+ name=datasets.Split.VALIDATION,
209
+ gen_kwargs={"corpus_path": corpus_path, "pmids_path": pmids_dev},
210
+ ),
211
+ ]
212
+
213
+ def _generate_examples(self, corpus_path, pmids_path):
214
+ with open(pmids_path, encoding="utf8") as infile:
215
+ pmids = infile.readlines()
216
+ pmids = {int(x.strip()) for x in pmids}
217
+
218
+ if self.config.schema == "source":
219
+ with open(corpus_path, encoding="utf8") as corpus:
220
+ for document in self._generate_parsed_documents(corpus, pmids):
221
+ yield document["pmid"], document
222
+
223
+ elif self.config.schema == "bigbio_kb":
224
+ uid = it.count(0)
225
+ with open(corpus_path, encoding="utf8") as corpus:
226
+ for document in self._generate_parsed_documents(corpus, pmids):
227
+ document["id"] = next(uid)
228
+ document["document_id"] = document.pop("pmid")
229
+
230
+ entities_ = []
231
+ for entity in document["entities"]:
232
+ for type in entity["semantic_type_id"]:
233
+ entities_.append(
234
+ {
235
+ "id": next(uid),
236
+ "type": type,
237
+ "text": entity["text"],
238
+ "offsets": entity["offsets"],
239
+ "normalized": [
240
+ {
241
+ "db_name": "UMLS",
242
+ "db_id": entity["concept_id"].split(":")[-1],
243
+ }
244
+ ],
245
+ }
246
+ )
247
+ document["entities"] = entities_
248
+
249
+ for passage in document["passages"]:
250
+ passage["id"] = next(uid)
251
+ document["relations"] = []
252
+ document["events"] = []
253
+ document["coreferences"] = []
254
+ yield document["document_id"], document
255
+
256
+ def _generate_parsed_documents(self, fstream, pmids):
257
+ for raw_document in self._generate_raw_documents(fstream):
258
+ if self._parse_pmid(raw_document) in pmids:
259
+ yield self._parse_document(raw_document)
260
+
261
+ def _generate_raw_documents(self, fstream):
262
+ raw_document = []
263
+ for line in fstream:
264
+ if line.strip():
265
+ raw_document.append(line.strip())
266
+ elif raw_document:
267
+ yield raw_document
268
+ raw_document = []
269
+ # needed for last document
270
+ if raw_document:
271
+ yield raw_document
272
+
273
+ def _parse_pmid(self, raw_document):
274
+ pmid, _ = raw_document[0].split("|", 1)
275
+ return int(pmid)
276
+
277
+ def _parse_document(self, raw_document):
278
+ pmid, type, title = raw_document[0].split("|", 2)
279
+ pmid_, type, abstract = raw_document[1].split("|", 2)
280
+ passages = [
281
+ {"type": "title", "text": [title], "offsets": [[0, len(title)]]},
282
+ {
283
+ "type": "abstract",
284
+ "text": [abstract],
285
+ "offsets": [[len(title) + 1, len(title) + len(abstract) + 1]],
286
+ },
287
+ ]
288
+
289
+ entities = []
290
+ for line in raw_document[2:]:
291
+ (
292
+ pmid_,
293
+ start_idx,
294
+ end_idx,
295
+ mention,
296
+ semantic_type_id,
297
+ entity_id,
298
+ ) = line.split("\t")
299
+ entity = {
300
+ "offsets": [[int(start_idx), int(end_idx)]],
301
+ "text": [mention],
302
+ "semantic_type_id": semantic_type_id.split(","),
303
+ "concept_id": entity_id,
304
+ }
305
+ entities.append(entity)
306
+
307
+ return {"pmid": int(pmid), "entities": entities, "passages": passages}