File size: 17,445 Bytes
5e4cf3f
9b1768c
 
 
 
 
 
 
 
 
 
 
 
 
5e4cf3f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9b1768c
b16bc40
 
 
 
 
 
 
 
 
 
 
a0fc9c5
b16bc40
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
184ed37
b16bc40
184ed37
 
 
 
 
 
 
 
 
 
 
 
 
 
 
b16bc40
 
 
184ed37
b16bc40
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
70a85f7
 
 
 
 
 
 
 
 
9b1768c
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
---
language:
- en
- fr
- es
- zh
- de
- it
- pt
- ko
- ru
pretty_name: Biomed-Enriched
task_categories:
- text-classification
dataset_info:
  features:
  - name: id
    dtype: string
  - name: article_id
    dtype: string
  - name: path
    dtype: string
  - name: text
    dtype: string
  - name: language
    dtype: string
  - name: section_title
    dtype: string
  - name: domain
    dtype: string
  - name: document_type
    dtype: string
  - name: educational_score
    dtype: float64
  - name: domain_scores
    sequence: float64
  - name: document_type_scores
    sequence: float64
  - name: language_score
    dtype: float64
  - name: authors
    sequence: string
  - name: article_url
    dtype: string
  - name: license_url
    dtype: string
  splits:
  - name: commercial
    num_bytes: 126945006652
    num_examples: 98595508
  - name: noncommercial
    num_bytes: 19998737015
    num_examples: 47449067
  download_size: 52228107549
  dataset_size: 146943743667
configs:
- config_name: default
  data_files:
  - split: commercial
    path: data/commercial-*
  - split: noncommercial
    path: data/noncommercial-*
---

![Biomed-Enriched](logo.png)

# Biomed-Enriched: A Biomedical Dataset Enriched with LLMs for Pretraining and Extracting Rare and Hidden Content 

### Dataset Authors

**Rian Touchent, Nathan Godey & Eric de la Clergerie**  
*Sorbonne Université, INRIA Paris*

### Overview

Biomed-Enriched is a PubMed-derived dataset created using a two-stage annotation process. Initially, [Llama 3.1 70B Instruct](https://huggingface.co/meta-llama/Llama-3.1-70B-Instruct) annotated 400K paragraphs for document type, domain, and educational quality. These annotations were then used to fine-tune a [smaller model](https://huggingface.co/almanach/Biomed-Enriched-classifier), which propagated the labels across the entire PubMed Central Open Access corpus. This process yielded 2M clinical case paragraphs, with over 450K high-quality paragraphs licensed for commercial use. This dataset provides a large-scale, openly available alternative to private clinical text. In continual pre-training experiments with OLMo2, curated subsets showed targeted improvements: clinical upsampling boosted MMLU ProfMed scores by ~5%, and educational quality filtering improved MedQA and MedMCQA by ~1%. Combining these methods achieved similar performance compared to standard continual-pretraining with just one-third of the training tokens, highlighting the potential for more efficient biomedical pretraining.

The dataset is structured into two primary splits:

* **Commercial**
* **Non-Commercial**

## Dataset Structure

### Commercial Split

* **text**: Textual content of the paragraphs.
* **path**: Precise XML path referencing original paragraph locations.
* **license\_url**: URL linking to the license.
* **authors**: Comprehensive list of authors per paragraph for proper attribution compliance.

### Non-Commercial Split

* **path**: Precise XML path referencing original paragraph locations.
* **license\_url**: URL linking to the license.
* **authors**: Comprehensive list of authors per paragraph for proper attribution compliance.

> **Note:** The non-commercial split does not contain text data due to licensing restrictions. However, we provide scripts to populate the `text` field from a local PMC Open Access XML dump. See below for installation and usage instructions.

```bash
pip install biomed-enriched
```

### With Python
```python
from biomed_enriched import populate

DATASET_DIR = "/path/to/biomed-enriched"  # input dataset
PMC_XML_ROOT = "/path/to/pmc/non-comm/xml"          # PMC XML dump
OUTPUT_DIR = "/path/to/populated-biomed-enriched"  # drop arg to overwrite in-place

populate(DATASET_DIR, PMC_XML_ROOT, output_path=OUTPUT_DIR, splits="noncommercial", num_proc=1)
```
The call overwrites the dataset in-place, adding a new `text` column as the third column (after `article_id`, `path`).

### With CLI
```bash
biomed-enriched \
  --input /path/to/biomed-enriched \
  --xml-root /path/to/pmc/non-comm/xml \
  --num-proc 8
```
Add `--output DIR` if you prefer writing to a new directory instead of overwriting.

## Annotation Process

The dataset was created using a two-stage annotation framework:

1. **Initial Annotation by Large Language Model**:

* Annotated a subset of paragraphs for the following categories:
   
 - **Document Type**: Categorizes the structure and purpose of the content.
   - **Clinical Case**: Detailed report of symptoms, diagnosis, treatment, and follow-up of individual patients.
   - **Study**: Research paragraph with methods, results, and discussion of experiments or observations.
   - **Review**: Summary or synthesis of current knowledge on a specific topic.
   - **Other**: Content not fitting above categories (editorials, commentaries, policy paragraphs).
 - **Domain**: Identifies the subject area focus.
   - **Clinical**: Content relating to patient care, clinical trials, case reports, or practice guidelines.
   - **Biomedical**: Scientific aspects of medicine and biology.
   - **Other**: Content mentioning biomedical topics but focusing on administrative, policy, or general communications.
 - **Educational Quality**: Assesses pedagogical value for college-level biomedical learning on a scale from 1 (minimal value) to 5 (exceptional value) inspired by FineWeb-edu.
   - **Score 1**: Basic information relevant to biomedical topics, may contain irrelevant content.
   - **Score 2**: Addresses biomedical education elements but with limitations in coherence or depth.
   - **Score 3**: Appropriate for college-level curricula, introduces key concepts with reasonable coherence.
   - **Score 4**: Highly relevant educational content with clear writing style, minimal irrelevant information.
   - **Score 5**: Outstanding educational value, detailed reasoning with profound insights for college-level learning.

2. **Annotation Scaling via Model Distillation**:

* These annotations were distilled into a XLM-RoBERTa-base model, enabling scalability to the entire PMC dataset.

## Annotation Statistics

Here is the distribution of educational scores per domain:

| Educational Score | Biomedical (n=116 221 134) | Clinical (n=2 182 784) | Other (n=15 213 051) |
| :---------------: | :------------------------: | :--------------------: | :------------------: |
|         1         |            1.8 %           |          6.0 %         |        60.1 %        |
|         2         |            9.4 %           |         23.4 %         |        29.4 %        |
|         3         |           10.9 %           |         26.6 %         |         8.3 %        |
|         4         |           75.3 %           |         44.0 %         |         2.1 %        |
|         5         |            2.6 %           |            –           |           –          |

Here is the distribution of educational scores per document type:

| Educational Score | Study (n=100 387 809) | Review (n=6 811 226) | Clinical case (n=2 122 403) | Other (n=24 295 531) |
|:-----------------:|:---------------------:|:--------------------:|:---------------------------:|:--------------------:|
| 1                 | 0.7 %                 | 0.3 %                | 4.0 %                       | 43.4 %               |
| 2                 | 7.9 %                 | 1.6 %                | 14.4 %                      | 30.9 %               |
| 3                 | 10.0 %                | 5.0 %                | 24.6 %                      | 14.6 %               |
| 4                 | 78.7 %                | 86.9 %               | 57.0 %                      | 11.1 %               |
| 5                 | 2.6 %                 | 6.1 %                | 0.0 %                       | 0.0 %                |

### Language Distribution


| Language | Articles | Paragraphs | Clinical Case Paragraphs | % Clinical Cases |
|----------|----------|------------|-------------------------|-----------------|
| en       | 4,113,275| 131,579,445| 2,113,185               | 1.61            |
| es       | 4,339    | 181,779    | 1,235                   | 0.68            |
| zh-cn    | 3,649    | 59,719     | 0                       | 0.00            |
| fr       | 3,410    | 173,325    | 2,586                   | 1.49            |
| de       | 2,976    | 248,608    | 51                      | 0.02            |
| it       | 2,708    | 274,819    | 521                     | 0.19            |
| pt       | 934      | 85,242     | 4,540                   | 5.33            |
| ko       | 636      | 25,535     | 0                       | 0.00            |
| ru       | 222      | 10,553     | 0                       | 0.00            |
| id       | 189      | 91,865     | 15                      | 0.02            |


## Key Applications

* Improve efficiency in biomedical pretraining by focusing on high-quality, specific content.
* Create new biomedical subsets tailored to specific research needs based on document type and domain.

## Evaluation

Our evaluation focuses on isolating the effects of data curation rather than pursuing state-of-the-art scores on benchmarks. A more powerful foundation model would likely yield higher absolute scores but would obscure the precise impact of our dataset. We therefore selected `OLMo2-7B-stage1` as our foundation model, as this intermediate checkpoint provides strong baseline capabilities while allowing for a clear attribution of performance gains to our enrichment strategies. This model has already developed strong language modeling capabilities but precedes the knowledge-intensive tuning of stage 2, providing an ideal balance without the risk of catastrophic forgetting of instruction-following abilities during domain adaptation. Notably, the data mix used in phase 1 includes DCLM, a dataset filtered from web data using a classifier trained on instruction data, which gives OLMo2-7B relatively strong question-answering capabilities even after stage 1.

Each Biomed-Enriched variant was trained for exactly 33.6 billion tokens using identical hyperparameters. We follow the annealing strategy of OLMo2 used in the mid-training phase. By maintaining strict parameter parity across experiments, we created a controlled environment focused solely on measuring the effectiveness of different data curation strategies.

These experiments are designed to illustrate how our granular annotations enable targeted improvements in model capabilities. For instance, by specifically upsampling clinical content (`BE-Clinical` and `BE-ClinicalCase` variants), we expect to see a notable increase in performance on the MMLU Professional Medicine benchmark, underscoring the dataset's potential for developing specialized models.

The following variants were created for this evaluation:
*   **BE-Base:** The complete unmodified PMC Open Access Subset serving as baseline.
*   **BE-Educational:** Preserves all articles but removes paragraphs with educational quality scores below 3.
*   **BE-Clinical:** Replicates articles with predominantly clinical domain content 10× in the training mix.
*   **BE-ClinicalCase:** Replicates articles containing at least one clinical case paragraph 10× to increase exposure to clinical narratives.
*   **BE-Prefix:** Prefixes each paragraph with its predicted annotations to allow modeling of metadata-content relationships.
*   **BE-French:** Upsamples articles containing French text 10× to address language imbalance.
*   **BE-All:** Combines quality filtering (score ≥ 3), upsampling of clinical content, French text, and clinical cases, plus metadata prefixing.

### Performance Results

**SOTA Models for reference**

| Model             | MedQA | MedMCQA | PubMedQA | Anat  | Clin  | Bio   | Med   | Gen   | Prof  | Avg   |
| ----------------- | :---: | :-----: | :------: | :---: | :---: | :---: | :---: | :---: | :---: | :---: |
| Llama-3-8B        | 59.70 |  57.47  |  74.80   | 68.89 | 74.72 | 78.47 | 61.85 | 83.00 | 70.22 | 69.90 |
| Meditron-70B      | 57.10 |  46.80  |  76.60   | 53.30 | 66.70 | 76.30 | 63.00 | 69.00 | 71.60 | 64.49 |

**Benchmark Results by Dataset Variant (continual pre-training of OLMo2-7B-stage1)**

| Variant         | MedQA          | MedMCQA        | PubMedQA       | Anat           | Clin           | Bio            | Med            | Gen            | Prof           | Avg            |
| --------------- | :------------: | :------------: | :------------: | :------------: | :------------: | :------------: | :------------: | :------------: | :------------: | :------------: |
| OLMo2-7B-stage1   | 45.33 |  41.14  |  75.60   | 54.81 | 63.40 | 69.44 | 53.18 | 69.00 | 59.93 | 59.09 |
| BE-Base         | 44.85          | 41.91          | 76.40          | 57.04          | 64.15          | **70.83**      | **59.54**      | 69.00          | 59.93          | 60.41          |
| BE-Clinical     | 41.95          | 39.35          | 76.60          | 53.33          | 63.40          | 65.28          | 58.38          | 66.00          | **63.97**      | 58.70          |
| BE-ClinicalCase | 42.11          | 39.52          | 76.60          | 57.04          | 64.91          | 66.67          | **59.54**      | 69.00          | 62.87          | 59.81          |
| BE-Prefix       | 45.72          | 41.76          | **77.80**      | 57.04          | 64.53          | 68.75          | 57.23          | 66.00          | 61.76          | 60.07          |
| BE-Educational  | 45.64          | **43.08**      | 77.00          | 57.04          | 65.28          | 68.06          | 56.65          | **71.00**      | 58.82          | 60.29          |
| BE-All          | **47.21**      | 42.79          | 76.60          | **60.00**      | **65.66**      | 68.06          | 58.96          | 69.00          | 61.40          | **61.08**      |

*Note: The first three columns represent Medical QA benchmarks. The following six (Anat, Clin, Bio, Med, Gen, Prof) are sub-tasks from MMLU Medical. MMLU abbreviations: Anat=Anatomy, Clin=Clinical Knowledge, Bio=College Biology, Med=College Medicine, Gen=Medical Genetics, Prof=Professional Medicine.*

### Results Analysis

**Overall performance.** BE-All achieved the highest average performance across benchmarks at 61.08%, surpassing BE-Base (60.41%) by a small but consistent margin (+0.67 pts, see table above). Its strongest improvements appeared in MedQA (47.21%), MMLU Anatomy (60.00%), and Clinical Knowledge (65.66%), suggesting the effectiveness of combining multiple targeted enrichment strategies.

**Clinical enrichment.** Clinical enrichment (BE-Clinical) significantly boosted performance on MMLU Professional Medicine benchmark (63.97%, +4.04 pts vs. BE-Base, Figure 2). This improvement was stable from early training, highlighting how clinical narratives enhance the model’s clinical reasoning abilities efficiently.

**Educational filtering.** Educational filtering (BE-Educational) consistently improved performance on medical question-answering tasks, notably Medical Genetics (71.00%, +2 pts), MedMCQA (43.08%, +1.17 pts), and PubMedQA (77.00%, +0.6 pts). These tasks likely benefit from the knowledge present in educationally high-quality paragraphs (Figure 2).

**Metadata prefixing.** Metadata prefixing (BE-Prefix) specifically improved performance on PubMedQA (77.80%, +1.4 pts vs. BE-Base). Providing explicit paragraph-level metadata helped primarily with structured document comprehension, but it had limited benefits for other tasks.

**General biomedical knowledge trade-off.** BE-Base performed better on College Biology (70.83%) than others. Building a biology variant (BE-Bio) could be an interesting future direction, as the current dataset does not specifically target this domain.

**Non-English enrichment.** BE-French showed clear improvements in French medical QA (FrenchMedMCQA), achieving 40.5% accuracy, significantly surpassing BE-Base and the OLMo2-7B-stage1 baseline (38.32%, Figure 1). These results illustrate effective adaptation to non-English contexts without modifying the underlying model architecture.

![Benchmark Results](french_medmcqa_performance.png)

**Data efficiency and training stability.** As shown in Figure 2, BE-All reached robust benchmark performance using roughly one-third of the tokens required by BE-Base. Individual enrichments (Educational, Clinical) also displayed early and stable improvements, underscoring potential reductions in training time and computational cost.

![Benchmark Results](combined_plot.png)

## Licensing

The Biomed-Enriched annotations (document type, domain, educational quality scores, and metadata) are released under the MIT License.

The textual content licensing depends on the individual article licenses from PubMed Central Open Access. Each paragraph includes a `license_url` field pointing to the specific license. Users must comply with the respective license terms when using the textual data.

## How to Cite

Please cite Biomed-Enriched using:

```
@misc{touchent2025biomedenrichedbiomedicaldatasetenriched,
      title={Biomed-Enriched: A Biomedical Dataset Enriched with LLMs for Pretraining and Extracting Rare and Hidden Content}, 
      author={Rian Touchent and Nathan Godey and Eric de la Clergerie},
      year={2025},
      eprint={2506.20331},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2506.20331}, 
}
```

[Paper](https://huggingface.co/papers/2506.20331)