|
|
--- |
|
|
license: mit |
|
|
language: |
|
|
- en |
|
|
tags: |
|
|
- life-sciences |
|
|
- clinical |
|
|
- biomedical |
|
|
- bio |
|
|
- medical |
|
|
- biology |
|
|
- synthetic |
|
|
pretty_name: TransCorpus-bio |
|
|
size_categories: |
|
|
- 10M<n<100M |
|
|
--- |
|
|
# TransCorpus-bio |
|
|
**TransCorpus-bio** is a large-scale, parallel biomedical corpus consisting of PubMed abstracts. This dataset is used in the [TransCorpus Toolkit](https://github.com/jknafou/TransCorpus) and is designed to enable high-quality multi-lingual biomedical language modeling and downstream NLP research. |
|
|
|
|
|
# Currently Translated with [TransCorpus Toolkit](https://github.com/jknafou/TransCorpus) |
|
|
- In French : [TransCorpus-bio-fr 🤗](https://huggingface.co/datasets/jknafou/TransCorpus-bio-fr) |
|
|
- In Spanish : [TransCorpus-bio-es 🤗](https://huggingface.co/datasets/jknafou/TransCorpus-bio-es) |
|
|
|
|
|
# Dataset Details |
|
|
- **Source**: PubMed abstracts (English) |
|
|
- **Size**: 22 million abstracts, 30.2GB of text |
|
|
- **Domain**: Biomedical, clinical, life sciences |
|
|
- **Format**: one abstract per line |
|
|
|
|
|
# Motivation |
|
|
Non-English languages are low-resource languages for biomedical NLP, with limited availability of large, high-quality corpora. **TransCorpus-bio** bridges this gap by leveraging state-of-the-art neural machine translation to generate a massive, high-quality synthetic corpus, enabling robust pretraining and evaluation of Spanish biomedical language models. |
|
|
|
|
|
```python |
|
|
from datasets import load_dataset |
|
|
|
|
|
dataset = load_dataset("jknafou/TransCorpus-bio", split="train") |
|
|
|
|
|
print(dataset) |
|
|
# Output: |
|
|
# Dataset({ |
|
|
# features: ['text'], |
|
|
# num_rows: 21567136 |
|
|
# }) |
|
|
|
|
|
print(dataset[0]) |
|
|
``` |
|
|
|
|
|
# Benchmark Results in our French Experiment |
|
|
[TransBERT-bio-fr](https://huggingface.co/jknafou/TransBERT-bio-fr) pretrained on TransCorpus-bio-fr achieve state-of-the-art results on the French biomedical benchmark DrBenchmark, outperforming both general-domain and previous domain-specific models on classification, NER, POS, and STS tasks. See TransBERT-bio-fr for details. |
|
|
|
|
|
# Why Synthetic Translation? |
|
|
- **Scalable**: Enables creation of large-scale corpora for any language with a strong MT system. |
|
|
- **Effective**: Supports state-of-the-art performance in downstream tasks. |
|
|
- **Accessible**: Makes domain-specific NLP feasible for any languages. |
|
|
|
|
|
# Citation |
|
|
If you use this corpus, please cite: |
|
|
```text |
|
|
@inproceedings{knafou-etal-2025-transbert, |
|
|
title = "{T}rans{BERT}: A Framework for Synthetic Translation in Domain-Specific Language Modeling", |
|
|
author = {Knafou, Julien and |
|
|
Mottin, Luc and |
|
|
Mottaz, Ana{\"i}s and |
|
|
Flament, Alexandre and |
|
|
Ruch, Patrick}, |
|
|
editor = "Christodoulopoulos, Christos and |
|
|
Chakraborty, Tanmoy and |
|
|
Rose, Carolyn and |
|
|
Peng, Violet", |
|
|
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2025", |
|
|
month = nov, |
|
|
year = "2025", |
|
|
address = "Suzhou, China", |
|
|
publisher = "Association for Computational Linguistics", |
|
|
url = "https://aclanthology.org/2025.findings-emnlp.1053/", |
|
|
doi = "10.18653/v1/2025.findings-emnlp.1053", |
|
|
pages = "19338--19354", |
|
|
ISBN = "979-8-89176-335-7", |
|
|
abstract = "The scarcity of non-English language data in specialized domains significantly limits the development of effective Natural Language Processing (NLP) tools. We present TransBERT, a novel framework for pre-training language models using exclusively synthetically translated text, and introduce TransCorpus, a scalable translation toolkit. Focusing on the life sciences domain in French, our approach demonstrates that state-of-the-art performance on various downstream tasks can be achieved solely by leveraging synthetically translated data. We release the TransCorpus toolkit, the TransCorpus-bio-fr corpus (36.4GB of French life sciences text), TransBERT-bio-fr, its associated pre-trained language model and reproducible code for both pre-training and fine-tuning. Our results highlight the viability of synthetic translation in a high-resource translation direction for building high-quality NLP resources in low-resource language/domain pairs." |
|
|
} |
|
|
``` |
|
|
|
|
|
|