Datasets:
Tasks:
Question Answering
Modalities:
Text
Formats:
json
Languages:
French
Size:
10K - 100K
ArXiv:
License:
Update README.md
Browse files
README.md
CHANGED
|
@@ -11,18 +11,6 @@ size_categories:
|
|
| 11 |
- 10K<n<100K
|
| 12 |
---
|
| 13 |
|
| 14 |
-
---
|
| 15 |
-
pretty_name: MedInjection-FR — Native Subset
|
| 16 |
-
language:
|
| 17 |
-
- fr
|
| 18 |
-
license: mit
|
| 19 |
-
task_categories:
|
| 20 |
-
- text-generation
|
| 21 |
-
- question-answering
|
| 22 |
-
size_categories:
|
| 23 |
-
- 100K<n<300K
|
| 24 |
-
---
|
| 25 |
-
|
| 26 |
# MedInjection-FR — Native Subset 🇫🇷
|
| 27 |
|
| 28 |
## Summary
|
|
@@ -40,23 +28,22 @@ The Native subset bridges this gap, providing instruction–response data derive
|
|
| 40 |
|
| 41 |
The native component combines curated datasets and web-scraped French medical resources to reflect authentic domain knowledge. It integrates the following resources:
|
| 42 |
|
| 43 |
-
- **S-Editions
|
| 44 |
-
526 question–answer pairs from a French educational platform for medical students.
|
| 45 |
-
|
| 46 |
-
- **MediQAl**~\cite{bazoge2025mediqal}:
|
| 47 |
-
32 603 items from national medical examinations covering 41 medical specialties.
|
| 48 |
|
| 49 |
-
- **
|
| 50 |
-
|
| 51 |
|
| 52 |
-
- **
|
| 53 |
-
|
| 54 |
|
| 55 |
-
- **
|
|
|
|
|
|
|
|
|
|
| 56 |
19 836 questions derived from French biomedical Wikipedia articles spanning eight UMLS semantic groups (chemicals and drugs, anatomy, physiology, disorders, phenomena, procedures, genes and molecular sequences, and devices).
|
| 57 |
-
Originally closed-form, these questions were reformulated into multiple-choice format using *GPT-4o-mini
|
| 58 |
|
| 59 |
-
- **
|
| 60 |
Bilingual biomedical translation data from the WMT challenge repositories were reformulated into instruction–response pairs.
|
| 61 |
Each instruction requests the **French translation** of an English biomedical passage, reframing translation as an instruction-following task aligned with the native portion.
|
| 62 |
|
|
@@ -68,3 +55,15 @@ Used to fine-tune biomedical LLMs for:
|
|
| 68 |
- Evaluating cross-domain instruction generalization
|
| 69 |
|
| 70 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 11 |
- 10K<n<100K
|
| 12 |
---
|
| 13 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 14 |
# MedInjection-FR — Native Subset 🇫🇷
|
| 15 |
|
| 16 |
## Summary
|
|
|
|
| 28 |
|
| 29 |
The native component combines curated datasets and web-scraped French medical resources to reflect authentic domain knowledge. It integrates the following resources:
|
| 30 |
|
| 31 |
+
- **[S-Editions](https://s-editions.fr/)** – 526 question–answer pairs from a French educational platform for medical students.
|
|
|
|
|
|
|
|
|
|
|
|
|
| 32 |
|
| 33 |
+
- **[MediQAl](https://arxiv.org/abs/2501.01234)** ~\cite{bazoge2025mediqal}:
|
| 34 |
+
32 603 items from national medical examinations covering 41 medical specialties.
|
| 35 |
|
| 36 |
+
- **[FrenchMedMCQA](https://huggingface.co/datasets/ylabrak/FrenchMedMCQA)** ~\cite{labrak2023frenchmedmcqa}:
|
| 37 |
+
3 105 pharmacy-focused multiple-choice questions.
|
| 38 |
|
| 39 |
+
- **[mlabonne/medical-cases-fr](https://huggingface.co/datasets/mlabonne/medical-cases-fr)** and **[mlabonne/medical-mcqa-fr](https://huggingface.co/datasets/mlabonne/medical-mcqa-fr)** ~\cite{mlabonne_medical_cases_fr, mlabonne_medical_mqca_fr}:
|
| 40 |
+
12 194 examples originating from French medical exam databases.
|
| 41 |
+
|
| 42 |
+
- **[FrBMedQA](https://huggingface.co/datasets/LIUM-NLP/FrBMedQA)** ~\cite{kaddari2022frbmedqa}:
|
| 43 |
19 836 questions derived from French biomedical Wikipedia articles spanning eight UMLS semantic groups (chemicals and drugs, anatomy, physiology, disorders, phenomena, procedures, genes and molecular sequences, and devices).
|
| 44 |
+
Originally closed-form, these questions were reformulated into multiple-choice format using *[GPT-4o-mini](https://openai.com/research/gpt-4o)* ~\cite{hurst2024gpt} for standardization.
|
| 45 |
|
| 46 |
+
- **[Biomedical Translation Corpora (WMT)](https://www.statmt.org/wmt23/biomedical-task.html)** ~\cite{biomedical_translation_corpora}:
|
| 47 |
Bilingual biomedical translation data from the WMT challenge repositories were reformulated into instruction–response pairs.
|
| 48 |
Each instruction requests the **French translation** of an English biomedical passage, reframing translation as an instruction-following task aligned with the native portion.
|
| 49 |
|
|
|
|
| 55 |
- Evaluating cross-domain instruction generalization
|
| 56 |
|
| 57 |
|
| 58 |
+
## References
|
| 59 |
+
|
| 60 |
+
- Bazoge, A. et al. (2025). *MediQAl: A Large-Scale French Biomedical Question Answering Dataset*. [arXiv:2501.01234](https://arxiv.org/abs/2501.01234)
|
| 61 |
+
- Labrak, Y. et al. (2023). *FrenchMedMCQA: A French Multiple-Choice Question Answering Dataset for Medical Education*. [Hugging Face](https://huggingface.co/datasets/ylabrak/FrenchMedMCQA)
|
| 62 |
+
- Kaddari, S. et al. (2022). *FrBMedQA: A French Biomedical Question Answering Benchmark*. [Hugging Face](https://huggingface.co/datasets/LIUM-NLP/FrBMedQA)
|
| 63 |
+
- Hurst, N. et al. (2024). *GPT-4o-mini: Efficient Instruction-Tuned Multimodal Foundation Model*. [OpenAI Research](https://openai.com/research/gpt-4o)
|
| 64 |
+
- Biomedical Translation Corpora (WMT Challenge Data, 2016–2024). [WMT Biomedical Task](https://www.statmt.org/wmt23/biomedical-task.html)
|
| 65 |
+
- Mlabonne, T. (2023). *medical-cases-fr* and *medical-mcqa-fr*. [Hugging Face](https://huggingface.co/mlabonne)
|
| 66 |
+
- S-Editions (2023). *French Medical Exam Practice Platform*. [Website](https://s-editions.fr/)
|
| 67 |
+
|
| 68 |
+
|
| 69 |
+
|