Update README.md
Browse files
README.md
CHANGED
|
@@ -1,3 +1,158 @@
|
|
| 1 |
---
|
| 2 |
license: cc-by-4.0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
license: cc-by-4.0
|
| 3 |
+
language:
|
| 4 |
+
- en
|
| 5 |
+
- ja
|
| 6 |
+
- zu
|
| 7 |
+
- yo
|
| 8 |
+
- zh
|
| 9 |
+
- ko
|
| 10 |
+
- th
|
| 11 |
+
- sw
|
| 12 |
+
tags:
|
| 13 |
+
- medical
|
| 14 |
+
size_categories:
|
| 15 |
+
- 1K<n<10K
|
| 16 |
---
|
| 17 |
+
|
| 18 |
+
# MultiMed-X
|
| 19 |
+
|
| 20 |
+
**MultiMed-X** is a multilingual benchmark for **medical reasoning evaluation** across **natural language inference (NLI)** and **open-ended question answering (QA)**.
|
| 21 |
+
The dataset is designed to assess **reasoning quality, factual accuracy, and localization** of large language models in **non-English medical settings**, with particular emphasis on **low-resource languages**.
|
| 22 |
+
|
| 23 |
+
This dataset accompanies the paper: [**MED-COREASONER: Reducing Language Disparities in Medical Reasoning via Language-Informed Co-Reasoning**](https://arxiv.org/pdf/2601.08267).
|
| 24 |
+
|
| 25 |
+
|
| 26 |
+
---
|
| 27 |
+
|
| 28 |
+
## Dataset Overview
|
| 29 |
+
|
| 30 |
+
MultiMed-X-350 is constructed by translating and expert-validating two established English medical benchmarks:
|
| 31 |
+
|
| 32 |
+
- **BioNLI** → Multilingual **medical natural language inference (NLI)**, original data from [BioNLI: Generating a Biomedical NLI Dataset Using Lexico-semantic Constraints for Adversarial Examples](https://arxiv.org/abs/2210.14814).
|
| 33 |
+
- **LiveQA** → Multilingual **open-ended medical question answering (QA)**, original data from [LiveQA: A Question Answering Dataset over Sports Live](https://aclanthology.org/2020.ccl-1.98.pdf).
|
| 34 |
+
|
| 35 |
+
Each instance is translated into multiple target languages and **independently reviewed and revised by bilingual medical experts** to ensure clinical correctness and linguistic naturalness.
|
| 36 |
+
|
| 37 |
+
### Languages
|
| 38 |
+
|
| 39 |
+
The dataset covers **7 non-English languages**:
|
| 40 |
+
|
| 41 |
+
- Chinese (**ZH**)
|
| 42 |
+
- Japanese (**JA**)
|
| 43 |
+
- Korean (**KO**)
|
| 44 |
+
- Swahili (**SW**)
|
| 45 |
+
- Thai (**TH**)
|
| 46 |
+
- Yoruba (**YO**)
|
| 47 |
+
- Zulu (**ZU**)
|
| 48 |
+
|
| 49 |
+
---
|
| 50 |
+
|
| 51 |
+
|
| 52 |
+
## Data Format
|
| 53 |
+
|
| 54 |
+
All data are released as a **single unified table** (e.g., JSONL / Parquet compatible with Hugging Face `datasets`).
|
| 55 |
+
|
| 56 |
+
### Common Fields
|
| 57 |
+
|
| 58 |
+
| Field | Type | Description |
|
| 59 |
+
|--------|--------|-------------|
|
| 60 |
+
| `id` | string | Unique instance ID |
|
| 61 |
+
| `lang` | string | Language code (e.g., `zu`, `sw`) |
|
| 62 |
+
| `task` | string | Task type: `nli` or `qa` |
|
| 63 |
+
| `source` | string | Data source (`BioNLI` or `LiveQA`) |
|
| 64 |
+
| `text` | string | Original content in the target language |
|
| 65 |
+
| `label` | string / null | Gold label (NLI only) |
|
| 66 |
+
|
| 67 |
+
---
|
| 68 |
+
|
| 69 |
+
### ID Convention
|
| 70 |
+
|
| 71 |
+
- **NLI (BioNLI)**
|
| 72 |
+
`bionli-<lang>-XYZ`
|
| 73 |
+
|
| 74 |
+
- **QA (LiveQA)**
|
| 75 |
+
`qa-<lang>-XYZ`
|
| 76 |
+
|
| 77 |
+
Only **3-digit numeric suffixes** are used.
|
| 78 |
+
|
| 79 |
+
---
|
| 80 |
+
|
| 81 |
+
### Example Entries
|
| 82 |
+
|
| 83 |
+
#### NLI Example
|
| 84 |
+
|
| 85 |
+
```json
|
| 86 |
+
{
|
| 87 |
+
"id": "bionli-zu-042",
|
| 88 |
+
"lang": "zu",
|
| 89 |
+
"task": "nli",
|
| 90 |
+
"source": "BioNLI",
|
| 91 |
+
"text": "Premise: ... Hypothesis: ...",
|
| 92 |
+
"label": "entailment"
|
| 93 |
+
}
|
| 94 |
+
```
|
| 95 |
+
|
| 96 |
+
#### QA Example
|
| 97 |
+
|
| 98 |
+
```json
|
| 99 |
+
{
|
| 100 |
+
"id": "qa-sw-117",
|
| 101 |
+
"lang": "sw",
|
| 102 |
+
"task": "qa",
|
| 103 |
+
"source": "LiveQA",
|
| 104 |
+
"text": "Swali: ... Jibu: ...",
|
| 105 |
+
"label": null
|
| 106 |
+
}
|
| 107 |
+
```
|
| 108 |
+
|
| 109 |
+
---
|
| 110 |
+
|
| 111 |
+
## Data Statistics
|
| 112 |
+
|
| 113 |
+
- **350 instances per language**
|
| 114 |
+
- 150 NLI (BioNLI)
|
| 115 |
+
- 200 QA (LiveQA)
|
| 116 |
+
- **~2,450 total instances**
|
| 117 |
+
- Annotated and validated by **~12 physicians or senior medical students**
|
| 118 |
+
|
| 119 |
+
---
|
| 120 |
+
|
| 121 |
+
## Intended Use
|
| 122 |
+
|
| 123 |
+
MultiMed-X-350 is intended for:
|
| 124 |
+
|
| 125 |
+
- Multilingual medical reasoning evaluation
|
| 126 |
+
- Cross-lingual robustness analysis
|
| 127 |
+
- Low-resource language benchmarking
|
| 128 |
+
- Evaluation of reasoning strategies (e.g., CoT, structured reasoning, agentic systems)
|
| 129 |
+
|
| 130 |
+
⚠️ **Not intended for clinical deployment or direct medical decision-making.**
|
| 131 |
+
|
| 132 |
+
---
|
| 133 |
+
|
| 134 |
+
## Ethical Considerations
|
| 135 |
+
|
| 136 |
+
- All data are derived from **publicly available datasets**
|
| 137 |
+
- Translations are **expert-reviewed**
|
| 138 |
+
- No private patient data are included
|
| 139 |
+
- Annotators were formally recruited and compensated or credited as co-authors
|
| 140 |
+
|
| 141 |
+
---
|
| 142 |
+
|
| 143 |
+
## Citation
|
| 144 |
+
|
| 145 |
+
```bibtex
|
| 146 |
+
@article{gao2026medcoreasoner,
|
| 147 |
+
title={MED-COREASONER: Reducing Language Disparities in Medical Reasoning via Language-Informed Co-Reasoning},
|
| 148 |
+
author={Gao, Fan and Tong, Sherry T. and Sohn, Jiwoong and Huang, Jiahao and Jiang, Junfeng and Xia, Ding and Ittichaiwong, Piyalitt and Veerakanjana, Kanyakorn and Kim, Hyunjae and Chen, Qingyu and Taylor, Edison Marrese and Kobayashi, Kazuma and Aizawa, Akiko and Li, Irene},
|
| 149 |
+
journal={arXiv preprint arXiv:2601.08267},
|
| 150 |
+
year={2026}
|
| 151 |
+
}
|
| 152 |
+
```
|
| 153 |
+
|
| 154 |
+
---
|
| 155 |
+
|
| 156 |
+
## License
|
| 157 |
+
|
| 158 |
+
This dataset is released for **research and evaluation purposes only**, under the same licensing terms as the original source datasets (BioNLI, LiveQA).
|