File size: 2,650 Bytes
a52de97
 
0be6a46
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6f08bc3
0be6a46
a52de97
 
 
 
65e1f16
0be6a46
a52de97
 
 
 
 
6f08bc3
 
0be6a46
 
 
a52de97
 
 
 
0be6a46
a52de97
0be6a46
 
 
 
 
 
a52de97
 
0be6a46
a52de97
0be6a46
 
 
a52de97
 
 
 
0be6a46
a52de97
 
0be6a46
 
 
 
 
 
 
 
 
 
a52de97
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
---
library_name: transformers
tags:
- reasoning
- text-generation
- medical-ai
- multilingual-ai
- healthcare
- LLMs
license: apache-2.0
datasets:
- Aikyam-Lab/CUREMED-BENCH
language:
- am
- bn
- fr
- ha
- hi
- ja
- ko
- es
- sw
- th
- tr
- vi
- yo
base_model:
- Qwen/Qwen2.5-3B-Instruct
pipeline_tag: text-generation
---

# Model Card for Model ID

CURE-MED-3B is a 3 billion parameter large language model specialized for multilingual medical reasoning, fine-tuned from Qwen/Qwen2.5-3B-instruct using a
curriculum-informed reinforcement learning framework to enhance logical correctness and language stability in healthcare applications.



## Model Details

CURE-MED-3B is part of the CURE-MED family of models, designed to address the challenges of multilingual medical reasoning in large language models (LLMs).
Built on the Qwen2.5-3B-instruct model, it incorporates a curriculum-informed reinforcement learning approach that integrates code-switching-aware supervised fine-tuning (SFT)
and Group Relative Policy Optimization (GRPO) to improve performance on open-ended medical queries across 13 languages, including underrepresented ones such as Amharic, Yoruba, and Swahili.
The model is trained and evaluated using CUREMED-BENCH, a high-quality multilingual open-ended medical reasoning benchmark with single verifiable answers.

### Model Description

<!-- Provide a longer summary of what this model is. -->

This is the model card of a 🤗 transformers model that has been pushed on the Hub. 

- **Developed by:** Eric Onyame, Akash Ghosh, Subhadip Baidya, Sriparna Saha, Xiuying Chen, Chirag Agarwal (Aikyam Lab and collaborators)
- **Shared by:** Aikyam Lab
- **Model type:** Multilingual medical reasoning large language model
- **Language(s) (NLP):** Amharic, Bengali, French, Hausa, Hindi, Japanese, Korean, Spanish, Swahili, Thai, Turkish, Vietnamese, Yoruba
- **License:** Apache 2.0
- **Finetuned from model:** Qwen2.5-Instruct (1.5B, 3B, 7B, 14B, 32B variants)


### Model Sources

- **Repository:** https://github.com/AikyamLab/cure-med
- **Paper:** https://arxiv.org/abs/2601.13262
- **Demo:** https://cure-med.github.io/




## Citation

**BibTeX:**
```bibtex
@misc{onyame2026curemedcurriculuminformedreinforcementlearning,
  title        = {CURE-Med: Curriculum-Informed Reinforcement Learning for Multilingual Medical Reasoning},
  author       = {Onyame, Eric and Ghosh, Akash and Baidya, Subhadip and Saha, Sriparna and Chen, Xiuying and Agarwal, Chirag},
  year         = {2026},
  eprint       = {2601.13262},
  archivePrefix= {arXiv},
  primaryClass = {cs.AI},
  url          = {https://arxiv.org/abs/2601.13262}
}