EricOnyame commited on
Commit
6f08bc3
·
verified ·
1 Parent(s): 0be6a46

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -25,21 +25,21 @@ language:
25
  - vi
26
  - yo
27
  base_model:
28
- - Qwen/Qwen2.5-7B-Instruct
29
  pipeline_tag: text-generation
30
  ---
31
 
32
  # Model Card for Model ID
33
 
34
- CURE-MED-3B is a 7 billion parameter large language model specialized for multilingual medical reasoning, fine-tuned from Qwen/Qwen2.5-7B using a
35
  curriculum-informed reinforcement learning framework to enhance logical correctness and language stability in healthcare applications.
36
 
37
 
38
 
39
  ## Model Details
40
 
41
- CURE-MED-7B is part of the CURE-MED family of models, designed to address the challenges of multilingual medical reasoning in large language models (LLMs).
42
- Built on the Qwen2.5-7B base model, it incorporates a curriculum-informed reinforcement learning approach that integrates code-switching-aware supervised fine-tuning (SFT)
43
  and Group Relative Policy Optimization (GRPO) to improve performance on open-ended medical queries across 13 languages, including underrepresented ones such as Amharic, Yoruba, and Swahili.
44
  The model is trained and evaluated using CUREMED-BENCH, a high-quality multilingual open-ended medical reasoning benchmark with single verifiable answers.
45
 
 
25
  - vi
26
  - yo
27
  base_model:
28
+ - Qwen/Qwen2.5-3B-Instruct
29
  pipeline_tag: text-generation
30
  ---
31
 
32
  # Model Card for Model ID
33
 
34
+ CURE-MED-3B is a 3 billion parameter large language model specialized for multilingual medical reasoning, fine-tuned from Qwen/Qwen2.5-7B using a
35
  curriculum-informed reinforcement learning framework to enhance logical correctness and language stability in healthcare applications.
36
 
37
 
38
 
39
  ## Model Details
40
 
41
+ CURE-MED-3B is part of the CURE-MED family of models, designed to address the challenges of multilingual medical reasoning in large language models (LLMs).
42
+ Built on the Qwen2.5-3B-instruct model, it incorporates a curriculum-informed reinforcement learning approach that integrates code-switching-aware supervised fine-tuning (SFT)
43
  and Group Relative Policy Optimization (GRPO) to improve performance on open-ended medical queries across 13 languages, including underrepresented ones such as Amharic, Yoruba, and Swahili.
44
  The model is trained and evaluated using CUREMED-BENCH, a high-quality multilingual open-ended medical reasoning benchmark with single verifiable answers.
45