File size: 2,034 Bytes
17072b9
 
562c2ad
 
 
 
 
 
 
 
 
 
 
17072b9
 
 
 
562c2ad
17072b9
562c2ad
 
17072b9
562c2ad
17072b9
562c2ad
17072b9
562c2ad
 
 
 
 
 
 
 
 
17072b9
562c2ad
17072b9
562c2ad
17072b9
562c2ad
 
 
 
 
 
 
 
17072b9
562c2ad
17072b9
562c2ad
17072b9
562c2ad
17072b9
562c2ad
 
17072b9
562c2ad
17072b9
 
 
562c2ad
17072b9
562c2ad
17072b9
562c2ad
17072b9
562c2ad
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
---
library_name: transformers
tags:
- medical
license: mit
datasets:
- MedInjection-FR/Native
- MedInjection-FR/Translated
language:
- fr
- en
base_model:
- Qwen/Qwen3-4B-Instruct-2507
---



# 🩺 QWEN-4B-TRAD

**QWEN-4B-TRAD** is a fine-tuned version of **Qwen-4B-Instruct** trained on the [MedInjection-FR](https://huggingface.co/MedInjection-FR) dataset, a French biomedical instruction corpus combining *native, synthetic, and translated* medical question–answer pairs.  
This model was fine-tuned using **Supervised Fine-Tuning (SFT)** with **DoRA adapters**, designed to study how the origin of supervision data influences model adaptation.

---

## 🧠 Model overview

| Property | Description |
|-----------|--------------|
| **Base model** | Qwen3-4B-Instruct-2507 |
| **Fine-tuning method** | DoRA (Weight-Decomposed Low-Rank Adaptation) |
| **Architecture size** | ~4B parameters |
| **Language** | French 🇫🇷 |
| **Domain** | Biomedical, Clinical, Health |
| **Intended use** | Research on instruction tuning and domain adaptation |
| **Caution** | Not for clinical or diagnostic use |

---

## ⚙️ Training setup

Fine-tuning was performed on **30k multiple-choice (MCQ and MCQU)** examples for each configuration, using:
- 10 epochs  
- Batch size: 12  
- Learning rate: 1e-4  
- Gradient accumulation: 8  
- Cosine scheduler with 5% warmup  
- LoRA rank: 16, α = 16, dropout = 0.05  
- Adapters applied to: `q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj`

All runs used identical hyperparameters to isolate the effect of **data provenance**.

---

## 📊 Evaluation summary

Evaluation was conducted on French biomedical benchmarks (MCQ, MCQU, OEQ).  
Metrics include **Exact Match (EM)** and **Hamming Score** for multiple-choice tasks, and **BLEU/ROUGE/BERTScore + LLM-as-a-judge** for open-ended QA.  

> See [MedInjection-FR GitHub](https://github.com/yourusername/MedInjection-FR) for full results and plots.



## 📚 Citation

If you use this model, please cite:

```bibtex

```