File size: 2,032 Bytes
3d19824
 
9764996
 
 
 
 
 
 
 
 
 
 
3d19824
 
 
 
9764996
3d19824
9764996
 
3d19824
9764996
3d19824
9764996
3d19824
9764996
 
 
 
 
 
 
 
 
3d19824
9764996
3d19824
9764996
3d19824
9764996
 
 
 
 
 
 
 
3d19824
9764996
3d19824
9764996
3d19824
9764996
3d19824
9764996
 
3d19824
9764996
3d19824
 
 
9764996
3d19824
9764996
3d19824
9764996
3d19824
9764996
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
---
library_name: transformers
tags:
- medical
license: mit
datasets:
- MedInjection-FR/Native
- MedInjection-FR/Translated
language:
- fr
- en
base_model:
- Qwen/Qwen3-4B-Instruct-2507
---



# 🩺 QWEN-4B-NAT

**QWEN-4B-NAT** is a fine-tuned version of **Qwen-4B-Instruct** trained on the [MedInjection-FR](https://huggingface.co/MedInjection-FR) dataset, a French biomedical instruction corpus combining *native, synthetic, and translated* medical question–answer pairs.  
This model was fine-tuned using **Supervised Fine-Tuning (SFT)** with **DoRA adapters**, designed to study how the origin of supervision data influences model adaptation.

---

## 🧠 Model overview

| Property | Description |
|-----------|--------------|
| **Base model** | Qwen3-4B-Instruct-2507 |
| **Fine-tuning method** | DoRA (Weight-Decomposed Low-Rank Adaptation) |
| **Architecture size** | ~4B parameters |
| **Language** | French 🇫🇷 |
| **Domain** | Biomedical, Clinical, Health |
| **Intended use** | Research on instruction tuning and domain adaptation |
| **Caution** | Not for clinical or diagnostic use |

---

## ⚙️ Training setup

Fine-tuning was performed on **30k multiple-choice (MCQ and MCQU)** examples for each configuration, using:
- 10 epochs  
- Batch size: 12  
- Learning rate: 1e-4  
- Gradient accumulation: 8  
- Cosine scheduler with 5% warmup  
- LoRA rank: 16, α = 16, dropout = 0.05  
- Adapters applied to: `q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj`

All runs used identical hyperparameters to isolate the effect of **data provenance**.

---

## 📊 Evaluation summary

Evaluation was conducted on French biomedical benchmarks (MCQ, MCQU, OEQ).  
Metrics include **Exact Match (EM)** and **Hamming Score** for multiple-choice tasks, and **BLEU/ROUGE/BERTScore + LLM-as-a-judge** for open-ended QA.  

> See [MedInjection-FR GitHub](https://github.com/yourusername/MedInjection-FR) for full results and plots.



## 📚 Citation

If you use this model, please cite:

```bibtex

```