Update README.md
Browse files
README.md
CHANGED
|
@@ -8,10 +8,14 @@ base_model:
|
|
| 8 |
- google/gemma-2-2b-it
|
| 9 |
---
|
| 10 |
|
|
|
|
| 11 |
The goal of this project is to adapt large language models for the Arabic language. Due to the scarcity of Arabic instruction fine-tuning data, the focus is on creating a high-quality instruction fine-tuning (IFT) dataset. The project aims to finetune models on this dataset and evaluate their performance across various benchmarks.
|
| 12 |
|
|
|
|
| 13 |
This model is the 2B version. It was trained for 2 days on 1 A100 GPU using LoRA with a rank of 128, a learning rate of 1e-4, and a cosine learning rate schedule.
|
| 14 |
|
|
|
|
|
|
|
| 15 |
| Metric | Slim205/Barka-2b-it |
|
| 16 |
|----------------------|---------------------|
|
| 17 |
| Average | 46.98 |
|
|
|
|
| 8 |
- google/gemma-2-2b-it
|
| 9 |
---
|
| 10 |
|
| 11 |
+
# Motivation :
|
| 12 |
The goal of this project is to adapt large language models for the Arabic language. Due to the scarcity of Arabic instruction fine-tuning data, the focus is on creating a high-quality instruction fine-tuning (IFT) dataset. The project aims to finetune models on this dataset and evaluate their performance across various benchmarks.
|
| 13 |
|
| 14 |
+
# Training :
|
| 15 |
This model is the 2B version. It was trained for 2 days on 1 A100 GPU using LoRA with a rank of 128, a learning rate of 1e-4, and a cosine learning rate schedule.
|
| 16 |
|
| 17 |
+
# Evaluation :
|
| 18 |
+
|
| 19 |
| Metric | Slim205/Barka-2b-it |
|
| 20 |
|----------------------|---------------------|
|
| 21 |
| Average | 46.98 |
|