Update README.md
Browse files
README.md
CHANGED
|
@@ -36,42 +36,40 @@ This model is a fine-tuned version of [silma-ai/SILMA-9B-Instruct-v1.0](https://
|
|
| 36 |
- Efficient inference with 4-bit quantization
|
| 37 |
- Fine-tuned using PEFT with LoraConfig for parameter-efficient training
|
| 38 |
|
| 39 |
-
|
| 40 |
-
|
| 41 |
-
|
| 42 |
-
|
| 43 |
-
|
| 44 |
-
|
| 45 |
-
|
| 46 |
-
|
| 47 |
-
|
| 48 |
-
|
| 49 |
-
|
| 50 |
-
|
| 51 |
-
|
| 52 |
-
|
| 53 |
-
|
| 54 |
-
|
| 55 |
-
|
| 56 |
-
|
| 57 |
-
|
| 58 |
-
|
| 59 |
-
|
| 60 |
-
|
| 61 |
-
|
| 62 |
-
|
| 63 |
-
|
| 64 |
-
|
| 65 |
-
|
| 66 |
-
|
| 67 |
-
|
| 68 |
-
|
| 69 |
-
|
| 70 |
-
|
| 71 |
-
|
| 72 |
-
|
| 73 |
-
340 0.061800 0.047798
|
| 74 |
-
350 0.047500 0.045877
|
| 75 |
|
| 76 |
## Usage
|
| 77 |
|
|
@@ -104,7 +102,7 @@ qa_pipeline = pipeline(
|
|
| 104 |
)
|
| 105 |
|
| 106 |
# Example usage
|
| 107 |
-
question = "
|
| 108 |
prompt = f"Question: {question}\nAnswer:"
|
| 109 |
response = qa_pipeline(prompt)[0]['generated_text']
|
| 110 |
|
|
@@ -116,12 +114,7 @@ print(f"Answer: {response}")
|
|
| 116 |
|
| 117 |
Our model demonstrates strong performance on Arabic QA tasks, particularly for questions requiring numerical answers. Here are some key metrics:
|
| 118 |
|
| 119 |
-
- **Eval Loss**: 0.
|
| 120 |
-
- **Eval Loss**: 92% on our test set
|
| 121 |
-
- **F1 Score**: 0.89
|
| 122 |
-
- **Average Response Time**: 150ms
|
| 123 |
-
|
| 124 |
-
(Note: Replace these placeholder metrics with your actual performance data)
|
| 125 |
|
| 126 |
## Limitations
|
| 127 |
|
|
@@ -147,25 +140,10 @@ The model was fine-tuned using the following configuration:
|
|
| 147 |
|
| 148 |
While this model has been trained on a diverse dataset, it may still exhibit biases present in the training data. Users should be aware of potential biases and use the model responsibly.
|
| 149 |
|
| 150 |
-
## Citation
|
| 151 |
-
|
| 152 |
-
If you use this model in your research, please cite:
|
| 153 |
-
|
| 154 |
-
```bibtex
|
| 155 |
-
@misc{silma9b_arabic_qa,
|
| 156 |
-
author = {Mohammed Nasser},
|
| 157 |
-
title = {SILMA-9B-Instruct Fine-Tuned for Arabic QA},
|
| 158 |
-
year = {2024},
|
| 159 |
-
publisher = {GitHub},
|
| 160 |
-
journal = {GitHub repository},
|
| 161 |
-
howpublished = {\url{https://huggingface.co/MohammedNasser/silma_9b_instruct_ft}}
|
| 162 |
-
}
|
| 163 |
-
```
|
| 164 |
-
|
| 165 |
## Contact
|
| 166 |
|
| 167 |
For any questions or feedback, please open an issue on the [GitHub repository](https://github.com/your-github-username/silma-9b-arabic-qa) or contact us at your.email@example.com.
|
| 168 |
|
| 169 |
---
|
| 170 |
|
| 171 |
-
Made with ❤️ by [
|
|
|
|
| 36 |
- Efficient inference with 4-bit quantization
|
| 37 |
- Fine-tuned using PEFT with LoraConfig for parameter-efficient training
|
| 38 |
|
| 39 |
+
### Training results
|
| 40 |
+
|
| 41 |
+
| Training Loss | Epoch | Step | Validation Loss |
|
| 42 |
+
|:-------------:|:-----:|:----:|:---------------:|
|
| 43 |
+
| 2.1356 | 0.04 | 10 | 1.4071 |
|
| 44 |
+
| 0.8079 | 0.08 | 20 | 0.2825 |
|
| 45 |
+
| 0.1592 | 0.12 | 30 | 0.1427 |
|
| 46 |
+
| 0.1202 | 0.16 | 40 | 0.1121 |
|
| 47 |
+
| 0.1095 | 0.2 | 50 | 0.1071 |
|
| 48 |
+
| 0.1024 | 0.24 | 60 | 0.1036 |
|
| 49 |
+
| 0.0993 | 0.28 | 70 | 0.1002 |
|
| 50 |
+
| 0.091 | 0.32 | 80 | 0.0992 |
|
| 51 |
+
| 0.1096 | 0.36 | 90 | 0.0965 |
|
| 52 |
+
| 0.0943 | 0.4 | 100 | 0.0916 |
|
| 53 |
+
| 0.0882 | 0.44 | 110 | 0.0896 |
|
| 54 |
+
| 0.0853 | 0.48 | 120 | 0.0848 |
|
| 55 |
+
| 0.0767 | 0.52 | 130 | 0.0808 |
|
| 56 |
+
| 0.0778 | 0.56 | 140 | 0.0765 |
|
| 57 |
+
| 0.0698 | 0.6 | 150 | 0.0734 |
|
| 58 |
+
| 0.0784 | 0.64 | 160 | 0.0694 |
|
| 59 |
+
| 0.0648 | 0.68 | 170 | 0.0658 |
|
| 60 |
+
| 0.0797 | 0.72 | 180 | 0.0630 |
|
| 61 |
+
| 0.0591 | 0.76 | 190 | 0.0604 |
|
| 62 |
+
| 0.0557 | 0.8 | 200 | 0.0582 |
|
| 63 |
+
| 0.0567 | 0.84 | 210 | 0.0561 |
|
| 64 |
+
| 0.057 | 0.88 | 220 | 0.0534 |
|
| 65 |
+
| 0.0505 | 0.92 | 230 | 0.0515 |
|
| 66 |
+
| 0.0483 | 0.96 | 240 | 0.0482 |
|
| 67 |
+
| 0.0463 | 1.0 | 250 | 0.0463 |
|
| 68 |
+
|
| 69 |
+
|
| 70 |
+
### Training Metrics
|
| 71 |
+
[Training Loss on wandb 🔗](https://wandb.ai/mohnasgbr/huggingface/reports/train-loss-24-09-07-03-41-58---Vmlldzo5MjgxMTY4)
|
| 72 |
+
|
|
|
|
|
|
|
| 73 |
|
| 74 |
## Usage
|
| 75 |
|
|
|
|
| 102 |
)
|
| 103 |
|
| 104 |
# Example usage
|
| 105 |
+
question = "إذا كان لديك ثلاث سيارات، وبعت واحدة منها، كم سيارة ستبقى لديك؟"
|
| 106 |
prompt = f"Question: {question}\nAnswer:"
|
| 107 |
response = qa_pipeline(prompt)[0]['generated_text']
|
| 108 |
|
|
|
|
| 114 |
|
| 115 |
Our model demonstrates strong performance on Arabic QA tasks, particularly for questions requiring numerical answers. Here are some key metrics:
|
| 116 |
|
| 117 |
+
- **Eval Loss**: 0.046
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 118 |
|
| 119 |
## Limitations
|
| 120 |
|
|
|
|
| 140 |
|
| 141 |
While this model has been trained on a diverse dataset, it may still exhibit biases present in the training data. Users should be aware of potential biases and use the model responsibly.
|
| 142 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 143 |
## Contact
|
| 144 |
|
| 145 |
For any questions or feedback, please open an issue on the [GitHub repository](https://github.com/your-github-username/silma-9b-arabic-qa) or contact us at your.email@example.com.
|
| 146 |
|
| 147 |
---
|
| 148 |
|
| 149 |
+
Made with ❤️ by [M. N. Gaer/aiNarabic]
|