Question Answering
Transformers
Safetensors
Arabic
MohammedNasser commited on
Commit
55ddea5
·
verified ·
1 Parent(s): 92e1520

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +37 -59
README.md CHANGED
@@ -36,42 +36,40 @@ This model is a fine-tuned version of [silma-ai/SILMA-9B-Instruct-v1.0](https://
36
  - Efficient inference with 4-bit quantization
37
  - Fine-tuned using PEFT with LoraConfig for parameter-efficient training
38
 
39
- Step |Training Loss |Validation Loss
40
- 10 2.207200 1.487218
41
- 20 1.021200 0.331835
42
- 30 0.205000 0.138645
43
- 40 0.112900 0.111273
44
- 50 0.113500 0.105615
45
- 60 0.100400 0.104426
46
- 70 0.104500 0.101215
47
- 80 0.099500 0.098990
48
- 90 0.098300 0.098076
49
- 100 0.093900 0.099338
50
- 110 0.100200 0.095021
51
- 120 0.094700 0.093543
52
- 130 0.089800 0.090921
53
- 140 0.087000 0.088527
54
- 150 0.085700 0.085943
55
- 160 0.095600 0.082870
56
- 170 0.079100 0.080233
57
- 180 0.079200 0.077066
58
- 190 0.083400 0.074134
59
- 200 0.087500 0.074748
60
- 210 0.076200 0.069750
61
- 220 0.069400 0.067776
62
- 230 0.075400 0.065210
63
- 240 0.069600 0.063573
64
- 250 0.068100 0.061195
65
- 260 0.067900 0.059264
66
- 270 0.061500 0.057951
67
- 280 0.059900 0.056350
68
- 290 0.056100 0.054536
69
- 300 0.052200 0.052772
70
- 310 0.053400 0.052114
71
- 320 0.049700 0.050167
72
- 330 0.045900 0.048726
73
- 340 0.061800 0.047798
74
- 350 0.047500 0.045877
75
 
76
  ## Usage
77
 
@@ -104,7 +102,7 @@ qa_pipeline = pipeline(
104
  )
105
 
106
  # Example usage
107
- question = "كم عدد الكواكب في المجموعة الشمسية؟"
108
  prompt = f"Question: {question}\nAnswer:"
109
  response = qa_pipeline(prompt)[0]['generated_text']
110
 
@@ -116,12 +114,7 @@ print(f"Answer: {response}")
116
 
117
  Our model demonstrates strong performance on Arabic QA tasks, particularly for questions requiring numerical answers. Here are some key metrics:
118
 
119
- - **Eval Loss**: 0.045
120
- - **Eval Loss**: 92% on our test set
121
- - **F1 Score**: 0.89
122
- - **Average Response Time**: 150ms
123
-
124
- (Note: Replace these placeholder metrics with your actual performance data)
125
 
126
  ## Limitations
127
 
@@ -147,25 +140,10 @@ The model was fine-tuned using the following configuration:
147
 
148
  While this model has been trained on a diverse dataset, it may still exhibit biases present in the training data. Users should be aware of potential biases and use the model responsibly.
149
 
150
- ## Citation
151
-
152
- If you use this model in your research, please cite:
153
-
154
- ```bibtex
155
- @misc{silma9b_arabic_qa,
156
- author = {Mohammed Nasser},
157
- title = {SILMA-9B-Instruct Fine-Tuned for Arabic QA},
158
- year = {2024},
159
- publisher = {GitHub},
160
- journal = {GitHub repository},
161
- howpublished = {\url{https://huggingface.co/MohammedNasser/silma_9b_instruct_ft}}
162
- }
163
- ```
164
-
165
  ## Contact
166
 
167
  For any questions or feedback, please open an issue on the [GitHub repository](https://github.com/your-github-username/silma-9b-arabic-qa) or contact us at your.email@example.com.
168
 
169
  ---
170
 
171
- Made with ❤️ by [Your Name/Organization]
 
36
  - Efficient inference with 4-bit quantization
37
  - Fine-tuned using PEFT with LoraConfig for parameter-efficient training
38
 
39
+ ### Training results
40
+
41
+ | Training Loss | Epoch | Step | Validation Loss |
42
+ |:-------------:|:-----:|:----:|:---------------:|
43
+ | 2.1356 | 0.04 | 10 | 1.4071 |
44
+ | 0.8079 | 0.08 | 20 | 0.2825 |
45
+ | 0.1592 | 0.12 | 30 | 0.1427 |
46
+ | 0.1202 | 0.16 | 40 | 0.1121 |
47
+ | 0.1095 | 0.2 | 50 | 0.1071 |
48
+ | 0.1024 | 0.24 | 60 | 0.1036 |
49
+ | 0.0993 | 0.28 | 70 | 0.1002 |
50
+ | 0.091 | 0.32 | 80 | 0.0992 |
51
+ | 0.1096 | 0.36 | 90 | 0.0965 |
52
+ | 0.0943 | 0.4 | 100 | 0.0916 |
53
+ | 0.0882 | 0.44 | 110 | 0.0896 |
54
+ | 0.0853 | 0.48 | 120 | 0.0848 |
55
+ | 0.0767 | 0.52 | 130 | 0.0808 |
56
+ | 0.0778 | 0.56 | 140 | 0.0765 |
57
+ | 0.0698 | 0.6 | 150 | 0.0734 |
58
+ | 0.0784 | 0.64 | 160 | 0.0694 |
59
+ | 0.0648 | 0.68 | 170 | 0.0658 |
60
+ | 0.0797 | 0.72 | 180 | 0.0630 |
61
+ | 0.0591 | 0.76 | 190 | 0.0604 |
62
+ | 0.0557 | 0.8 | 200 | 0.0582 |
63
+ | 0.0567 | 0.84 | 210 | 0.0561 |
64
+ | 0.057 | 0.88 | 220 | 0.0534 |
65
+ | 0.0505 | 0.92 | 230 | 0.0515 |
66
+ | 0.0483 | 0.96 | 240 | 0.0482 |
67
+ | 0.0463 | 1.0 | 250 | 0.0463 |
68
+
69
+
70
+ ### Training Metrics
71
+ [Training Loss on wandb 🔗](https://wandb.ai/mohnasgbr/huggingface/reports/train-loss-24-09-07-03-41-58---Vmlldzo5MjgxMTY4)
72
+
 
 
73
 
74
  ## Usage
75
 
 
102
  )
103
 
104
  # Example usage
105
+ question = "إذا كان لديك ثلاث سيارات، وبعت واحدة منها، كم سيارة ستبقى لديك؟"
106
  prompt = f"Question: {question}\nAnswer:"
107
  response = qa_pipeline(prompt)[0]['generated_text']
108
 
 
114
 
115
  Our model demonstrates strong performance on Arabic QA tasks, particularly for questions requiring numerical answers. Here are some key metrics:
116
 
117
+ - **Eval Loss**: 0.046
 
 
 
 
 
118
 
119
  ## Limitations
120
 
 
140
 
141
  While this model has been trained on a diverse dataset, it may still exhibit biases present in the training data. Users should be aware of potential biases and use the model responsibly.
142
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
143
  ## Contact
144
 
145
  For any questions or feedback, please open an issue on the [GitHub repository](https://github.com/your-github-username/silma-9b-arabic-qa) or contact us at your.email@example.com.
146
 
147
  ---
148
 
149
+ Made with ❤️ by [M. N. Gaer/aiNarabic]