Question Answering
Transformers
Safetensors
Arabic
MohammedNasser commited on
Commit
b98c962
·
verified ·
1 Parent(s): 72cfe87

Delete README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -125
README.md DELETED
@@ -1,125 +0,0 @@
1
- ---
2
- base_model: silma-ai/SILMA-9B-Instruct-v1.0
3
- datasets:
4
- - MohammedNasser/ARabic_Reasoning_QA
5
- language:
6
- - ar
7
- library_name: transformers
8
- license: apache-2.0
9
- metrics:
10
- - accuracy
11
- pipeline_tag: question-answering
12
- ---
13
-
14
-
15
-
16
-
17
- # MohammedNasser/silma_9b_instruct_ft
18
-
19
-
20
- ## Model Description
21
-
22
- The **silma_9b_instruct_ft** is a state-of-the-art language model fine-tuned specifically for Arabic reasoning tasks. This model excels in understanding and processing complex reasoning questions in Arabic, making it suitable for applications that require nuanced comprehension and logical inference. Leveraging advanced transformer architecture and extensive training, this model is designed to handle various reasoning challenges with high accuracy and efficiency.
23
-
24
- ### Key Features
25
- - **Language**: Arabic
26
- - **Primary Task**: Reasoning
27
- - **Architecture**: Transformer-based, fine-tuned for Arabic language
28
- - **Sequence Length**: 20
29
-
30
- ## Dataset
31
-
32
- This model was trained on the ARabic_Reasoning_QA dataset.
33
-
34
- - **Dataset Repository:** [ARabic_Reasoning_QA](https://huggingface.co/datasets/MohammedNasser/ARabic_Reasoning_QA)
35
- - **Dataset Description:** For detailed information about the dataset, please refer to the [README.md of ARabic_Reasoning_QA](https://huggingface.co/datasets/MohammedNasser/ARabic_Reasoning_QA/README.md).
36
-
37
- -
38
- ## Intended Uses & Limitations
39
-
40
- ### Intended Uses
41
- - **Educational Tools**: Assist in creating intelligent tutoring systems and educational applications that require reasoning in Arabic.
42
- - **Research**: Facilitate research in Arabic natural language processing, especially in reasoning and inference tasks.
43
- - **Question Answering Systems**: Improve the accuracy of Arabic-based question-answering systems in various domains.
44
-
45
- ### Limitations
46
- - **Training Data Scope**: Performance is dependent on the diversity and quality of the training data. May not generalize well to highly specialized domains or uncommon dialects.
47
- - **Single Epoch**: Trained for only one epoch; performance may improve with additional training.
48
-
49
- ## Training and Evaluation Data
50
-
51
- ### Training Data
52
- - **Dataset**: The model was fine-tuned using the ARabic Reasoning QA dataset.
53
- - **Content**: The dataset comprises 1000 reasoning questions across various difficulty levels, designed to test logical reasoning in Arabic.
54
- - **Source**: Custom dataset created for reasoning tasks, ensuring diverse and representative examples.
55
-
56
- ### Evaluation Data
57
- - **Datasets Used**: Evaluated on `train`, `eval`, and `test` subsets of the ARabic Reasoning QA dataset.
58
- - **Metrics**: Accuracy and loss were measured to assess performance.
59
- - **Performance**: Achieved an evaluation loss of 0.038
60
-
61
- ## Training Procedure
62
-
63
- ### Preprocessing
64
- - **Tokenization**: Text was tokenized using a pre-trained Arabic tokenizer.
65
- - **Sequence Length**: Text sequences were truncated or padded to a maximum length of 20 tokens.
66
-
67
- ### Training
68
- - **Trainer**: Fine-tuned using the `SFTTrainer` class with the following parameters:
69
- - **Gradient Accumulation**: Steps of 2
70
- - **Learning Rate**: 2e-5
71
- - **Optimizer**: AdamW with 32-bit precision
72
- - **Gradient Clipping**: Max norm of 0.3
73
- - **Warmup Ratio**: 0.03
74
- - **Logging and Saving**: Logs and model checkpoints saved every 10 steps
75
-
76
- ### Evaluation
77
- - **Evaluation Strategy**: Evaluated every 10 steps to monitor model performance on validation and test datasets.
78
- - **Metrics**: Accuracy and loss were recorded to assess the model's reasoning capabilities.
79
-
80
- ---
81
-
82
-
83
-
84
- ### Training results
85
-
86
- | Training Loss | Epoch | Step | Validation Loss |
87
- |:-------------:|:-----:|:----:|:---------------:|
88
- | 2.1356 | 0.04 | 10 | 1.4071 |
89
- | 0.8079 | 0.08 | 20 | 0.2825 |
90
- | 0.1592 | 0.12 | 30 | 0.1427 |
91
- | 0.1202 | 0.16 | 40 | 0.1121 |
92
- | 0.1095 | 0.2 | 50 | 0.1071 |
93
- | 0.1024 | 0.24 | 60 | 0.1036 |
94
- | 0.0993 | 0.28 | 70 | 0.1002 |
95
- | 0.091 | 0.32 | 80 | 0.0992 |
96
- | 0.1096 | 0.36 | 90 | 0.0965 |
97
- | 0.0943 | 0.4 | 100 | 0.0916 |
98
- | 0.0882 | 0.44 | 110 | 0.0896 |
99
- | 0.0853 | 0.48 | 120 | 0.0848 |
100
- | 0.0767 | 0.52 | 130 | 0.0808 |
101
- | 0.0778 | 0.56 | 140 | 0.0765 |
102
- | 0.0698 | 0.6 | 150 | 0.0734 |
103
- | 0.0784 | 0.64 | 160 | 0.0694 |
104
- | 0.0648 | 0.68 | 170 | 0.0658 |
105
- | 0.0797 | 0.72 | 180 | 0.0630 |
106
- | 0.0591 | 0.76 | 190 | 0.0604 |
107
- | 0.0557 | 0.8 | 200 | 0.0582 |
108
- | 0.0567 | 0.84 | 210 | 0.0561 |
109
- | 0.057 | 0.88 | 220 | 0.0534 |
110
- | 0.0505 | 0.92 | 230 | 0.0515 |
111
- | 0.0483 | 0.96 | 240 | 0.0482 |
112
- | 0.0463 | 1.0 | 250 | 0.0463 |
113
-
114
-
115
- ### Training Metrics
116
- [Training Loss on wandb 🔗](https://wandb.ai/mohnasgbr/huggingface/reports/train-loss-24-09-07-03-41-58---Vmlldzo5MjgxMTY4)
117
-
118
-
119
- ### Framework versions
120
-
121
- - PEFT 0.12.1.dev0
122
- - Transformers 4.44.2
123
- - Pytorch 2.4.0+cu121
124
- - Datasets 2.21.0
125
- - Tokenizers 0.19.1