Delete README.md
Browse files
README.md
DELETED
|
@@ -1,125 +0,0 @@
|
|
| 1 |
-
---
|
| 2 |
-
base_model: silma-ai/SILMA-9B-Instruct-v1.0
|
| 3 |
-
datasets:
|
| 4 |
-
- MohammedNasser/ARabic_Reasoning_QA
|
| 5 |
-
language:
|
| 6 |
-
- ar
|
| 7 |
-
library_name: transformers
|
| 8 |
-
license: apache-2.0
|
| 9 |
-
metrics:
|
| 10 |
-
- accuracy
|
| 11 |
-
pipeline_tag: question-answering
|
| 12 |
-
---
|
| 13 |
-
|
| 14 |
-
|
| 15 |
-
|
| 16 |
-
|
| 17 |
-
# MohammedNasser/silma_9b_instruct_ft
|
| 18 |
-
|
| 19 |
-
|
| 20 |
-
## Model Description
|
| 21 |
-
|
| 22 |
-
The **silma_9b_instruct_ft** is a state-of-the-art language model fine-tuned specifically for Arabic reasoning tasks. This model excels in understanding and processing complex reasoning questions in Arabic, making it suitable for applications that require nuanced comprehension and logical inference. Leveraging advanced transformer architecture and extensive training, this model is designed to handle various reasoning challenges with high accuracy and efficiency.
|
| 23 |
-
|
| 24 |
-
### Key Features
|
| 25 |
-
- **Language**: Arabic
|
| 26 |
-
- **Primary Task**: Reasoning
|
| 27 |
-
- **Architecture**: Transformer-based, fine-tuned for Arabic language
|
| 28 |
-
- **Sequence Length**: 20
|
| 29 |
-
|
| 30 |
-
## Dataset
|
| 31 |
-
|
| 32 |
-
This model was trained on the ARabic_Reasoning_QA dataset.
|
| 33 |
-
|
| 34 |
-
- **Dataset Repository:** [ARabic_Reasoning_QA](https://huggingface.co/datasets/MohammedNasser/ARabic_Reasoning_QA)
|
| 35 |
-
- **Dataset Description:** For detailed information about the dataset, please refer to the [README.md of ARabic_Reasoning_QA](https://huggingface.co/datasets/MohammedNasser/ARabic_Reasoning_QA/README.md).
|
| 36 |
-
|
| 37 |
-
-
|
| 38 |
-
## Intended Uses & Limitations
|
| 39 |
-
|
| 40 |
-
### Intended Uses
|
| 41 |
-
- **Educational Tools**: Assist in creating intelligent tutoring systems and educational applications that require reasoning in Arabic.
|
| 42 |
-
- **Research**: Facilitate research in Arabic natural language processing, especially in reasoning and inference tasks.
|
| 43 |
-
- **Question Answering Systems**: Improve the accuracy of Arabic-based question-answering systems in various domains.
|
| 44 |
-
|
| 45 |
-
### Limitations
|
| 46 |
-
- **Training Data Scope**: Performance is dependent on the diversity and quality of the training data. May not generalize well to highly specialized domains or uncommon dialects.
|
| 47 |
-
- **Single Epoch**: Trained for only one epoch; performance may improve with additional training.
|
| 48 |
-
|
| 49 |
-
## Training and Evaluation Data
|
| 50 |
-
|
| 51 |
-
### Training Data
|
| 52 |
-
- **Dataset**: The model was fine-tuned using the ARabic Reasoning QA dataset.
|
| 53 |
-
- **Content**: The dataset comprises 1000 reasoning questions across various difficulty levels, designed to test logical reasoning in Arabic.
|
| 54 |
-
- **Source**: Custom dataset created for reasoning tasks, ensuring diverse and representative examples.
|
| 55 |
-
|
| 56 |
-
### Evaluation Data
|
| 57 |
-
- **Datasets Used**: Evaluated on `train`, `eval`, and `test` subsets of the ARabic Reasoning QA dataset.
|
| 58 |
-
- **Metrics**: Accuracy and loss were measured to assess performance.
|
| 59 |
-
- **Performance**: Achieved an evaluation loss of 0.038
|
| 60 |
-
|
| 61 |
-
## Training Procedure
|
| 62 |
-
|
| 63 |
-
### Preprocessing
|
| 64 |
-
- **Tokenization**: Text was tokenized using a pre-trained Arabic tokenizer.
|
| 65 |
-
- **Sequence Length**: Text sequences were truncated or padded to a maximum length of 20 tokens.
|
| 66 |
-
|
| 67 |
-
### Training
|
| 68 |
-
- **Trainer**: Fine-tuned using the `SFTTrainer` class with the following parameters:
|
| 69 |
-
- **Gradient Accumulation**: Steps of 2
|
| 70 |
-
- **Learning Rate**: 2e-5
|
| 71 |
-
- **Optimizer**: AdamW with 32-bit precision
|
| 72 |
-
- **Gradient Clipping**: Max norm of 0.3
|
| 73 |
-
- **Warmup Ratio**: 0.03
|
| 74 |
-
- **Logging and Saving**: Logs and model checkpoints saved every 10 steps
|
| 75 |
-
|
| 76 |
-
### Evaluation
|
| 77 |
-
- **Evaluation Strategy**: Evaluated every 10 steps to monitor model performance on validation and test datasets.
|
| 78 |
-
- **Metrics**: Accuracy and loss were recorded to assess the model's reasoning capabilities.
|
| 79 |
-
|
| 80 |
-
---
|
| 81 |
-
|
| 82 |
-
|
| 83 |
-
|
| 84 |
-
### Training results
|
| 85 |
-
|
| 86 |
-
| Training Loss | Epoch | Step | Validation Loss |
|
| 87 |
-
|:-------------:|:-----:|:----:|:---------------:|
|
| 88 |
-
| 2.1356 | 0.04 | 10 | 1.4071 |
|
| 89 |
-
| 0.8079 | 0.08 | 20 | 0.2825 |
|
| 90 |
-
| 0.1592 | 0.12 | 30 | 0.1427 |
|
| 91 |
-
| 0.1202 | 0.16 | 40 | 0.1121 |
|
| 92 |
-
| 0.1095 | 0.2 | 50 | 0.1071 |
|
| 93 |
-
| 0.1024 | 0.24 | 60 | 0.1036 |
|
| 94 |
-
| 0.0993 | 0.28 | 70 | 0.1002 |
|
| 95 |
-
| 0.091 | 0.32 | 80 | 0.0992 |
|
| 96 |
-
| 0.1096 | 0.36 | 90 | 0.0965 |
|
| 97 |
-
| 0.0943 | 0.4 | 100 | 0.0916 |
|
| 98 |
-
| 0.0882 | 0.44 | 110 | 0.0896 |
|
| 99 |
-
| 0.0853 | 0.48 | 120 | 0.0848 |
|
| 100 |
-
| 0.0767 | 0.52 | 130 | 0.0808 |
|
| 101 |
-
| 0.0778 | 0.56 | 140 | 0.0765 |
|
| 102 |
-
| 0.0698 | 0.6 | 150 | 0.0734 |
|
| 103 |
-
| 0.0784 | 0.64 | 160 | 0.0694 |
|
| 104 |
-
| 0.0648 | 0.68 | 170 | 0.0658 |
|
| 105 |
-
| 0.0797 | 0.72 | 180 | 0.0630 |
|
| 106 |
-
| 0.0591 | 0.76 | 190 | 0.0604 |
|
| 107 |
-
| 0.0557 | 0.8 | 200 | 0.0582 |
|
| 108 |
-
| 0.0567 | 0.84 | 210 | 0.0561 |
|
| 109 |
-
| 0.057 | 0.88 | 220 | 0.0534 |
|
| 110 |
-
| 0.0505 | 0.92 | 230 | 0.0515 |
|
| 111 |
-
| 0.0483 | 0.96 | 240 | 0.0482 |
|
| 112 |
-
| 0.0463 | 1.0 | 250 | 0.0463 |
|
| 113 |
-
|
| 114 |
-
|
| 115 |
-
### Training Metrics
|
| 116 |
-
[Training Loss on wandb 🔗](https://wandb.ai/mohnasgbr/huggingface/reports/train-loss-24-09-07-03-41-58---Vmlldzo5MjgxMTY4)
|
| 117 |
-
|
| 118 |
-
|
| 119 |
-
### Framework versions
|
| 120 |
-
|
| 121 |
-
- PEFT 0.12.1.dev0
|
| 122 |
-
- Transformers 4.44.2
|
| 123 |
-
- Pytorch 2.4.0+cu121
|
| 124 |
-
- Datasets 2.21.0
|
| 125 |
-
- Tokenizers 0.19.1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|