Upload fine-tuned AraBERT for Saudi EOU detection
Browse files- README.md +83 -0
- adapter_config.json +45 -0
- adapter_model.safetensors +3 -0
- runs/Dec09_04-32-29_ee40c7ab3786/events.out.tfevents.1765254750.ee40c7ab3786.3977.0 +3 -0
- runs/Dec09_04-45-55_ee40c7ab3786/events.out.tfevents.1765255558.ee40c7ab3786.3977.1 +3 -0
- runs/Dec09_05-01-35_ee40c7ab3786/events.out.tfevents.1765256497.ee40c7ab3786.3977.3 +3 -0
- runs/Dec09_05-20-34_ee40c7ab3786/events.out.tfevents.1765257635.ee40c7ab3786.3977.4 +3 -0
- runs/Dec09_05-21-46_ee40c7ab3786/events.out.tfevents.1765257707.ee40c7ab3786.3977.5 +3 -0
- runs/Dec09_05-44-19_ee40c7ab3786/events.out.tfevents.1765259061.ee40c7ab3786.3977.6 +3 -0
- runs/Dec09_06-13-21_ee40c7ab3786/events.out.tfevents.1765260803.ee40c7ab3786.3977.7 +3 -0
- runs/Dec09_06-19-52_ee40c7ab3786/events.out.tfevents.1765261193.ee40c7ab3786.3977.8 +3 -0
- runs/Dec09_06-21-31_ee40c7ab3786/events.out.tfevents.1765261292.ee40c7ab3786.3977.9 +3 -0
- special_tokens_map.json +37 -0
- tokenizer.json +0 -0
- tokenizer_config.json +87 -0
- training_args.bin +3 -0
- vocab.txt +0 -0
README.md
ADDED
|
@@ -0,0 +1,83 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
library_name: peft
|
| 3 |
+
base_model: aubmindlab/bert-base-arabertv02
|
| 4 |
+
tags:
|
| 5 |
+
- base_model:adapter:aubmindlab/bert-base-arabertv02
|
| 6 |
+
- lora
|
| 7 |
+
- transformers
|
| 8 |
+
metrics:
|
| 9 |
+
- accuracy
|
| 10 |
+
model-index:
|
| 11 |
+
- name: bert-eou-classifier_teacher
|
| 12 |
+
results: []
|
| 13 |
+
---
|
| 14 |
+
|
| 15 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
| 16 |
+
should probably proofread and complete it, then remove this comment. -->
|
| 17 |
+
|
| 18 |
+
# bert-eou-classifier_teacher
|
| 19 |
+
|
| 20 |
+
This model is a fine-tuned version of [aubmindlab/bert-base-arabertv02](https://huggingface.co/aubmindlab/bert-base-arabertv02) on an unknown dataset.
|
| 21 |
+
It achieves the following results on the evaluation set:
|
| 22 |
+
- Loss: 1.1555
|
| 23 |
+
- Accuracy: 0.791
|
| 24 |
+
- Auc: 0.865
|
| 25 |
+
|
| 26 |
+
## Model description
|
| 27 |
+
|
| 28 |
+
More information needed
|
| 29 |
+
|
| 30 |
+
## Intended uses & limitations
|
| 31 |
+
|
| 32 |
+
More information needed
|
| 33 |
+
|
| 34 |
+
## Training and evaluation data
|
| 35 |
+
|
| 36 |
+
More information needed
|
| 37 |
+
|
| 38 |
+
## Training procedure
|
| 39 |
+
|
| 40 |
+
### Training hyperparameters
|
| 41 |
+
|
| 42 |
+
The following hyperparameters were used during training:
|
| 43 |
+
- learning_rate: 0.0002
|
| 44 |
+
- train_batch_size: 8
|
| 45 |
+
- eval_batch_size: 8
|
| 46 |
+
- seed: 42
|
| 47 |
+
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
| 48 |
+
- lr_scheduler_type: linear
|
| 49 |
+
- num_epochs: 20
|
| 50 |
+
|
| 51 |
+
### Training results
|
| 52 |
+
|
| 53 |
+
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Auc |
|
| 54 |
+
|:-------------:|:-----:|:-----:|:---------------:|:--------:|:-----:|
|
| 55 |
+
| 0.5875 | 1.0 | 622 | 0.4866 | 0.75 | 0.845 |
|
| 56 |
+
| 0.4703 | 2.0 | 1244 | 0.5337 | 0.76 | 0.855 |
|
| 57 |
+
| 0.4097 | 3.0 | 1866 | 0.5273 | 0.785 | 0.869 |
|
| 58 |
+
| 0.3598 | 4.0 | 2488 | 0.5383 | 0.795 | 0.868 |
|
| 59 |
+
| 0.3278 | 5.0 | 3110 | 0.6127 | 0.803 | 0.878 |
|
| 60 |
+
| 0.3019 | 6.0 | 3732 | 0.6487 | 0.804 | 0.878 |
|
| 61 |
+
| 0.2616 | 7.0 | 4354 | 0.7659 | 0.801 | 0.874 |
|
| 62 |
+
| 0.2451 | 8.0 | 4976 | 0.8012 | 0.793 | 0.871 |
|
| 63 |
+
| 0.2241 | 9.0 | 5598 | 0.8936 | 0.802 | 0.87 |
|
| 64 |
+
| 0.2044 | 10.0 | 6220 | 0.9513 | 0.8 | 0.869 |
|
| 65 |
+
| 0.2015 | 11.0 | 6842 | 0.9689 | 0.802 | 0.869 |
|
| 66 |
+
| 0.1834 | 12.0 | 7464 | 0.9756 | 0.799 | 0.869 |
|
| 67 |
+
| 0.1731 | 13.0 | 8086 | 0.9917 | 0.796 | 0.866 |
|
| 68 |
+
| 0.1455 | 14.0 | 8708 | 1.0958 | 0.794 | 0.863 |
|
| 69 |
+
| 0.1557 | 15.0 | 9330 | 1.0042 | 0.796 | 0.869 |
|
| 70 |
+
| 0.1316 | 16.0 | 9952 | 1.0996 | 0.796 | 0.865 |
|
| 71 |
+
| 0.1335 | 17.0 | 10574 | 1.2024 | 0.794 | 0.863 |
|
| 72 |
+
| 0.1201 | 18.0 | 11196 | 1.1508 | 0.791 | 0.865 |
|
| 73 |
+
| 0.1204 | 19.0 | 11818 | 1.1580 | 0.798 | 0.865 |
|
| 74 |
+
| 0.1137 | 20.0 | 12440 | 1.1555 | 0.791 | 0.865 |
|
| 75 |
+
|
| 76 |
+
|
| 77 |
+
### Framework versions
|
| 78 |
+
|
| 79 |
+
- PEFT 0.18.0
|
| 80 |
+
- Transformers 4.57.3
|
| 81 |
+
- Pytorch 2.5.1+cu124
|
| 82 |
+
- Datasets 3.2.0
|
| 83 |
+
- Tokenizers 0.22.1
|
adapter_config.json
ADDED
|
@@ -0,0 +1,45 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"alora_invocation_tokens": null,
|
| 3 |
+
"alpha_pattern": {},
|
| 4 |
+
"arrow_config": null,
|
| 5 |
+
"auto_mapping": null,
|
| 6 |
+
"base_model_name_or_path": "aubmindlab/bert-base-arabertv02",
|
| 7 |
+
"bias": "none",
|
| 8 |
+
"corda_config": null,
|
| 9 |
+
"ensure_weight_tying": false,
|
| 10 |
+
"eva_config": null,
|
| 11 |
+
"exclude_modules": null,
|
| 12 |
+
"fan_in_fan_out": false,
|
| 13 |
+
"inference_mode": true,
|
| 14 |
+
"init_lora_weights": true,
|
| 15 |
+
"layer_replication": null,
|
| 16 |
+
"layers_pattern": null,
|
| 17 |
+
"layers_to_transform": null,
|
| 18 |
+
"loftq_config": {},
|
| 19 |
+
"lora_alpha": 16,
|
| 20 |
+
"lora_bias": false,
|
| 21 |
+
"lora_dropout": 0.1,
|
| 22 |
+
"megatron_config": null,
|
| 23 |
+
"megatron_core": "megatron.core",
|
| 24 |
+
"modules_to_save": [
|
| 25 |
+
"classifier",
|
| 26 |
+
"score"
|
| 27 |
+
],
|
| 28 |
+
"peft_type": "LORA",
|
| 29 |
+
"peft_version": "0.18.0",
|
| 30 |
+
"qalora_group_size": 16,
|
| 31 |
+
"r": 8,
|
| 32 |
+
"rank_pattern": {},
|
| 33 |
+
"revision": null,
|
| 34 |
+
"target_modules": [
|
| 35 |
+
"query",
|
| 36 |
+
"key",
|
| 37 |
+
"value"
|
| 38 |
+
],
|
| 39 |
+
"target_parameters": null,
|
| 40 |
+
"task_type": "SEQ_CLS",
|
| 41 |
+
"trainable_token_indices": null,
|
| 42 |
+
"use_dora": false,
|
| 43 |
+
"use_qalora": false,
|
| 44 |
+
"use_rslora": false
|
| 45 |
+
}
|
adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1a3884b3841e3a759768e3286cf5b3fff6bfb995975b83513629e77398065107
|
| 3 |
+
size 1785816
|
runs/Dec09_04-32-29_ee40c7ab3786/events.out.tfevents.1765254750.ee40c7ab3786.3977.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bcf1bb7e90234eedf11d760fdd4f673a31bea7c95436794fda610242ea715792
|
| 3 |
+
size 8013
|
runs/Dec09_04-45-55_ee40c7ab3786/events.out.tfevents.1765255558.ee40c7ab3786.3977.1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6f74d68e6ddf8a06db466ae8b571a41387562bf78302c7f4d5965609f4e083b8
|
| 3 |
+
size 7432
|
runs/Dec09_05-01-35_ee40c7ab3786/events.out.tfevents.1765256497.ee40c7ab3786.3977.3
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f4059d4c3955f6b8cbd79cc34a896acdb55e055aab9a5c31c9d330b18f8acc71
|
| 3 |
+
size 9186
|
runs/Dec09_05-20-34_ee40c7ab3786/events.out.tfevents.1765257635.ee40c7ab3786.3977.4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:78eed4d3891fcef462d8b64b1658b3c1b2971c1715c8e7d3a873022b03ba76e7
|
| 3 |
+
size 5171
|
runs/Dec09_05-21-46_ee40c7ab3786/events.out.tfevents.1765257707.ee40c7ab3786.3977.5
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:eb79d1054abb6f4fa38a9666d181982f9b70beeab3659c65980f5822156a217b
|
| 3 |
+
size 5689
|
runs/Dec09_05-44-19_ee40c7ab3786/events.out.tfevents.1765259061.ee40c7ab3786.3977.6
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0d5997853492bb5e27d3dcfa3827566f27f01726f2dbc4b9bfd73a54ddef5414
|
| 3 |
+
size 6851
|
runs/Dec09_06-13-21_ee40c7ab3786/events.out.tfevents.1765260803.ee40c7ab3786.3977.7
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dae7a122af25d28e6a326e0b115e259bdeeb94f1d125b05695f2b7b645ff0537
|
| 3 |
+
size 11272
|
runs/Dec09_06-19-52_ee40c7ab3786/events.out.tfevents.1765261193.ee40c7ab3786.3977.8
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9b3e85ebdd5505fa6891c5d2109ca2bb945e62d5009a1119c9720dbaee334c10
|
| 3 |
+
size 6319
|
runs/Dec09_06-21-31_ee40c7ab3786/events.out.tfevents.1765261292.ee40c7ab3786.3977.9
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b18c718a1787a62d78b2d08e559fa111cd4e74cb2886895d5cea1809b07f7365
|
| 3 |
+
size 17082
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cls_token": {
|
| 3 |
+
"content": "[CLS]",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": false,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"mask_token": {
|
| 10 |
+
"content": "[MASK]",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"pad_token": {
|
| 17 |
+
"content": "[PAD]",
|
| 18 |
+
"lstrip": false,
|
| 19 |
+
"normalized": false,
|
| 20 |
+
"rstrip": false,
|
| 21 |
+
"single_word": false
|
| 22 |
+
},
|
| 23 |
+
"sep_token": {
|
| 24 |
+
"content": "[SEP]",
|
| 25 |
+
"lstrip": false,
|
| 26 |
+
"normalized": false,
|
| 27 |
+
"rstrip": false,
|
| 28 |
+
"single_word": false
|
| 29 |
+
},
|
| 30 |
+
"unk_token": {
|
| 31 |
+
"content": "[UNK]",
|
| 32 |
+
"lstrip": false,
|
| 33 |
+
"normalized": false,
|
| 34 |
+
"rstrip": false,
|
| 35 |
+
"single_word": false
|
| 36 |
+
}
|
| 37 |
+
}
|
tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer_config.json
ADDED
|
@@ -0,0 +1,87 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"added_tokens_decoder": {
|
| 3 |
+
"0": {
|
| 4 |
+
"content": "[PAD]",
|
| 5 |
+
"lstrip": false,
|
| 6 |
+
"normalized": false,
|
| 7 |
+
"rstrip": false,
|
| 8 |
+
"single_word": false,
|
| 9 |
+
"special": true
|
| 10 |
+
},
|
| 11 |
+
"1": {
|
| 12 |
+
"content": "[UNK]",
|
| 13 |
+
"lstrip": false,
|
| 14 |
+
"normalized": false,
|
| 15 |
+
"rstrip": false,
|
| 16 |
+
"single_word": false,
|
| 17 |
+
"special": true
|
| 18 |
+
},
|
| 19 |
+
"2": {
|
| 20 |
+
"content": "[CLS]",
|
| 21 |
+
"lstrip": false,
|
| 22 |
+
"normalized": false,
|
| 23 |
+
"rstrip": false,
|
| 24 |
+
"single_word": false,
|
| 25 |
+
"special": true
|
| 26 |
+
},
|
| 27 |
+
"3": {
|
| 28 |
+
"content": "[SEP]",
|
| 29 |
+
"lstrip": false,
|
| 30 |
+
"normalized": false,
|
| 31 |
+
"rstrip": false,
|
| 32 |
+
"single_word": false,
|
| 33 |
+
"special": true
|
| 34 |
+
},
|
| 35 |
+
"4": {
|
| 36 |
+
"content": "[MASK]",
|
| 37 |
+
"lstrip": false,
|
| 38 |
+
"normalized": false,
|
| 39 |
+
"rstrip": false,
|
| 40 |
+
"single_word": false,
|
| 41 |
+
"special": true
|
| 42 |
+
},
|
| 43 |
+
"5": {
|
| 44 |
+
"content": "[رابط]",
|
| 45 |
+
"lstrip": false,
|
| 46 |
+
"normalized": true,
|
| 47 |
+
"rstrip": false,
|
| 48 |
+
"single_word": true,
|
| 49 |
+
"special": true
|
| 50 |
+
},
|
| 51 |
+
"6": {
|
| 52 |
+
"content": "[بريد]",
|
| 53 |
+
"lstrip": false,
|
| 54 |
+
"normalized": true,
|
| 55 |
+
"rstrip": false,
|
| 56 |
+
"single_word": true,
|
| 57 |
+
"special": true
|
| 58 |
+
},
|
| 59 |
+
"7": {
|
| 60 |
+
"content": "[مستخدم]",
|
| 61 |
+
"lstrip": false,
|
| 62 |
+
"normalized": true,
|
| 63 |
+
"rstrip": false,
|
| 64 |
+
"single_word": true,
|
| 65 |
+
"special": true
|
| 66 |
+
}
|
| 67 |
+
},
|
| 68 |
+
"clean_up_tokenization_spaces": false,
|
| 69 |
+
"cls_token": "[CLS]",
|
| 70 |
+
"do_basic_tokenize": true,
|
| 71 |
+
"do_lower_case": false,
|
| 72 |
+
"extra_special_tokens": {},
|
| 73 |
+
"mask_token": "[MASK]",
|
| 74 |
+
"max_len": 512,
|
| 75 |
+
"model_max_length": 512,
|
| 76 |
+
"never_split": [
|
| 77 |
+
"[بريد]",
|
| 78 |
+
"[مستخدم]",
|
| 79 |
+
"[رابط]"
|
| 80 |
+
],
|
| 81 |
+
"pad_token": "[PAD]",
|
| 82 |
+
"sep_token": "[SEP]",
|
| 83 |
+
"strip_accents": null,
|
| 84 |
+
"tokenize_chinese_chars": true,
|
| 85 |
+
"tokenizer_class": "BertTokenizer",
|
| 86 |
+
"unk_token": "[UNK]"
|
| 87 |
+
}
|
training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:502492914bf5c9271da60a8876cab0250843a3df2c2a9bec59cd72b91be169be
|
| 3 |
+
size 5432
|
vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|