update model card README.md
Browse files
README.md
CHANGED
|
@@ -4,19 +4,21 @@ tags:
|
|
| 4 |
metrics:
|
| 5 |
- accuracy
|
| 6 |
model-index:
|
| 7 |
-
- name: barec-finetune-
|
| 8 |
results: []
|
| 9 |
---
|
| 10 |
|
| 11 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
| 12 |
should probably proofread and complete it, then remove this comment. -->
|
| 13 |
|
| 14 |
-
# barec-finetune-
|
| 15 |
|
| 16 |
-
This model
|
| 17 |
It achieves the following results on the evaluation set:
|
| 18 |
-
- Loss: 6.
|
| 19 |
-
- Accuracy: 0.
|
|
|
|
|
|
|
| 20 |
|
| 21 |
## Model description
|
| 22 |
|
|
@@ -35,9 +37,9 @@ More information needed
|
|
| 35 |
### Training hyperparameters
|
| 36 |
|
| 37 |
The following hyperparameters were used during training:
|
| 38 |
-
- learning_rate: 5e-05
|
| 39 |
-
- train_batch_size:
|
| 40 |
-
- eval_batch_size:
|
| 41 |
- seed: 42
|
| 42 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 43 |
- lr_scheduler_type: linear
|
|
@@ -46,50 +48,58 @@ The following hyperparameters were used during training:
|
|
| 46 |
|
| 47 |
### Training results
|
| 48 |
|
| 49 |
-
| Training Loss | Epoch | Step |
|
| 50 |
-
|
| 51 |
-
| 1.
|
| 52 |
-
| 1.
|
| 53 |
-
| 1.
|
| 54 |
-
| 1.
|
| 55 |
-
| 1.
|
| 56 |
-
| 1.
|
| 57 |
-
| 0.
|
| 58 |
-
| 0.
|
| 59 |
-
| 0.
|
| 60 |
-
| 0.
|
| 61 |
-
| 0.
|
| 62 |
-
| 0.
|
| 63 |
-
| 0.
|
| 64 |
-
| 0.
|
| 65 |
-
| 0.
|
| 66 |
-
| 0.
|
| 67 |
-
| 0.
|
| 68 |
-
| 0.
|
| 69 |
-
| 0.
|
| 70 |
-
| 0.
|
| 71 |
-
| 0.
|
| 72 |
-
| 0.
|
| 73 |
-
| 0.
|
| 74 |
-
| 0.
|
| 75 |
-
| 0.
|
| 76 |
-
| 0.
|
| 77 |
-
| 0.
|
| 78 |
-
| 0.
|
| 79 |
-
| 0.
|
| 80 |
-
| 0.
|
| 81 |
-
| 0.
|
| 82 |
-
| 0.
|
| 83 |
-
| 0.
|
| 84 |
-
| 0.
|
| 85 |
-
| 0.
|
| 86 |
-
| 0.
|
| 87 |
-
| 0.
|
| 88 |
-
| 0.
|
| 89 |
-
| 0.
|
| 90 |
-
| 0.
|
| 91 |
-
| 0.
|
| 92 |
-
| 0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 93 |
|
| 94 |
|
| 95 |
### Framework versions
|
|
|
|
| 4 |
metrics:
|
| 5 |
- accuracy
|
| 6 |
model-index:
|
| 7 |
+
- name: barec-base-finetune-sent
|
| 8 |
results: []
|
| 9 |
---
|
| 10 |
|
| 11 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
| 12 |
should probably proofread and complete it, then remove this comment. -->
|
| 13 |
|
| 14 |
+
# barec-base-finetune-sent
|
| 15 |
|
| 16 |
+
This model is a fine-tuned version of [bensapir/pixel-barec-pretrain](https://huggingface.co/bensapir/pixel-barec-pretrain) on an unknown dataset.
|
| 17 |
It achieves the following results on the evaluation set:
|
| 18 |
+
- Loss: 6.5546
|
| 19 |
+
- Accuracy: 0.3538
|
| 20 |
+
- Qwk: 0.6191
|
| 21 |
+
- Mae: 1.9391
|
| 22 |
|
| 23 |
## Model description
|
| 24 |
|
|
|
|
| 37 |
### Training hyperparameters
|
| 38 |
|
| 39 |
The following hyperparameters were used during training:
|
| 40 |
+
- learning_rate: 2.5e-05
|
| 41 |
+
- train_batch_size: 64
|
| 42 |
+
- eval_batch_size: 16
|
| 43 |
- seed: 42
|
| 44 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 45 |
- lr_scheduler_type: linear
|
|
|
|
| 48 |
|
| 49 |
### Training results
|
| 50 |
|
| 51 |
+
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Qwk | Mae |
|
| 52 |
+
|:-------------:|:-----:|:-----:|:---------------:|:--------:|:------:|:------:|
|
| 53 |
+
| 1.9165 | 1.17 | 1000 | 1.9485 | 0.3074 | 0.6338 | 1.9832 |
|
| 54 |
+
| 1.7575 | 2.33 | 2000 | 1.9038 | 0.3438 | 0.6283 | 1.8642 |
|
| 55 |
+
| 1.5789 | 3.5 | 3000 | 1.8731 | 0.3595 | 0.6409 | 1.9171 |
|
| 56 |
+
| 1.4313 | 4.67 | 4000 | 1.9007 | 0.3622 | 0.6230 | 1.8940 |
|
| 57 |
+
| 1.2578 | 5.83 | 5000 | 2.0439 | 0.3674 | 0.6296 | 1.9421 |
|
| 58 |
+
| 1.086 | 7.0 | 6000 | 2.2058 | 0.3584 | 0.6337 | 1.9475 |
|
| 59 |
+
| 0.7325 | 8.17 | 7000 | 2.6200 | 0.3566 | 0.6300 | 1.9146 |
|
| 60 |
+
| 0.5493 | 9.33 | 8000 | 2.8226 | 0.3495 | 0.6096 | 1.9665 |
|
| 61 |
+
| 0.4671 | 10.5 | 9000 | 3.1466 | 0.3368 | 0.6115 | 1.9643 |
|
| 62 |
+
| 0.3892 | 11.67 | 10000 | 3.3661 | 0.3446 | 0.6094 | 1.9643 |
|
| 63 |
+
| 0.3393 | 12.84 | 11000 | 3.4706 | 0.3183 | 0.5977 | 2.0435 |
|
| 64 |
+
| 0.297 | 14.0 | 12000 | 3.5575 | 0.3453 | 0.6022 | 1.9891 |
|
| 65 |
+
| 0.1983 | 15.17 | 13000 | 3.9345 | 0.3371 | 0.6095 | 2.0380 |
|
| 66 |
+
| 0.1615 | 16.34 | 14000 | 4.0767 | 0.3509 | 0.6155 | 1.9029 |
|
| 67 |
+
| 0.1622 | 17.5 | 15000 | 4.0375 | 0.3446 | 0.6190 | 1.9390 |
|
| 68 |
+
| 0.1366 | 18.67 | 16000 | 4.1847 | 0.3427 | 0.6183 | 1.9622 |
|
| 69 |
+
| 0.126 | 19.84 | 17000 | 4.2534 | 0.3487 | 0.6212 | 1.9285 |
|
| 70 |
+
| 0.1198 | 21.0 | 18000 | 4.4174 | 0.3413 | 0.6080 | 1.9911 |
|
| 71 |
+
| 0.0885 | 22.17 | 19000 | 4.6410 | 0.3339 | 0.6149 | 2.0514 |
|
| 72 |
+
| 0.0818 | 23.34 | 20000 | 4.6826 | 0.3409 | 0.6079 | 1.9544 |
|
| 73 |
+
| 0.0792 | 24.5 | 21000 | 4.8105 | 0.3477 | 0.6197 | 1.9428 |
|
| 74 |
+
| 0.0811 | 25.67 | 22000 | 4.8232 | 0.3438 | 0.6069 | 1.9865 |
|
| 75 |
+
| 0.0761 | 26.84 | 23000 | 4.8745 | 0.3389 | 0.6183 | 1.9248 |
|
| 76 |
+
| 0.0704 | 28.0 | 24000 | 4.8785 | 0.3518 | 0.5999 | 1.9681 |
|
| 77 |
+
| 0.0603 | 29.17 | 25000 | 5.2340 | 0.3394 | 0.6196 | 1.9793 |
|
| 78 |
+
| 0.0513 | 30.34 | 26000 | 5.3277 | 0.3383 | 0.6078 | 1.9665 |
|
| 79 |
+
| 0.0521 | 31.51 | 27000 | 5.3942 | 0.3360 | 0.6033 | 2.0167 |
|
| 80 |
+
| 0.0518 | 32.67 | 28000 | 5.4359 | 0.3356 | 0.6072 | 2.0219 |
|
| 81 |
+
| 0.0491 | 33.84 | 29000 | 5.5302 | 0.3398 | 0.6109 | 1.9944 |
|
| 82 |
+
| 0.0407 | 35.01 | 30000 | 5.6421 | 0.3408 | 0.6118 | 1.9773 |
|
| 83 |
+
| 0.033 | 36.17 | 31000 | 5.6559 | 0.3568 | 0.6133 | 1.9461 |
|
| 84 |
+
| 0.0355 | 37.34 | 32000 | 5.8496 | 0.3458 | 0.6159 | 1.9594 |
|
| 85 |
+
| 0.0381 | 38.51 | 33000 | 5.9810 | 0.3457 | 0.5981 | 2.0120 |
|
| 86 |
+
| 0.0335 | 39.67 | 34000 | 6.0368 | 0.3465 | 0.5942 | 2.0148 |
|
| 87 |
+
| 0.0282 | 40.84 | 35000 | 6.1605 | 0.3423 | 0.5993 | 2.0044 |
|
| 88 |
+
| 0.0262 | 42.01 | 36000 | 6.2392 | 0.3468 | 0.5993 | 1.9595 |
|
| 89 |
+
| 0.0177 | 43.17 | 37000 | 6.3454 | 0.3456 | 0.6096 | 1.9534 |
|
| 90 |
+
| 0.022 | 44.34 | 38000 | 6.2977 | 0.3492 | 0.6176 | 1.9588 |
|
| 91 |
+
| 0.02 | 45.51 | 39000 | 6.4121 | 0.3431 | 0.5953 | 1.9843 |
|
| 92 |
+
| 0.0178 | 46.67 | 40000 | 6.4294 | 0.3528 | 0.6129 | 1.9550 |
|
| 93 |
+
| 0.0219 | 47.84 | 41000 | 6.5038 | 0.3506 | 0.6039 | 1.9653 |
|
| 94 |
+
| 0.0172 | 49.01 | 42000 | 6.5427 | 0.3454 | 0.6152 | 1.9949 |
|
| 95 |
+
| 0.0169 | 50.18 | 43000 | 6.5222 | 0.3505 | 0.6076 | 1.9714 |
|
| 96 |
+
| 0.0132 | 51.34 | 44000 | 6.4575 | 0.3568 | 0.6194 | 1.9421 |
|
| 97 |
+
| 0.0127 | 52.51 | 45000 | 6.5038 | 0.3613 | 0.5985 | 1.9590 |
|
| 98 |
+
| 0.0131 | 53.68 | 46000 | 6.5518 | 0.3466 | 0.6074 | 1.9647 |
|
| 99 |
+
| 0.0113 | 54.84 | 47000 | 6.5016 | 0.3535 | 0.6187 | 1.9436 |
|
| 100 |
+
| 0.0131 | 56.01 | 48000 | 6.5359 | 0.3505 | 0.6172 | 1.9465 |
|
| 101 |
+
| 0.0112 | 57.18 | 49000 | 6.5665 | 0.3542 | 0.6195 | 1.9379 |
|
| 102 |
+
| 0.0116 | 58.34 | 50000 | 6.5546 | 0.3538 | 0.6191 | 1.9391 |
|
| 103 |
|
| 104 |
|
| 105 |
### Framework versions
|