appearance
This model is a fine-tuned version of on an unknown dataset. It achieves the following results on the evaluation set:
- Loss: 1.0131
- Accuracy: 0.6805
- F1 Macro: 0.6261
- Precision Macro: 0.6468
- Recall Macro: 0.6250
- Total Tf: [279, 131, 1099, 131]
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 38
- num_epochs: 25
Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 Macro | Precision Macro | Recall Macro | Total Tf |
|---|---|---|---|---|---|---|---|---|
| 1.0882 | 1.0 | 39 | 1.0748 | 0.6122 | 0.4014 | 0.3585 | 0.5 | [251, 159, 1071, 159] |
| 1.0566 | 2.0 | 78 | 1.0612 | 0.6122 | 0.4014 | 0.3585 | 0.5 | [251, 159, 1071, 159] |
| 1.0157 | 3.0 | 117 | 1.0543 | 0.6171 | 0.4700 | 0.4692 | 0.5254 | [253, 157, 1073, 157] |
| 0.9566 | 4.0 | 156 | 1.0220 | 0.6585 | 0.5851 | 0.6129 | 0.5931 | [270, 140, 1090, 140] |
| 0.8942 | 5.0 | 195 | 1.0177 | 0.6707 | 0.6268 | 0.6288 | 0.6265 | [275, 135, 1095, 135] |
| 0.8334 | 6.0 | 234 | 1.0868 | 0.5902 | 0.5460 | 0.5782 | 0.5687 | [242, 168, 1062, 168] |
| 0.7717 | 7.0 | 273 | 1.0260 | 0.6585 | 0.5920 | 0.6165 | 0.5945 | [270, 140, 1090, 140] |
| 0.8031 | 8.0 | 312 | 1.0290 | 0.6585 | 0.5821 | 0.6298 | 0.5865 | [270, 140, 1090, 140] |
| 0.7367 | 9.0 | 351 | 1.0135 | 0.6732 | 0.6175 | 0.6326 | 0.6166 | [276, 134, 1096, 134] |
| 0.7453 | 10.0 | 390 | 1.0400 | 0.6439 | 0.5868 | 0.6096 | 0.5929 | [264, 146, 1084, 146] |
| 0.7362 | 11.0 | 429 | 1.0152 | 0.6707 | 0.5985 | 0.6256 | 0.6053 | [275, 135, 1095, 135] |
| 0.6926 | 12.0 | 468 | 1.0143 | 0.6805 | 0.6156 | 0.6429 | 0.6179 | [279, 131, 1099, 131] |
| 0.6821 | 13.0 | 507 | 1.0325 | 0.6561 | 0.6133 | 0.6199 | 0.6160 | [269, 141, 1089, 141] |
| 0.6613 | 14.0 | 546 | 1.0184 | 0.6683 | 0.5984 | 0.6287 | 0.6036 | [274, 136, 1094, 136] |
| 0.6479 | 15.0 | 585 | 1.0198 | 0.6659 | 0.6176 | 0.6272 | 0.6158 | [273, 137, 1093, 137] |
| 0.6612 | 16.0 | 624 | 1.0137 | 0.6780 | 0.6191 | 0.6387 | 0.6194 | [278, 132, 1098, 132] |
| 0.6382 | 17.0 | 663 | 1.0194 | 0.6732 | 0.6107 | 0.6364 | 0.6126 | [276, 134, 1096, 134] |
| 0.6392 | 18.0 | 702 | 1.0085 | 0.6805 | 0.6288 | 0.6438 | 0.6272 | [279, 131, 1099, 131] |
| 0.6439 | 19.0 | 741 | 1.0100 | 0.6805 | 0.6266 | 0.6446 | 0.6259 | [279, 131, 1099, 131] |
| 0.6198 | 20.0 | 780 | 1.0145 | 0.6780 | 0.6305 | 0.6426 | 0.6309 | [278, 132, 1098, 132] |
| 0.6223 | 21.0 | 819 | 1.0200 | 0.6634 | 0.6079 | 0.6229 | 0.6089 | [272, 138, 1092, 138] |
| 0.6238 | 22.0 | 858 | 1.0049 | 0.6829 | 0.6389 | 0.6479 | 0.6372 | [280, 130, 1100, 130] |
| 0.6317 | 23.0 | 897 | 1.0042 | 0.6878 | 0.6410 | 0.6539 | 0.6378 | [282, 128, 1102, 128] |
| 0.6089 | 24.0 | 936 | 1.0130 | 0.6829 | 0.6308 | 0.6503 | 0.6292 | [280, 130, 1100, 130] |
| 0.6203 | 25.0 | 975 | 1.0131 | 0.6805 | 0.6261 | 0.6468 | 0.6250 | [279, 131, 1099, 131] |
Framework versions
- Transformers 4.44.2
- Pytorch 2.4.1+cu121
- Datasets 3.2.0
- Tokenizers 0.19.1
- Downloads last month
- -
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support