rlcc-new-taste-class-weight-absa-max
This model is a fine-tuned version of on an unknown dataset. It achieves the following results on the evaluation set:
- Loss: 1.7481
- Accuracy: 0.5781
- F1 Macro: 0.5865
- Precision Macro: 0.6190
- Recall Macro: 0.5779
- F1 Micro: 0.5781
- Precision Micro: 0.5781
- Recall Micro: 0.5781
- Total Tf: [211, 154, 576, 154]
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 45
- num_epochs: 25
Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 Macro | Precision Macro | Recall Macro | F1 Micro | Precision Micro | Recall Micro | Total Tf |
|---|---|---|---|---|---|---|---|---|---|---|---|
| 1.162 | 1.0 | 46 | 1.0874 | 0.3863 | 0.3082 | 0.2599 | 0.3786 | 0.3863 | 0.3863 | 0.3863 | [141, 224, 506, 224] |
| 1.0057 | 2.0 | 92 | 0.9374 | 0.5534 | 0.4872 | 0.5408 | 0.5454 | 0.5534 | 0.5534 | 0.5534 | [202, 163, 567, 163] |
| 0.8157 | 3.0 | 138 | 0.9043 | 0.5671 | 0.5528 | 0.5525 | 0.5627 | 0.5671 | 0.5671 | 0.5671 | [207, 158, 572, 158] |
| 0.6872 | 4.0 | 184 | 0.9179 | 0.5863 | 0.5739 | 0.5741 | 0.5812 | 0.5863 | 0.5863 | 0.5863 | [214, 151, 579, 151] |
| 0.555 | 5.0 | 230 | 1.0917 | 0.5589 | 0.5516 | 0.5665 | 0.5561 | 0.5589 | 0.5589 | 0.5589 | [204, 161, 569, 161] |
| 0.4627 | 6.0 | 276 | 1.1116 | 0.5781 | 0.5729 | 0.5718 | 0.5742 | 0.5781 | 0.5781 | 0.5781 | [211, 154, 576, 154] |
| 0.356 | 7.0 | 322 | 1.1857 | 0.5753 | 0.5794 | 0.5933 | 0.5733 | 0.5753 | 0.5753 | 0.5753 | [210, 155, 575, 155] |
| 0.3358 | 8.0 | 368 | 1.1866 | 0.5808 | 0.5834 | 0.5922 | 0.5783 | 0.5808 | 0.5808 | 0.5808 | [212, 153, 577, 153] |
| 0.2119 | 9.0 | 414 | 1.3293 | 0.5644 | 0.5719 | 0.5944 | 0.5633 | 0.5644 | 0.5644 | 0.5644 | [206, 159, 571, 159] |
| 0.1631 | 10.0 | 460 | 1.4271 | 0.5726 | 0.5730 | 0.5781 | 0.5696 | 0.5726 | 0.5726 | 0.5726 | [209, 156, 574, 156] |
| 0.1637 | 11.0 | 506 | 1.4893 | 0.6055 | 0.6097 | 0.6277 | 0.6036 | 0.6055 | 0.6055 | 0.6055 | [221, 144, 586, 144] |
| 0.1432 | 12.0 | 552 | 1.5199 | 0.5808 | 0.5820 | 0.5877 | 0.5779 | 0.5808 | 0.5808 | 0.5808 | [212, 153, 577, 153] |
| 0.125 | 13.0 | 598 | 1.5459 | 0.5945 | 0.5999 | 0.6165 | 0.5929 | 0.5945 | 0.5945 | 0.5945 | [217, 148, 582, 148] |
| 0.0873 | 14.0 | 644 | 1.5985 | 0.6055 | 0.6102 | 0.6249 | 0.6038 | 0.6055 | 0.6055 | 0.6055 | [221, 144, 586, 144] |
| 0.0939 | 15.0 | 690 | 1.6536 | 0.5973 | 0.6036 | 0.6225 | 0.5960 | 0.5973 | 0.5973 | 0.5973 | [218, 147, 583, 147] |
| 0.0788 | 16.0 | 736 | 1.7481 | 0.5781 | 0.5865 | 0.6190 | 0.5779 | 0.5781 | 0.5781 | 0.5781 | [211, 154, 576, 154] |
Framework versions
- Transformers 4.52.4
- Pytorch 2.6.0+cu124
- Datasets 3.6.0
- Tokenizers 0.21.2
- Downloads last month
- -
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support