rlcc-new-taste-class-weight-absa-None
This model is a fine-tuned version of on an unknown dataset. It achieves the following results on the evaluation set:
- Loss: 1.5234
- Accuracy: 0.5918
- F1 Macro: 0.5945
- Precision Macro: 0.6035
- Recall Macro: 0.5896
- F1 Micro: 0.5918
- Precision Micro: 0.5918
- Recall Micro: 0.5918
- Total Tf: [216, 149, 581, 149]
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 45
- num_epochs: 25
Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 Macro | Precision Macro | Recall Macro | F1 Micro | Precision Micro | Recall Micro | Total Tf |
|---|---|---|---|---|---|---|---|---|---|---|---|
| 1.0969 | 1.0 | 46 | 1.0961 | 0.3425 | 0.2272 | 0.5217 | 0.3424 | 0.3425 | 0.3425 | 0.3425 | [125, 240, 490, 240] |
| 0.9759 | 2.0 | 92 | 0.9606 | 0.5315 | 0.5261 | 0.5292 | 0.5287 | 0.5315 | 0.5315 | 0.5315 | [194, 171, 559, 171] |
| 0.8335 | 3.0 | 138 | 0.9300 | 0.5644 | 0.5370 | 0.5513 | 0.5576 | 0.5644 | 0.5644 | 0.5644 | [206, 159, 571, 159] |
| 0.6809 | 4.0 | 184 | 0.9330 | 0.5863 | 0.5745 | 0.5817 | 0.5812 | 0.5863 | 0.5863 | 0.5863 | [214, 151, 579, 151] |
| 0.5874 | 5.0 | 230 | 1.0094 | 0.5781 | 0.5680 | 0.5786 | 0.5750 | 0.5781 | 0.5781 | 0.5781 | [211, 154, 576, 154] |
| 0.4379 | 6.0 | 276 | 1.1100 | 0.5863 | 0.5795 | 0.5791 | 0.5823 | 0.5863 | 0.5863 | 0.5863 | [214, 151, 579, 151] |
| 0.3543 | 7.0 | 322 | 1.1689 | 0.5945 | 0.5951 | 0.6041 | 0.5919 | 0.5945 | 0.5945 | 0.5945 | [217, 148, 582, 148] |
| 0.3305 | 8.0 | 368 | 1.2335 | 0.5808 | 0.5826 | 0.5889 | 0.5787 | 0.5808 | 0.5808 | 0.5808 | [212, 153, 577, 153] |
| 0.2577 | 9.0 | 414 | 1.3390 | 0.5808 | 0.5851 | 0.6031 | 0.5796 | 0.5808 | 0.5808 | 0.5808 | [212, 153, 577, 153] |
| 0.223 | 10.0 | 460 | 1.4179 | 0.5589 | 0.5666 | 0.5881 | 0.5579 | 0.5589 | 0.5589 | 0.5589 | [204, 161, 569, 161] |
| 0.1873 | 11.0 | 506 | 1.4582 | 0.5616 | 0.5652 | 0.5817 | 0.5595 | 0.5616 | 0.5616 | 0.5616 | [205, 160, 570, 160] |
| 0.1449 | 12.0 | 552 | 1.5234 | 0.5918 | 0.5945 | 0.6035 | 0.5896 | 0.5918 | 0.5918 | 0.5918 | [216, 149, 581, 149] |
Framework versions
- Transformers 4.52.4
- Pytorch 2.6.0+cu124
- Datasets 3.6.0
- Tokenizers 0.21.2
- Downloads last month
- 2
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support