rlcc-new-palate-upsample_replacement-absa-None
This model is a fine-tuned version of on an unknown dataset. It achieves the following results on the evaluation set:
- Loss: 1.6556
- Accuracy: 0.5730
- F1 Macro: 0.5804
- Precision Macro: 0.6202
- Recall Macro: 0.5711
- F1 Micro: 0.5730
- Precision Micro: 0.5730
- Recall Micro: 0.5730
- Total Tf: [102, 76, 280, 76]
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 21
- num_epochs: 25
Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 Macro | Precision Macro | Recall Macro | F1 Micro | Precision Micro | Recall Micro | Total Tf |
|---|---|---|---|---|---|---|---|---|---|---|---|
| 1.1158 | 1.0 | 22 | 1.0903 | 0.3820 | 0.3137 | 0.5856 | 0.3706 | 0.3820 | 0.3820 | 0.3820 | [68, 110, 246, 110] |
| 1.0469 | 2.0 | 44 | 1.0469 | 0.4494 | 0.4257 | 0.4347 | 0.4571 | 0.4494 | 0.4494 | 0.4494 | [80, 98, 258, 98] |
| 0.8449 | 3.0 | 66 | 0.9927 | 0.5169 | 0.4983 | 0.5137 | 0.5241 | 0.5169 | 0.5169 | 0.5169 | [92, 86, 270, 86] |
| 0.6818 | 4.0 | 88 | 1.0175 | 0.5225 | 0.5081 | 0.5275 | 0.5193 | 0.5225 | 0.5225 | 0.5225 | [93, 85, 271, 85] |
| 0.5959 | 5.0 | 110 | 1.0224 | 0.5281 | 0.5335 | 0.5438 | 0.5307 | 0.5281 | 0.5281 | 0.5281 | [94, 84, 272, 84] |
| 0.4688 | 6.0 | 132 | 1.0541 | 0.5393 | 0.5407 | 0.5552 | 0.5367 | 0.5393 | 0.5393 | 0.5393 | [96, 82, 274, 82] |
| 0.3176 | 7.0 | 154 | 1.0975 | 0.5618 | 0.5684 | 0.5856 | 0.5618 | 0.5618 | 0.5618 | 0.5618 | [100, 78, 278, 78] |
| 0.2808 | 8.0 | 176 | 1.1576 | 0.5562 | 0.5570 | 0.6056 | 0.5507 | 0.5562 | 0.5562 | 0.5562 | [99, 79, 277, 79] |
| 0.2042 | 9.0 | 198 | 1.1906 | 0.5506 | 0.5581 | 0.5964 | 0.5501 | 0.5506 | 0.5506 | 0.5506 | [98, 80, 276, 80] |
| 0.1776 | 10.0 | 220 | 1.2535 | 0.5787 | 0.5818 | 0.6134 | 0.5749 | 0.5787 | 0.5787 | 0.5787 | [103, 75, 281, 75] |
| 0.158 | 11.0 | 242 | 1.2912 | 0.5449 | 0.5430 | 0.6015 | 0.5384 | 0.5449 | 0.5449 | 0.5449 | [97, 81, 275, 81] |
| 0.1295 | 12.0 | 264 | 1.2984 | 0.5899 | 0.5934 | 0.6266 | 0.5863 | 0.5899 | 0.5899 | 0.5899 | [105, 73, 283, 73] |
| 0.0902 | 13.0 | 286 | 1.3874 | 0.5618 | 0.5681 | 0.5815 | 0.5614 | 0.5618 | 0.5618 | 0.5618 | [100, 78, 278, 78] |
| 0.0972 | 14.0 | 308 | 1.4799 | 0.5787 | 0.5827 | 0.6333 | 0.5757 | 0.5787 | 0.5787 | 0.5787 | [103, 75, 281, 75] |
| 0.0707 | 15.0 | 330 | 1.4690 | 0.5843 | 0.5896 | 0.6204 | 0.5815 | 0.5843 | 0.5843 | 0.5843 | [104, 74, 282, 74] |
| 0.0694 | 16.0 | 352 | 1.5353 | 0.5843 | 0.5904 | 0.6282 | 0.5814 | 0.5843 | 0.5843 | 0.5843 | [104, 74, 282, 74] |
| 0.0529 | 17.0 | 374 | 1.6556 | 0.5730 | 0.5804 | 0.6202 | 0.5711 | 0.5730 | 0.5730 | 0.5730 | [102, 76, 280, 76] |
Framework versions
- Transformers 4.52.4
- Pytorch 2.6.0+cu124
- Datasets 3.6.0
- Tokenizers 0.21.2
- Downloads last month
- -
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support