gs-Logion

This model is a fine-tuned version of cabrooks/LOGION-50k_wordpiece on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 4.7435
  • Bertscore Precision Top1: 69.4231
  • Bertscore Recall Top1: 73.3247
  • Bertscore F1 Top1: 71.2742
  • Bertscore Precision Top1 Mean: 69.4231
  • Bertscore Recall Top1 Mean: 73.3247
  • Bertscore F1 Top1 Mean: 71.2742
  • Bertscore Precision Top3: 72.4567
  • Bertscore Recall Top3: 75.8122
  • Bertscore F1 Top3: 73.9344
  • Bertscore Precision Top3 Mean: 69.3369
  • Bertscore Recall Top3 Mean: 73.4405
  • Bertscore F1 Top3 Mean: 71.2808
  • Bertscore Precision Top5: 73.9557
  • Bertscore Recall Top5: 76.7091
  • Bertscore F1 Top5: 75.1040
  • Bertscore Precision Top5 Mean: 69.3624
  • Bertscore Recall Top5 Mean: 73.4929
  • Bertscore F1 Top5 Mean: 71.3172
  • Bertscore Precision Top10: 75.3599
  • Bertscore Recall Top10: 77.8343
  • Bertscore F1 Top10: 76.3667
  • Bertscore Precision Top10 Mean: 69.1599
  • Bertscore Recall Top10 Mean: 73.4324
  • Bertscore F1 Top10 Mean: 71.1816

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 128
  • eval_batch_size: 64
  • seed: 42
  • optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 0.06
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Bertscore Precision Top1 Bertscore Recall Top1 Bertscore F1 Top1 Bertscore Precision Top1 Mean Bertscore Recall Top1 Mean Bertscore F1 Top1 Mean Bertscore Precision Top3 Bertscore Recall Top3 Bertscore F1 Top3 Bertscore Precision Top3 Mean Bertscore Recall Top3 Mean Bertscore F1 Top3 Mean Bertscore Precision Top5 Bertscore Recall Top5 Bertscore F1 Top5 Bertscore Precision Top5 Mean Bertscore Recall Top5 Mean Bertscore F1 Top5 Mean Bertscore Precision Top10 Bertscore Recall Top10 Bertscore F1 Top10 Bertscore Precision Top10 Mean Bertscore Recall Top10 Mean Bertscore F1 Top10 Mean
No log 1.0 606 4.9901 68.7520 72.9461 70.7337 68.7520 72.9461 70.7337 72.1198 75.4872 73.6212 68.8316 73.0485 70.8266 73.1629 76.2283 74.4758 68.6738 72.9180 70.6819 75.1303 77.6449 76.1507 68.7304 72.9866 70.7462
No log 2.0 1212 4.8858 69.2416 73.3064 71.1667 69.2416 73.3064 71.1667 72.4188 75.6314 73.8351 69.2952 73.4484 71.2634 73.5839 76.4682 74.8175 69.2015 73.3693 71.1773 75.2007 77.7680 76.2653 69.0913 73.3655 71.1144
No log 3.0 1818 4.8845 69.0534 73.1345 70.9860 69.0534 73.1345 70.9860 72.2111 75.4684 73.6579 69.0691 73.3228 71.0816 73.3766 76.4212 74.6816 69.0125 73.3447 71.0603 75.1277 77.7571 76.2173 68.9938 73.3990 71.0762
No log 4.0 2424 4.8869 69.3035 73.2836 71.1877 69.3035 73.2836 71.1877 72.4419 75.6757 73.8648 69.3172 73.4788 71.2876 73.6316 76.5581 74.8813 69.1884 73.4034 71.1839 75.1194 77.7941 76.1902 69.0444 73.3921 71.1014
No log 5.0 3030 4.8059 69.4452 73.4349 71.3335 69.4452 73.4349 71.3335 72.6238 75.8750 74.0494 69.4236 73.5510 71.3804 73.8090 76.7330 75.0386 69.3161 73.4983 71.2958 75.3875 77.8775 76.4057 69.1467 73.4308 71.1751
No log 6.0 3636 4.8282 69.3672 73.3103 71.2332 69.3672 73.3103 71.2332 72.5659 75.7743 73.9856 69.3733 73.4869 71.3226 73.8688 76.7862 75.1133 69.2316 73.4670 71.2346 75.3331 77.9016 76.3904 69.0741 73.3896 71.1164
No log 7.0 4242 4.7246 69.3667 73.3375 71.2496 69.3667 73.3375 71.2496 72.5054 75.7440 73.9124 69.4398 73.5218 71.3728 73.6991 76.6942 74.9626 69.2684 73.4909 71.2654 75.1182 77.8364 76.2124 69.1234 73.4519 71.1719
No log 8.0 4848 4.7093 69.4197 73.3537 71.2865 69.4197 73.3537 71.2865 72.5558 75.8420 73.9847 69.3971 73.5186 71.3496 73.9247 76.7740 75.1221 69.2820 73.4882 71.2719 75.2892 77.8283 76.3209 69.1216 73.4226 71.1562
3.4957 9.0 5454 4.8471 69.4257 73.2362 71.2324 69.4257 73.2362 71.2324 72.3364 75.7030 73.8146 69.3297 73.4323 71.2719 73.9476 76.6215 75.0459 69.3692 73.4637 71.3062 75.4202 77.8834 76.4114 69.2006 73.4434 71.2081
3.4957 10.0 6060 4.6768 69.4231 73.3247 71.2742 69.4231 73.3247 71.2742 72.4567 75.8122 73.9344 69.3369 73.4405 71.2808 73.9557 76.7091 75.1040 69.3624 73.4929 71.3172 75.3599 77.8343 76.3667 69.1599 73.4324 71.1816

Framework versions

  • Transformers 5.8.0
  • Pytorch 2.11.0+cu126
  • Datasets 3.6.0
  • Tokenizers 0.22.2
Downloads last month
32
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for CNR-ILC/gs-Logion

Finetuned
(1)
this model