roberta-large_MBTI_F

This model is a fine-tuned version of roberta-large on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4707
  • F1: 0.7379
  • Threshold: 0.62
  • F1 At 05: 0.6636

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 1e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • distributed_type: multi-GPU
  • num_devices: 2
  • total_train_batch_size: 32
  • total_eval_batch_size: 32
  • optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 400
  • num_epochs: 5

Training results

Training Loss Epoch Step Validation Loss F1 Threshold F1 At 05
0.5355 0.2018 200 0.4714 0.6356 0.6 0.3552
0.4641 0.4036 400 0.4142 0.6833 0.65 0.3640
0.4016 0.6054 600 0.3815 0.7132 0.61 0.3689
0.3790 0.8073 800 0.3709 0.7220 0.64 0.4053
0.3164 1.0091 1000 0.3741 0.7335 0.59 0.3670
0.3028 1.2109 1200 0.3641 0.7281 0.66 0.3552
0.2624 1.4127 1400 0.3711 0.7298 0.63 0.4856
0.2428 1.6145 1600 0.3751 0.7320 0.63 0.3584
0.2456 1.8163 1800 0.3891 0.7326 0.5700 0.4160
0.2084 2.0182 2000 0.3749 0.7396 0.62 0.4083
0.1631 2.2200 2200 0.4496 0.7359 0.53 0.6828
0.1683 2.4218 2400 0.4192 0.7344 0.6 0.4471
0.1333 2.6236 2600 0.4223 0.7361 0.59 0.6424
0.1594 2.8254 2800 0.4317 0.7399 0.5700 0.5270
0.1244 3.0272 3000 0.4147 0.7359 0.6 0.5211
0.1074 3.2291 3200 0.4412 0.7373 0.64 0.6532
0.1018 3.4309 3400 0.4660 0.7363 0.71 0.5559
0.1089 3.6327 3600 0.4614 0.7329 0.56 0.6998
0.1099 3.8345 3800 0.4304 0.7391 0.61 0.5878
0.1005 4.0363 4000 0.4650 0.7388 0.63 0.5550
0.0829 4.2381 4200 0.4607 0.7375 0.61 0.5115
0.0714 4.4400 4400 0.4740 0.7348 0.6 0.5643
0.0768 4.6418 4600 0.4659 0.7354 0.61 0.5102
0.0884 4.8436 4800 0.4707 0.7379 0.62 0.6636

Framework versions

  • Transformers 5.3.0
  • Pytorch 2.10.0+cu128
  • Datasets 4.6.1
  • Tokenizers 0.22.2
Downloads last month
79
Safetensors
Model size
0.4B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for DrinkIcedT/roberta-large_MBTI_F

Finetuned
(456)
this model