ArabicNewSplits6_FineTuningAraBERTFreeze_run1_AugV5_k15_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7035
  • Qwk: 0.4606
  • Mse: 0.7035
  • Rmse: 0.8387

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0541 2 6.4115 -0.0230 6.4115 2.5321
No log 0.1081 4 4.3567 -0.0299 4.3567 2.0873
No log 0.1622 6 3.0427 -0.0039 3.0427 1.7443
No log 0.2162 8 2.1635 0.0283 2.1635 1.4709
No log 0.2703 10 1.5615 0.0545 1.5615 1.2496
No log 0.3243 12 1.2503 0.0025 1.2503 1.1182
No log 0.3784 14 1.1634 -0.0252 1.1634 1.0786
No log 0.4324 16 1.0931 -0.0165 1.0931 1.0455
No log 0.4865 18 0.9489 -0.0067 0.9489 0.9741
No log 0.5405 20 0.8926 -0.0190 0.8926 0.9448
No log 0.5946 22 0.8574 0.0098 0.8574 0.9259
No log 0.6486 24 0.8121 0.0761 0.8121 0.9012
No log 0.7027 26 0.7979 0.0835 0.7979 0.8933
No log 0.7568 28 0.7699 0.1382 0.7699 0.8774
No log 0.8108 30 0.7201 0.2253 0.7201 0.8486
No log 0.8649 32 0.7077 0.2413 0.7077 0.8413
No log 0.9189 34 0.7226 0.1667 0.7226 0.8501
No log 0.9730 36 0.7339 0.2333 0.7339 0.8567
No log 1.0270 38 0.7329 0.2405 0.7329 0.8561
No log 1.0811 40 0.7151 0.2445 0.7151 0.8457
No log 1.1351 42 0.7027 0.2658 0.7027 0.8383
No log 1.1892 44 0.6759 0.3083 0.6759 0.8221
No log 1.2432 46 0.6696 0.3369 0.6696 0.8183
No log 1.2973 48 0.6547 0.3354 0.6547 0.8092
No log 1.3514 50 0.6485 0.3324 0.6485 0.8053
No log 1.4054 52 0.6922 0.3676 0.6922 0.8320
No log 1.4595 54 0.7792 0.3205 0.7792 0.8827
No log 1.5135 56 0.8405 0.3207 0.8405 0.9168
No log 1.5676 58 0.9099 0.3009 0.9099 0.9539
No log 1.6216 60 0.8903 0.3009 0.8903 0.9436
No log 1.6757 62 0.8059 0.3326 0.8059 0.8977
No log 1.7297 64 0.7128 0.3537 0.7128 0.8443
No log 1.7838 66 0.6782 0.3685 0.6782 0.8236
No log 1.8378 68 0.7101 0.3372 0.7101 0.8427
No log 1.8919 70 0.7591 0.3525 0.7591 0.8713
No log 1.9459 72 0.7103 0.3530 0.7103 0.8428
No log 2.0 74 0.6128 0.4731 0.6128 0.7828
No log 2.0541 76 0.5603 0.4556 0.5603 0.7485
No log 2.1081 78 0.5518 0.4476 0.5518 0.7428
No log 2.1622 80 0.5452 0.4732 0.5452 0.7384
No log 2.2162 82 0.5476 0.4393 0.5476 0.7400
No log 2.2703 84 0.5688 0.4420 0.5688 0.7542
No log 2.3243 86 0.5916 0.4533 0.5916 0.7692
No log 2.3784 88 0.6035 0.4476 0.6035 0.7768
No log 2.4324 90 0.5857 0.4422 0.5857 0.7653
No log 2.4865 92 0.5934 0.4465 0.5934 0.7703
No log 2.5405 94 0.6169 0.5004 0.6169 0.7854
No log 2.5946 96 0.6243 0.5055 0.6243 0.7902
No log 2.6486 98 0.6347 0.5004 0.6347 0.7967
No log 2.7027 100 0.6544 0.4646 0.6544 0.8089
No log 2.7568 102 0.6277 0.4612 0.6277 0.7923
No log 2.8108 104 0.6009 0.4765 0.6009 0.7752
No log 2.8649 106 0.6020 0.4841 0.6020 0.7759
No log 2.9189 108 0.6110 0.4841 0.6110 0.7817
No log 2.9730 110 0.5838 0.5101 0.5838 0.7641
No log 3.0270 112 0.5752 0.5079 0.5752 0.7584
No log 3.0811 114 0.5959 0.5025 0.5959 0.7720
No log 3.1351 116 0.6193 0.4953 0.6193 0.7869
No log 3.1892 118 0.6218 0.5215 0.6218 0.7886
No log 3.2432 120 0.5862 0.5407 0.5862 0.7656
No log 3.2973 122 0.5718 0.4996 0.5718 0.7562
No log 3.3514 124 0.5824 0.5463 0.5824 0.7631
No log 3.4054 126 0.5890 0.5049 0.5890 0.7675
No log 3.4595 128 0.6156 0.5805 0.6156 0.7846
No log 3.5135 130 0.6206 0.5241 0.6206 0.7878
No log 3.5676 132 0.5956 0.5260 0.5956 0.7717
No log 3.6216 134 0.5993 0.5106 0.5993 0.7742
No log 3.6757 136 0.6001 0.5011 0.6001 0.7747
No log 3.7297 138 0.6086 0.5338 0.6086 0.7801
No log 3.7838 140 0.6184 0.5338 0.6184 0.7864
No log 3.8378 142 0.6107 0.5049 0.6107 0.7815
No log 3.8919 144 0.6276 0.5312 0.6276 0.7922
No log 3.9459 146 0.6161 0.5040 0.6161 0.7849
No log 4.0 148 0.5846 0.5223 0.5846 0.7646
No log 4.0541 150 0.5814 0.5411 0.5814 0.7625
No log 4.1081 152 0.5830 0.5429 0.5830 0.7635
No log 4.1622 154 0.5896 0.5244 0.5896 0.7678
No log 4.2162 156 0.6281 0.4909 0.6281 0.7925
No log 4.2703 158 0.6687 0.5358 0.6687 0.8178
No log 4.3243 160 0.6464 0.5189 0.6464 0.8040
No log 4.3784 162 0.6128 0.4968 0.6128 0.7828
No log 4.4324 164 0.5871 0.5283 0.5871 0.7662
No log 4.4865 166 0.5920 0.5065 0.5920 0.7694
No log 4.5405 168 0.6367 0.4591 0.6367 0.7980
No log 4.5946 170 0.7408 0.5032 0.7408 0.8607
No log 4.6486 172 0.7997 0.5019 0.7997 0.8943
No log 4.7027 174 0.7554 0.4943 0.7554 0.8691
No log 4.7568 176 0.6540 0.4924 0.6540 0.8087
No log 4.8108 178 0.5736 0.5347 0.5736 0.7574
No log 4.8649 180 0.5585 0.5511 0.5585 0.7473
No log 4.9189 182 0.5654 0.5034 0.5654 0.7520
No log 4.9730 184 0.5919 0.4485 0.5919 0.7694
No log 5.0270 186 0.6460 0.4994 0.6460 0.8037
No log 5.0811 188 0.6755 0.4907 0.6755 0.8219
No log 5.1351 190 0.7000 0.4390 0.7000 0.8367
No log 5.1892 192 0.7180 0.4663 0.7180 0.8474
No log 5.2432 194 0.7402 0.4607 0.7402 0.8603
No log 5.2973 196 0.7742 0.4686 0.7742 0.8799
No log 5.3514 198 0.7798 0.4686 0.7798 0.8831
No log 5.4054 200 0.7832 0.4596 0.7832 0.8850
No log 5.4595 202 0.7777 0.4724 0.7777 0.8819
No log 5.5135 204 0.7789 0.4824 0.7789 0.8825
No log 5.5676 206 0.8012 0.4885 0.8012 0.8951
No log 5.6216 208 0.7901 0.4731 0.7901 0.8889
No log 5.6757 210 0.7721 0.4853 0.7721 0.8787
No log 5.7297 212 0.7737 0.4966 0.7737 0.8796
No log 5.7838 214 0.7659 0.4803 0.7659 0.8752
No log 5.8378 216 0.7520 0.4740 0.7520 0.8672
No log 5.8919 218 0.7373 0.4473 0.7373 0.8587
No log 5.9459 220 0.7420 0.4741 0.7420 0.8614
No log 6.0 222 0.7573 0.4539 0.7573 0.8702
No log 6.0541 224 0.7477 0.4731 0.7477 0.8647
No log 6.1081 226 0.7189 0.4718 0.7189 0.8479
No log 6.1622 228 0.7199 0.4627 0.7199 0.8484
No log 6.2162 230 0.7510 0.5072 0.7510 0.8666
No log 6.2703 232 0.7670 0.4992 0.7670 0.8758
No log 6.3243 234 0.7543 0.4773 0.7543 0.8685
No log 6.3784 236 0.7353 0.4832 0.7353 0.8575
No log 6.4324 238 0.7181 0.4727 0.7181 0.8474
No log 6.4865 240 0.7124 0.4785 0.7124 0.8441
No log 6.5405 242 0.7039 0.4869 0.7039 0.8390
No log 6.5946 244 0.6924 0.5176 0.6924 0.8321
No log 6.6486 246 0.6674 0.4945 0.6674 0.8169
No log 6.7027 248 0.6413 0.5114 0.6413 0.8008
No log 6.7568 250 0.6476 0.5314 0.6476 0.8047
No log 6.8108 252 0.6757 0.4789 0.6757 0.8220
No log 6.8649 254 0.6974 0.4639 0.6974 0.8351
No log 6.9189 256 0.7150 0.4563 0.7150 0.8455
No log 6.9730 258 0.7161 0.4602 0.7161 0.8462
No log 7.0270 260 0.7059 0.4636 0.7059 0.8402
No log 7.0811 262 0.6928 0.4368 0.6928 0.8323
No log 7.1351 264 0.6555 0.4672 0.6555 0.8096
No log 7.1892 266 0.6267 0.4931 0.6267 0.7916
No log 7.2432 268 0.6337 0.4489 0.6337 0.7961
No log 7.2973 270 0.6516 0.5030 0.6516 0.8072
No log 7.3514 272 0.6857 0.4811 0.6857 0.8281
No log 7.4054 274 0.6871 0.4811 0.6871 0.8289
No log 7.4595 276 0.6629 0.5392 0.6629 0.8142
No log 7.5135 278 0.6684 0.4905 0.6684 0.8176
No log 7.5676 280 0.6725 0.5079 0.6725 0.8201
No log 7.6216 282 0.6922 0.4407 0.6922 0.8320
No log 7.6757 284 0.7112 0.4540 0.7112 0.8433
No log 7.7297 286 0.7174 0.4788 0.7174 0.8470
No log 7.7838 288 0.7129 0.5035 0.7129 0.8443
No log 7.8378 290 0.7134 0.5040 0.7134 0.8447
No log 7.8919 292 0.7259 0.4946 0.7259 0.8520
No log 7.9459 294 0.7158 0.5254 0.7158 0.8460
No log 8.0 296 0.6930 0.5178 0.6930 0.8324
No log 8.0541 298 0.6874 0.5104 0.6874 0.8291
No log 8.1081 300 0.6798 0.4904 0.6798 0.8245
No log 8.1622 302 0.6629 0.5069 0.6629 0.8142
No log 8.2162 304 0.6693 0.4895 0.6693 0.8181
No log 8.2703 306 0.7121 0.5135 0.7121 0.8438
No log 8.3243 308 0.7167 0.5206 0.7167 0.8466
No log 8.3784 310 0.6837 0.5079 0.6837 0.8269
No log 8.4324 312 0.6629 0.5 0.6629 0.8142
No log 8.4865 314 0.6628 0.4737 0.6628 0.8141
No log 8.5405 316 0.6572 0.4811 0.6572 0.8107
No log 8.5946 318 0.6501 0.4922 0.6501 0.8063
No log 8.6486 320 0.6798 0.5173 0.6798 0.8245
No log 8.7027 322 0.7027 0.4767 0.7027 0.8383
No log 8.7568 324 0.6815 0.5056 0.6815 0.8255
No log 8.8108 326 0.6645 0.5405 0.6645 0.8152
No log 8.8649 328 0.6725 0.5055 0.6725 0.8201
No log 8.9189 330 0.6900 0.5315 0.6900 0.8307
No log 8.9730 332 0.6745 0.5395 0.6745 0.8213
No log 9.0270 334 0.6613 0.5341 0.6613 0.8132
No log 9.0811 336 0.6759 0.5363 0.6759 0.8221
No log 9.1351 338 0.6870 0.5076 0.6870 0.8289
No log 9.1892 340 0.6846 0.5075 0.6846 0.8274
No log 9.2432 342 0.6797 0.4927 0.6797 0.8244
No log 9.2973 344 0.6877 0.5109 0.6877 0.8293
No log 9.3514 346 0.6776 0.5006 0.6776 0.8232
No log 9.4054 348 0.6884 0.5374 0.6884 0.8297
No log 9.4595 350 0.7374 0.4809 0.7374 0.8587
No log 9.5135 352 0.7519 0.5166 0.7519 0.8671
No log 9.5676 354 0.7259 0.5071 0.7259 0.8520
No log 9.6216 356 0.6952 0.5105 0.6952 0.8338
No log 9.6757 358 0.6810 0.5145 0.6810 0.8252
No log 9.7297 360 0.6701 0.4936 0.6701 0.8186
No log 9.7838 362 0.6497 0.5214 0.6497 0.8060
No log 9.8378 364 0.6672 0.5071 0.6672 0.8168
No log 9.8919 366 0.6960 0.4937 0.6960 0.8342
No log 9.9459 368 0.6861 0.5039 0.6861 0.8283
No log 10.0 370 0.6489 0.5370 0.6489 0.8055
No log 10.0541 372 0.6306 0.5290 0.6306 0.7941
No log 10.1081 374 0.6408 0.4813 0.6408 0.8005
No log 10.1622 376 0.6407 0.4813 0.6407 0.8004
No log 10.2162 378 0.6207 0.4988 0.6207 0.7879
No log 10.2703 380 0.6131 0.5564 0.6131 0.7830
No log 10.3243 382 0.6259 0.5566 0.6259 0.7911
No log 10.3784 384 0.6438 0.5466 0.6438 0.8024
No log 10.4324 386 0.6538 0.5540 0.6538 0.8086
No log 10.4865 388 0.6585 0.5724 0.6585 0.8115
No log 10.5405 390 0.6712 0.5622 0.6712 0.8193
No log 10.5946 392 0.6924 0.5201 0.6924 0.8321
No log 10.6486 394 0.7146 0.5268 0.7146 0.8453
No log 10.7027 396 0.7195 0.5257 0.7195 0.8482
No log 10.7568 398 0.7229 0.5378 0.7229 0.8502
No log 10.8108 400 0.7196 0.5378 0.7196 0.8483
No log 10.8649 402 0.6994 0.5387 0.6994 0.8363
No log 10.9189 404 0.6758 0.5211 0.6758 0.8221
No log 10.9730 406 0.6466 0.5758 0.6466 0.8041
No log 11.0270 408 0.6312 0.5697 0.6312 0.7945
No log 11.0811 410 0.6361 0.5803 0.6361 0.7975
No log 11.1351 412 0.6585 0.5372 0.6585 0.8115
No log 11.1892 414 0.6551 0.5372 0.6551 0.8094
No log 11.2432 416 0.6350 0.5731 0.6350 0.7968
No log 11.2973 418 0.6285 0.5434 0.6285 0.7928
No log 11.3514 420 0.6319 0.5332 0.6319 0.7949
No log 11.4054 422 0.6373 0.5099 0.6373 0.7983
No log 11.4595 424 0.6612 0.5321 0.6612 0.8131
No log 11.5135 426 0.6868 0.5377 0.6868 0.8287
No log 11.5676 428 0.7079 0.5003 0.7079 0.8414
No log 11.6216 430 0.7244 0.5059 0.7244 0.8511
No log 11.6757 432 0.7215 0.4969 0.7215 0.8494
No log 11.7297 434 0.6988 0.4994 0.6988 0.8359
No log 11.7838 436 0.6598 0.5023 0.6598 0.8123
No log 11.8378 438 0.6454 0.5686 0.6454 0.8034
No log 11.8919 440 0.6705 0.5424 0.6705 0.8189
No log 11.9459 442 0.6883 0.5275 0.6883 0.8296
No log 12.0 444 0.6563 0.5437 0.6563 0.8101
No log 12.0541 446 0.6145 0.5761 0.6145 0.7839
No log 12.1081 448 0.6173 0.5112 0.6173 0.7857
No log 12.1622 450 0.6560 0.4994 0.6560 0.8099
No log 12.2162 452 0.6666 0.5027 0.6666 0.8165
No log 12.2703 454 0.6453 0.5268 0.6453 0.8033
No log 12.3243 456 0.6206 0.5188 0.6206 0.7878
No log 12.3784 458 0.6151 0.5502 0.6151 0.7843
No log 12.4324 460 0.6231 0.5955 0.6231 0.7894
No log 12.4865 462 0.6260 0.5654 0.6260 0.7912
No log 12.5405 464 0.6311 0.5773 0.6311 0.7944
No log 12.5946 466 0.6396 0.5608 0.6396 0.7998
No log 12.6486 468 0.6512 0.5637 0.6512 0.8070
No log 12.7027 470 0.6638 0.5767 0.6638 0.8148
No log 12.7568 472 0.6593 0.5782 0.6593 0.8120
No log 12.8108 474 0.6521 0.5799 0.6521 0.8076
No log 12.8649 476 0.6437 0.5451 0.6437 0.8023
No log 12.9189 478 0.6308 0.5755 0.6308 0.7943
No log 12.9730 480 0.6250 0.5317 0.6250 0.7906
No log 13.0270 482 0.6189 0.5389 0.6189 0.7867
No log 13.0811 484 0.6137 0.5594 0.6137 0.7834
No log 13.1351 486 0.6060 0.5344 0.6060 0.7785
No log 13.1892 488 0.6024 0.5347 0.6024 0.7762
No log 13.2432 490 0.6081 0.5351 0.6081 0.7798
No log 13.2973 492 0.6235 0.5305 0.6235 0.7896
No log 13.3514 494 0.6345 0.5803 0.6345 0.7966
No log 13.4054 496 0.6497 0.5898 0.6497 0.8060
No log 13.4595 498 0.6614 0.5861 0.6614 0.8132
0.5211 13.5135 500 0.6754 0.5504 0.6754 0.8218
0.5211 13.5676 502 0.6963 0.5219 0.6963 0.8345
0.5211 13.6216 504 0.7023 0.5268 0.7023 0.8380
0.5211 13.6757 506 0.6935 0.5395 0.6935 0.8328
0.5211 13.7297 508 0.6922 0.5421 0.6922 0.8320
0.5211 13.7838 510 0.7035 0.4606 0.7035 0.8387

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.1+cu121
  • Datasets 3.2.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERTFreeze_run1_AugV5_k15_task2_organization

Finetuned
(4024)
this model