ArabicNewSplits6_FineTuningAraBERTFreeze_run1_AugV5_k20_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8345
  • Qwk: 0.6336
  • Mse: 0.8345
  • Rmse: 0.9135

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.04 2 7.8563 -0.0507 7.8563 2.8029
No log 0.08 4 5.5663 -0.0566 5.5663 2.3593
No log 0.12 6 3.9660 0.0409 3.9660 1.9915
No log 0.16 8 3.0009 0.0662 3.0009 1.7323
No log 0.2 10 2.3537 0.0159 2.3537 1.5342
No log 0.24 12 1.8207 0.1671 1.8207 1.3493
No log 0.28 14 1.5020 0.1735 1.5020 1.2256
No log 0.32 16 1.2822 0.2074 1.2822 1.1323
No log 0.36 18 1.1592 0.3263 1.1592 1.0766
No log 0.4 20 1.1140 0.4009 1.1140 1.0555
No log 0.44 22 1.1284 0.4155 1.1284 1.0623
No log 0.48 24 1.1839 0.3824 1.1839 1.0881
No log 0.52 26 1.1503 0.3824 1.1503 1.0725
No log 0.56 28 1.0423 0.4293 1.0423 1.0209
No log 0.6 30 1.0190 0.4102 1.0190 1.0095
No log 0.64 32 1.0203 0.4384 1.0203 1.0101
No log 0.68 34 1.0233 0.4229 1.0233 1.0116
No log 0.72 36 1.1016 0.4014 1.1016 1.0496
No log 0.76 38 1.2863 0.3649 1.2863 1.1342
No log 0.8 40 1.4000 0.3876 1.4000 1.1832
No log 0.84 42 1.2997 0.3629 1.2997 1.1400
No log 0.88 44 1.1830 0.3450 1.1830 1.0877
No log 0.92 46 1.0643 0.4057 1.0643 1.0316
No log 0.96 48 1.0079 0.4493 1.0079 1.0039
No log 1.0 50 1.0323 0.4431 1.0323 1.0160
No log 1.04 52 1.1497 0.3954 1.1497 1.0722
No log 1.08 54 1.3968 0.3697 1.3968 1.1819
No log 1.12 56 1.5225 0.3458 1.5225 1.2339
No log 1.16 58 1.4307 0.3857 1.4307 1.1961
No log 1.2 60 1.2206 0.3622 1.2206 1.1048
No log 1.24 62 1.0214 0.4475 1.0214 1.0107
No log 1.28 64 0.9998 0.4539 0.9998 0.9999
No log 1.32 66 0.9632 0.4539 0.9632 0.9814
No log 1.3600 68 0.9332 0.4839 0.9332 0.9660
No log 1.4 70 0.9160 0.4992 0.9160 0.9571
No log 1.44 72 0.9108 0.5196 0.9108 0.9544
No log 1.48 74 0.8868 0.5324 0.8868 0.9417
No log 1.52 76 0.8584 0.5627 0.8584 0.9265
No log 1.56 78 0.8375 0.5823 0.8375 0.9151
No log 1.6 80 0.8692 0.5593 0.8692 0.9323
No log 1.6400 82 0.9557 0.5327 0.9557 0.9776
No log 1.6800 84 1.0452 0.4590 1.0452 1.0223
No log 1.72 86 1.0918 0.4807 1.0918 1.0449
No log 1.76 88 0.9508 0.5564 0.9508 0.9751
No log 1.8 90 0.7703 0.6300 0.7703 0.8777
No log 1.8400 92 0.7451 0.6295 0.7451 0.8632
No log 1.88 94 0.7419 0.6149 0.7419 0.8613
No log 1.92 96 0.7389 0.6258 0.7389 0.8596
No log 1.96 98 0.7530 0.6164 0.7530 0.8677
No log 2.0 100 0.7581 0.6256 0.7581 0.8707
No log 2.04 102 0.7472 0.6253 0.7472 0.8644
No log 2.08 104 0.7592 0.6053 0.7592 0.8713
No log 2.12 106 0.7701 0.5985 0.7701 0.8776
No log 2.16 108 0.7656 0.6228 0.7656 0.8750
No log 2.2 110 0.8014 0.6618 0.8014 0.8952
No log 2.24 112 0.9364 0.5646 0.9364 0.9677
No log 2.2800 114 0.9559 0.5690 0.9559 0.9777
No log 2.32 116 0.8869 0.5972 0.8869 0.9418
No log 2.36 118 0.7834 0.6508 0.7834 0.8851
No log 2.4 120 0.7727 0.6388 0.7727 0.8791
No log 2.44 122 0.8925 0.5207 0.8925 0.9447
No log 2.48 124 0.8809 0.5285 0.8809 0.9386
No log 2.52 126 0.7654 0.6122 0.7654 0.8749
No log 2.56 128 0.7577 0.6438 0.7577 0.8705
No log 2.6 130 0.9400 0.5859 0.9400 0.9696
No log 2.64 132 1.0132 0.5199 1.0132 1.0066
No log 2.68 134 0.9161 0.6015 0.9161 0.9571
No log 2.7200 136 0.7676 0.6373 0.7676 0.8761
No log 2.76 138 0.6940 0.6304 0.6940 0.8331
No log 2.8 140 0.7053 0.6585 0.7053 0.8398
No log 2.84 142 0.7146 0.6220 0.7146 0.8453
No log 2.88 144 0.7011 0.6340 0.7011 0.8373
No log 2.92 146 0.7125 0.6064 0.7125 0.8441
No log 2.96 148 0.7274 0.5977 0.7274 0.8529
No log 3.0 150 0.7459 0.5927 0.7459 0.8636
No log 3.04 152 0.7486 0.5947 0.7486 0.8652
No log 3.08 154 0.7622 0.6034 0.7622 0.8730
No log 3.12 156 0.7712 0.5939 0.7712 0.8782
No log 3.16 158 0.7772 0.6161 0.7772 0.8816
No log 3.2 160 0.7837 0.5871 0.7837 0.8853
No log 3.24 162 0.7790 0.5782 0.7790 0.8826
No log 3.2800 164 0.7715 0.6617 0.7715 0.8784
No log 3.32 166 0.7903 0.6267 0.7903 0.8890
No log 3.36 168 0.7958 0.6222 0.7958 0.8921
No log 3.4 170 0.7911 0.6654 0.7911 0.8894
No log 3.44 172 0.8170 0.6125 0.8170 0.9039
No log 3.48 174 0.8222 0.6068 0.8222 0.9067
No log 3.52 176 0.7889 0.6671 0.7889 0.8882
No log 3.56 178 0.8321 0.6013 0.8321 0.9122
No log 3.6 180 0.8114 0.5989 0.8114 0.9008
No log 3.64 182 0.7668 0.6611 0.7668 0.8757
No log 3.68 184 0.7685 0.6733 0.7685 0.8766
No log 3.7200 186 0.7574 0.6774 0.7574 0.8703
No log 3.76 188 0.7483 0.6758 0.7483 0.8650
No log 3.8 190 0.7564 0.6841 0.7564 0.8697
No log 3.84 192 0.7408 0.6920 0.7408 0.8607
No log 3.88 194 0.7320 0.6723 0.7320 0.8556
No log 3.92 196 0.8709 0.5654 0.8709 0.9332
No log 3.96 198 0.9780 0.4821 0.9780 0.9889
No log 4.0 200 0.8890 0.5667 0.8890 0.9428
No log 4.04 202 0.7488 0.6550 0.7488 0.8653
No log 4.08 204 0.7494 0.6568 0.7494 0.8657
No log 4.12 206 0.8183 0.6561 0.8183 0.9046
No log 4.16 208 0.8017 0.6578 0.8017 0.8954
No log 4.2 210 0.7274 0.6871 0.7274 0.8529
No log 4.24 212 0.7149 0.6454 0.7149 0.8455
No log 4.28 214 0.7561 0.6153 0.7561 0.8695
No log 4.32 216 0.7463 0.6546 0.7463 0.8639
No log 4.36 218 0.7010 0.6419 0.7010 0.8373
No log 4.4 220 0.7140 0.6764 0.7140 0.8450
No log 4.44 222 0.7260 0.6722 0.7260 0.8521
No log 4.48 224 0.6997 0.6789 0.6997 0.8365
No log 4.52 226 0.7039 0.6647 0.7039 0.8390
No log 4.5600 228 0.7028 0.6930 0.7028 0.8383
No log 4.6 230 0.7202 0.6984 0.7202 0.8487
No log 4.64 232 0.7387 0.6614 0.7387 0.8595
No log 4.68 234 0.7328 0.6900 0.7328 0.8561
No log 4.72 236 0.7449 0.6411 0.7449 0.8631
No log 4.76 238 0.8093 0.6361 0.8093 0.8996
No log 4.8 240 0.9498 0.5120 0.9498 0.9746
No log 4.84 242 0.9376 0.5389 0.9376 0.9683
No log 4.88 244 0.7959 0.6042 0.7959 0.8921
No log 4.92 246 0.7539 0.6834 0.7539 0.8683
No log 4.96 248 0.7633 0.6717 0.7633 0.8737
No log 5.0 250 0.7543 0.6858 0.7543 0.8685
No log 5.04 252 0.8313 0.6081 0.8313 0.9118
No log 5.08 254 1.0345 0.3985 1.0345 1.0171
No log 5.12 256 1.0368 0.3983 1.0368 1.0182
No log 5.16 258 0.9021 0.5049 0.9021 0.9498
No log 5.2 260 0.8374 0.6018 0.8374 0.9151
No log 5.24 262 0.8459 0.5889 0.8459 0.9197
No log 5.28 264 0.8177 0.5833 0.8177 0.9042
No log 5.32 266 0.7586 0.6415 0.7586 0.8710
No log 5.36 268 0.7522 0.6525 0.7522 0.8673
No log 5.4 270 0.7412 0.6756 0.7412 0.8609
No log 5.44 272 0.7384 0.6749 0.7384 0.8593
No log 5.48 274 0.7514 0.6495 0.7514 0.8669
No log 5.52 276 0.7850 0.6221 0.7850 0.8860
No log 5.5600 278 0.8000 0.6116 0.8000 0.8944
No log 5.6 280 0.8231 0.6117 0.8231 0.9072
No log 5.64 282 0.8563 0.6037 0.8563 0.9254
No log 5.68 284 0.8534 0.6042 0.8534 0.9238
No log 5.72 286 0.8326 0.6244 0.8326 0.9124
No log 5.76 288 0.8394 0.6195 0.8394 0.9162
No log 5.8 290 0.8796 0.6050 0.8796 0.9379
No log 5.84 292 0.8743 0.6050 0.8743 0.9350
No log 5.88 294 0.8210 0.6374 0.8210 0.9061
No log 5.92 296 0.7870 0.6253 0.7870 0.8871
No log 5.96 298 0.7804 0.6408 0.7804 0.8834
No log 6.0 300 0.7898 0.6706 0.7898 0.8887
No log 6.04 302 0.7935 0.6880 0.7935 0.8908
No log 6.08 304 0.7530 0.6936 0.7530 0.8678
No log 6.12 306 0.7502 0.6661 0.7502 0.8662
No log 6.16 308 0.7584 0.6653 0.7584 0.8709
No log 6.2 310 0.7352 0.6679 0.7352 0.8574
No log 6.24 312 0.7302 0.6487 0.7302 0.8545
No log 6.28 314 0.7596 0.6509 0.7596 0.8716
No log 6.32 316 0.7608 0.6492 0.7608 0.8722
No log 6.36 318 0.7535 0.6584 0.7535 0.8680
No log 6.4 320 0.7665 0.6459 0.7665 0.8755
No log 6.44 322 0.7734 0.6393 0.7734 0.8794
No log 6.48 324 0.7784 0.6586 0.7784 0.8823
No log 6.52 326 0.7852 0.6528 0.7852 0.8861
No log 6.5600 328 0.8016 0.6310 0.8016 0.8953
No log 6.6 330 0.8351 0.6338 0.8351 0.9139
No log 6.64 332 0.8207 0.6209 0.8207 0.9059
No log 6.68 334 0.8393 0.6347 0.8393 0.9161
No log 6.72 336 0.8406 0.6269 0.8406 0.9168
No log 6.76 338 0.8036 0.6404 0.8036 0.8964
No log 6.8 340 0.8067 0.6448 0.8067 0.8982
No log 6.84 342 0.8347 0.6352 0.8347 0.9136
No log 6.88 344 0.8125 0.6546 0.8125 0.9014
No log 6.92 346 0.7832 0.6760 0.7832 0.8850
No log 6.96 348 0.8715 0.5345 0.8715 0.9335
No log 7.0 350 0.9094 0.5179 0.9094 0.9536
No log 7.04 352 0.8521 0.5693 0.8521 0.9231
No log 7.08 354 0.7815 0.6734 0.7815 0.8841
No log 7.12 356 0.7874 0.6387 0.7874 0.8873
No log 7.16 358 0.7781 0.6440 0.7781 0.8821
No log 7.2 360 0.7611 0.6864 0.7611 0.8724
No log 7.24 362 0.8052 0.6273 0.8052 0.8973
No log 7.28 364 0.9069 0.5751 0.9069 0.9523
No log 7.32 366 0.9091 0.5737 0.9091 0.9535
No log 7.36 368 0.8152 0.6123 0.8152 0.9029
No log 7.4 370 0.7632 0.6812 0.7632 0.8736
No log 7.44 372 0.7995 0.6469 0.7995 0.8941
No log 7.48 374 0.8042 0.6453 0.8042 0.8968
No log 7.52 376 0.7828 0.6518 0.7828 0.8848
No log 7.5600 378 0.7980 0.6674 0.7980 0.8933
No log 7.6 380 0.8576 0.625 0.8576 0.9260
No log 7.64 382 0.8750 0.6059 0.8750 0.9354
No log 7.68 384 0.8411 0.6357 0.8411 0.9171
No log 7.72 386 0.7864 0.6397 0.7864 0.8868
No log 7.76 388 0.7694 0.6355 0.7694 0.8772
No log 7.8 390 0.7656 0.6445 0.7656 0.8750
No log 7.84 392 0.7929 0.6260 0.7929 0.8904
No log 7.88 394 0.8232 0.6319 0.8232 0.9073
No log 7.92 396 0.8308 0.6193 0.8308 0.9115
No log 7.96 398 0.7930 0.6276 0.7930 0.8905
No log 8.0 400 0.7650 0.6298 0.7650 0.8746
No log 8.04 402 0.7925 0.6262 0.7925 0.8902
No log 8.08 404 0.8383 0.6248 0.8383 0.9156
No log 8.12 406 0.8917 0.6406 0.8917 0.9443
No log 8.16 408 0.8543 0.6234 0.8543 0.9243
No log 8.2 410 0.8128 0.6441 0.8128 0.9016
No log 8.24 412 0.8212 0.6441 0.8212 0.9062
No log 8.28 414 0.8449 0.6262 0.8449 0.9192
No log 8.32 416 0.8718 0.6386 0.8718 0.9337
No log 8.36 418 0.8664 0.6401 0.8664 0.9308
No log 8.4 420 0.8356 0.6382 0.8356 0.9141
No log 8.44 422 0.8340 0.6308 0.8340 0.9132
No log 8.48 424 0.8676 0.6223 0.8676 0.9314
No log 8.52 426 0.8959 0.6252 0.8959 0.9465
No log 8.56 428 0.8865 0.6017 0.8865 0.9415
No log 8.6 430 0.8700 0.6273 0.8700 0.9327
No log 8.64 432 0.8778 0.5992 0.8778 0.9369
No log 8.68 434 0.8870 0.5852 0.8870 0.9418
No log 8.72 436 0.8625 0.6162 0.8625 0.9287
No log 8.76 438 0.8608 0.6147 0.8608 0.9278
No log 8.8 440 0.8516 0.6218 0.8516 0.9228
No log 8.84 442 0.8315 0.6643 0.8315 0.9118
No log 8.88 444 0.8476 0.5826 0.8476 0.9206
No log 8.92 446 0.8410 0.5958 0.8410 0.9171
No log 8.96 448 0.8466 0.6472 0.8466 0.9201
No log 9.0 450 0.8734 0.6238 0.8734 0.9346
No log 9.04 452 0.8509 0.6433 0.8509 0.9224
No log 9.08 454 0.8114 0.6345 0.8114 0.9008
No log 9.12 456 0.7970 0.6316 0.7970 0.8928
No log 9.16 458 0.8027 0.6212 0.8027 0.8959
No log 9.2 460 0.8348 0.6028 0.8348 0.9137
No log 9.24 462 0.8644 0.6076 0.8644 0.9297
No log 9.28 464 0.8841 0.6050 0.8841 0.9403
No log 9.32 466 0.8369 0.6046 0.8369 0.9148
No log 9.36 468 0.8022 0.6297 0.8022 0.8957
No log 9.4 470 0.8063 0.6101 0.8063 0.8979
No log 9.44 472 0.8302 0.6351 0.8302 0.9112
No log 9.48 474 0.8816 0.6246 0.8816 0.9389
No log 9.52 476 0.9204 0.6320 0.9204 0.9594
No log 9.56 478 0.9004 0.6270 0.9004 0.9489
No log 9.6 480 0.8837 0.6284 0.8837 0.9401
No log 9.64 482 0.8816 0.6223 0.8816 0.9389
No log 9.68 484 0.8208 0.6305 0.8208 0.9060
No log 9.72 486 0.8047 0.6411 0.8047 0.8970
No log 9.76 488 0.8345 0.6291 0.8345 0.9135
No log 9.8 490 0.8630 0.6146 0.8630 0.9290
No log 9.84 492 0.9064 0.6097 0.9064 0.9520
No log 9.88 494 0.8986 0.6111 0.8986 0.9479
No log 9.92 496 0.8726 0.6260 0.8726 0.9341
No log 9.96 498 0.8265 0.6186 0.8265 0.9091
0.617 10.0 500 0.8035 0.6694 0.8035 0.8964
0.617 10.04 502 0.8052 0.6576 0.8052 0.8973
0.617 10.08 504 0.8323 0.6314 0.8323 0.9123
0.617 10.12 506 0.8979 0.5979 0.8979 0.9476
0.617 10.16 508 0.8829 0.5955 0.8829 0.9396
0.617 10.2 510 0.8345 0.6336 0.8345 0.9135

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.1+cu121
  • Datasets 3.2.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERTFreeze_run1_AugV5_k20_task1_organization

Finetuned
(4024)
this model