ArabicNewSplits5_FineTuningAraBERT_run1_AugV5_k9_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8348
  • Qwk: 0.3080
  • Mse: 0.8348
  • Rmse: 0.9136

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0370 2 3.1522 -0.0350 3.1522 1.7754
No log 0.0741 4 1.4073 -0.0070 1.4073 1.1863
No log 0.1111 6 0.9353 0.0632 0.9353 0.9671
No log 0.1481 8 0.8966 0.0977 0.8966 0.9469
No log 0.1852 10 0.9805 0.0294 0.9805 0.9902
No log 0.2222 12 1.1073 0.0 1.1073 1.0523
No log 0.2593 14 1.1223 0.0 1.1223 1.0594
No log 0.2963 16 0.9486 0.0677 0.9486 0.9740
No log 0.3333 18 0.9119 0.0431 0.9119 0.9549
No log 0.3704 20 0.9116 0.0345 0.9116 0.9548
No log 0.4074 22 0.8540 0.0345 0.8540 0.9241
No log 0.4444 24 0.8872 0.0345 0.8872 0.9419
No log 0.4815 26 0.7915 0.1667 0.7915 0.8897
No log 0.5185 28 0.6680 0.3548 0.6680 0.8173
No log 0.5556 30 0.6924 0.3231 0.6924 0.8321
No log 0.5926 32 1.3498 0.0596 1.3498 1.1618
No log 0.6296 34 1.6157 0.0264 1.6157 1.2711
No log 0.6667 36 0.8673 0.1347 0.8673 0.9313
No log 0.7037 38 0.6386 0.1813 0.6386 0.7991
No log 0.7407 40 0.6499 0.2749 0.6499 0.8062
No log 0.7778 42 0.6897 0.3118 0.6897 0.8305
No log 0.8148 44 0.6932 0.2994 0.6932 0.8326
No log 0.8519 46 0.6545 0.1111 0.6545 0.8090
No log 0.8889 48 0.7398 0.1398 0.7398 0.8601
No log 0.9259 50 0.9837 0.1169 0.9837 0.9918
No log 0.9630 52 0.9978 0.1453 0.9978 0.9989
No log 1.0 54 0.8539 0.1373 0.8539 0.9241
No log 1.0370 56 1.0273 -0.2440 1.0273 1.0136
No log 1.0741 58 1.0462 -0.1980 1.0462 1.0228
No log 1.1111 60 0.7698 -0.1667 0.7698 0.8774
No log 1.1481 62 0.6384 -0.0068 0.6384 0.7990
No log 1.1852 64 0.7964 0.1781 0.7964 0.8924
No log 1.2222 66 1.0575 0.1165 1.0575 1.0284
No log 1.2593 68 1.0022 0.1220 1.0022 1.0011
No log 1.2963 70 0.6335 0.2787 0.6335 0.7959
No log 1.3333 72 0.5628 0.0303 0.5628 0.7502
No log 1.3704 74 0.6382 0.0145 0.6382 0.7989
No log 1.4074 76 0.8029 -0.0370 0.8029 0.8960
No log 1.4444 78 0.8929 0.0531 0.8929 0.9450
No log 1.4815 80 0.7447 -0.0222 0.7447 0.8630
No log 1.5185 82 0.6178 0.0400 0.6178 0.7860
No log 1.5556 84 0.5615 0.1795 0.5615 0.7493
No log 1.5926 86 0.5783 0.2444 0.5783 0.7604
No log 1.6296 88 0.5615 0.2542 0.5615 0.7493
No log 1.6667 90 0.5163 0.2298 0.5163 0.7185
No log 1.7037 92 0.5273 0.3182 0.5273 0.7262
No log 1.7407 94 0.5965 0.2990 0.5965 0.7723
No log 1.7778 96 0.8870 0.264 0.8870 0.9418
No log 1.8148 98 0.7435 0.3043 0.7435 0.8623
No log 1.8519 100 0.5469 0.2289 0.5469 0.7396
No log 1.8889 102 0.5769 0.0811 0.5769 0.7595
No log 1.9259 104 0.5781 0.1083 0.5781 0.7603
No log 1.9630 106 0.6020 0.2857 0.6020 0.7759
No log 2.0 108 0.5774 0.1698 0.5774 0.7598
No log 2.0370 110 0.5686 0.0897 0.5686 0.7541
No log 2.0741 112 0.6161 0.1325 0.6161 0.7849
No log 2.1111 114 0.5673 0.3333 0.5673 0.7532
No log 2.1481 116 0.5462 0.3706 0.5462 0.7390
No log 2.1852 118 0.5402 0.375 0.5402 0.7350
No log 2.2222 120 0.6125 0.3118 0.6125 0.7826
No log 2.2593 122 0.6521 0.3769 0.6521 0.8075
No log 2.2963 124 0.7200 0.3303 0.7200 0.8486
No log 2.3333 126 0.6049 0.2893 0.6049 0.7777
No log 2.3704 128 0.6228 0.28 0.6228 0.7892
No log 2.4074 130 0.6251 0.3433 0.6251 0.7906
No log 2.4444 132 0.6994 0.3171 0.6994 0.8363
No log 2.4815 134 0.6909 0.3171 0.6909 0.8312
No log 2.5185 136 0.6641 0.2893 0.6641 0.8149
No log 2.5556 138 0.9355 0.1169 0.9355 0.9672
No log 2.5926 140 0.8967 0.1150 0.8967 0.9470
No log 2.6296 142 0.6575 0.1556 0.6575 0.8109
No log 2.6667 144 0.6403 0.2083 0.6403 0.8002
No log 2.7037 146 0.6888 0.1287 0.6888 0.8299
No log 2.7407 148 0.6316 0.1411 0.6316 0.7947
No log 2.7778 150 0.7056 0.1638 0.7056 0.8400
No log 2.8148 152 0.8160 0.2077 0.8160 0.9033
No log 2.8519 154 0.7447 0.2661 0.7447 0.8630
No log 2.8889 156 0.7600 0.2941 0.7600 0.8718
No log 2.9259 158 0.8448 0.2554 0.8448 0.9191
No log 2.9630 160 0.8390 0.3080 0.8390 0.9160
No log 3.0 162 0.8208 0.2922 0.8208 0.9060
No log 3.0370 164 0.7852 0.2212 0.7852 0.8861
No log 3.0741 166 0.8008 0.2727 0.8008 0.8949
No log 3.1111 168 0.8186 0.3226 0.8186 0.9048
No log 3.1481 170 0.9313 0.2741 0.9313 0.9651
No log 3.1852 172 0.8645 0.3134 0.8645 0.9298
No log 3.2222 174 0.7505 0.4419 0.7505 0.8663
No log 3.2593 176 0.6820 0.4375 0.6820 0.8258
No log 3.2963 178 0.7206 0.4234 0.7206 0.8489
No log 3.3333 180 0.6458 0.5094 0.6458 0.8036
No log 3.3704 182 0.5474 0.4286 0.5474 0.7399
No log 3.4074 184 0.6496 0.4605 0.6496 0.8060
No log 3.4444 186 0.8345 0.2593 0.8345 0.9135
No log 3.4815 188 1.0426 0.0938 1.0426 1.0211
No log 3.5185 190 0.9417 0.1525 0.9417 0.9704
No log 3.5556 192 0.8125 0.2877 0.8125 0.9014
No log 3.5926 194 0.8170 0.2877 0.8170 0.9039
No log 3.6296 196 0.9031 0.2479 0.9031 0.9503
No log 3.6667 198 1.2574 0.1572 1.2574 1.1213
No log 3.7037 200 1.5678 0.0881 1.5678 1.2521
No log 3.7407 202 1.5167 0.0638 1.5167 1.2315
No log 3.7778 204 1.1430 0.2168 1.1430 1.0691
No log 3.8148 206 0.8244 0.3619 0.8244 0.9079
No log 3.8519 208 0.7605 0.3859 0.7605 0.8720
No log 3.8889 210 0.8664 0.3000 0.8664 0.9308
No log 3.9259 212 1.2418 0.1847 1.2418 1.1144
No log 3.9630 214 1.2096 0.1661 1.2096 1.0998
No log 4.0 216 0.8083 0.2432 0.8083 0.8991
No log 4.0370 218 0.6281 0.3628 0.6281 0.7925
No log 4.0741 220 0.7134 0.3360 0.7134 0.8446
No log 4.1111 222 0.6291 0.4872 0.6291 0.7932
No log 4.1481 224 0.9616 0.25 0.9616 0.9806
No log 4.1852 226 0.9888 0.2308 0.9888 0.9944
No log 4.2222 228 0.7805 0.3195 0.7805 0.8835
No log 4.2593 230 0.7740 0.3719 0.7740 0.8798
No log 4.2963 232 0.7796 0.3701 0.7796 0.8830
No log 4.3333 234 0.8680 0.2615 0.8680 0.9317
No log 4.3704 236 1.1897 0.1254 1.1897 1.0907
No log 4.4074 238 1.2088 0.1254 1.2088 1.0994
No log 4.4444 240 0.9674 0.2908 0.9674 0.9835
No log 4.4815 242 0.8521 0.2548 0.8521 0.9231
No log 4.5185 244 0.8178 0.3893 0.8178 0.9043
No log 4.5556 246 0.7865 0.36 0.7865 0.8868
No log 4.5926 248 0.7988 0.2960 0.7988 0.8937
No log 4.6296 250 0.8298 0.2441 0.8298 0.9110
No log 4.6667 252 0.8006 0.2479 0.8006 0.8948
No log 4.7037 254 0.8669 0.2581 0.8669 0.9311
No log 4.7407 256 1.0008 0.2302 1.0008 1.0004
No log 4.7778 258 0.9535 0.2615 0.9535 0.9765
No log 4.8148 260 0.9979 0.2296 0.9979 0.9990
No log 4.8519 262 0.9658 0.2366 0.9658 0.9827
No log 4.8889 264 0.7978 0.3067 0.7978 0.8932
No log 4.9259 266 0.6832 0.3665 0.6832 0.8266
No log 4.9630 268 0.6760 0.2727 0.6760 0.8222
No log 5.0 270 0.7026 0.3010 0.7026 0.8382
No log 5.0370 272 0.9334 0.1811 0.9334 0.9661
No log 5.0741 274 1.0919 0.1587 1.0919 1.0450
No log 5.1111 276 1.0618 0.1822 1.0618 1.0305
No log 5.1481 278 0.8941 0.2803 0.8941 0.9456
No log 5.1852 280 0.8002 0.3391 0.8002 0.8945
No log 5.2222 282 0.7067 0.2511 0.7067 0.8407
No log 5.2593 284 0.6826 0.3271 0.6826 0.8262
No log 5.2963 286 0.7660 0.4035 0.7660 0.8752
No log 5.3333 288 0.8476 0.2327 0.8476 0.9206
No log 5.3704 290 0.7520 0.3333 0.7520 0.8672
No log 5.4074 292 0.6268 0.4081 0.6268 0.7917
No log 5.4444 294 0.6153 0.3684 0.6153 0.7844
No log 5.4815 296 0.6274 0.4286 0.6274 0.7921
No log 5.5185 298 0.7687 0.2922 0.7687 0.8768
No log 5.5556 300 0.9064 0.2771 0.9064 0.9520
No log 5.5926 302 0.9256 0.2441 0.9256 0.9621
No log 5.6296 304 0.9092 0.2741 0.9092 0.9535
No log 5.6667 306 0.8067 0.2961 0.8067 0.8981
No log 5.7037 308 0.7545 0.3058 0.7545 0.8686
No log 5.7407 310 0.7714 0.2982 0.7714 0.8783
No log 5.7778 312 0.8038 0.2348 0.8038 0.8965
No log 5.8148 314 0.7480 0.2364 0.7480 0.8649
No log 5.8519 316 0.7785 0.2072 0.7785 0.8823
No log 5.8889 318 0.9362 0.2131 0.9362 0.9675
No log 5.9259 320 1.0854 0.2401 1.0854 1.0418
No log 5.9630 322 1.1117 0.2401 1.1117 1.0544
No log 6.0 324 0.9154 0.2727 0.9154 0.9568
No log 6.0370 326 0.7475 0.3580 0.7475 0.8646
No log 6.0741 328 0.7623 0.3496 0.7623 0.8731
No log 6.1111 330 0.8691 0.3251 0.8691 0.9323
No log 6.1481 332 1.1089 0.2714 1.1089 1.0531
No log 6.1852 334 1.0979 0.2701 1.0979 1.0478
No log 6.2222 336 1.0823 0.2701 1.0823 1.0404
No log 6.2593 338 1.0786 0.2701 1.0786 1.0386
No log 6.2963 340 0.9132 0.2432 0.9132 0.9556
No log 6.3333 342 0.7736 0.3613 0.7736 0.8796
No log 6.3704 344 0.7483 0.3220 0.7483 0.8650
No log 6.4074 346 0.8274 0.3333 0.8274 0.9096
No log 6.4444 348 1.0003 0.2472 1.0003 1.0002
No log 6.4815 350 1.1100 0.1608 1.1100 1.0535
No log 6.5185 352 1.0656 0.1601 1.0656 1.0323
No log 6.5556 354 0.9736 0.2174 0.9736 0.9867
No log 6.5926 356 0.7964 0.2775 0.7964 0.8924
No log 6.6296 358 0.7107 0.3091 0.7107 0.8430
No log 6.6667 360 0.7559 0.2811 0.7559 0.8694
No log 6.7037 362 0.9195 0.2441 0.9195 0.9589
No log 6.7407 364 0.9753 0.2714 0.9753 0.9876
No log 6.7778 366 0.9402 0.2450 0.9402 0.9696
No log 6.8148 368 1.0394 0.2409 1.0394 1.0195
No log 6.8519 370 1.0264 0.2115 1.0264 1.0131
No log 6.8889 372 0.8542 0.2980 0.8542 0.9242
No log 6.9259 374 0.7166 0.3103 0.7166 0.8465
No log 6.9630 376 0.6975 0.3362 0.6975 0.8351
No log 7.0 378 0.7031 0.3000 0.7031 0.8385
No log 7.0370 380 0.7453 0.3306 0.7453 0.8633
No log 7.0741 382 0.8808 0.3115 0.8808 0.9385
No log 7.1111 384 1.0123 0.2701 1.0123 1.0061
No log 7.1481 386 0.9905 0.2714 0.9905 0.9953
No log 7.1852 388 0.8353 0.3162 0.8353 0.9139
No log 7.2222 390 0.7429 0.3128 0.7429 0.8619
No log 7.2593 392 0.7456 0.3128 0.7456 0.8635
No log 7.2963 394 0.7603 0.3128 0.7603 0.8719
No log 7.3333 396 0.7652 0.2696 0.7652 0.8747
No log 7.3704 398 0.8208 0.3128 0.8208 0.9060
No log 7.4074 400 0.9455 0.2756 0.9455 0.9724
No log 7.4444 402 0.9889 0.2727 0.9889 0.9944
No log 7.4815 404 0.9094 0.3092 0.9094 0.9536
No log 7.5185 406 0.7821 0.3128 0.7821 0.8843
No log 7.5556 408 0.7281 0.2711 0.7281 0.8533
No log 7.5926 410 0.7634 0.3043 0.7634 0.8737
No log 7.6296 412 0.8085 0.3000 0.8085 0.8992
No log 7.6667 414 0.9630 0.3333 0.9630 0.9813
No log 7.7037 416 1.0940 0.2340 1.0940 1.0459
No log 7.7407 418 1.0565 0.2635 1.0565 1.0279
No log 7.7778 420 0.9154 0.3058 0.9154 0.9568
No log 7.8148 422 0.8210 0.2618 0.8210 0.9061
No log 7.8519 424 0.8069 0.2618 0.8069 0.8983
No log 7.8889 426 0.8410 0.3043 0.8410 0.9170
No log 7.9259 428 0.8979 0.3058 0.8979 0.9476
No log 7.9630 430 0.9917 0.2347 0.9917 0.9958
No log 8.0 432 0.9469 0.3282 0.9469 0.9731
No log 8.0370 434 0.8835 0.3036 0.8835 0.9399
No log 8.0741 436 0.9157 0.3333 0.9157 0.9569
No log 8.1111 438 0.9012 0.3333 0.9012 0.9493
No log 8.1481 440 0.8439 0.3036 0.8439 0.9186
No log 8.1852 442 0.7646 0.3043 0.7646 0.8744
No log 8.2222 444 0.7742 0.3043 0.7742 0.8799
No log 8.2593 446 0.7580 0.3422 0.7580 0.8706
No log 8.2963 448 0.7424 0.3422 0.7424 0.8616
No log 8.3333 450 0.7213 0.3917 0.7213 0.8493
No log 8.3704 452 0.6642 0.3524 0.6642 0.8150
No log 8.4074 454 0.6440 0.3524 0.6440 0.8025
No log 8.4444 456 0.6465 0.3524 0.6465 0.8040
No log 8.4815 458 0.6937 0.3962 0.6937 0.8329
No log 8.5185 460 0.7037 0.3962 0.7037 0.8389
No log 8.5556 462 0.6610 0.3860 0.6610 0.8130
No log 8.5926 464 0.6201 0.4123 0.6201 0.7875
No log 8.6296 466 0.6072 0.4123 0.6072 0.7792
No log 8.6667 468 0.6148 0.4123 0.6148 0.7841
No log 8.7037 470 0.6558 0.4444 0.6558 0.8098
No log 8.7407 472 0.7269 0.3128 0.7269 0.8526
No log 8.7778 474 0.7998 0.3036 0.7998 0.8943
No log 8.8148 476 0.8210 0.3115 0.8210 0.9061
No log 8.8519 478 0.8018 0.3036 0.8018 0.8954
No log 8.8889 480 0.7744 0.3058 0.7744 0.8800
No log 8.9259 482 0.7283 0.3722 0.7283 0.8534
No log 8.9630 484 0.7010 0.3722 0.7010 0.8373
No log 9.0 486 0.6779 0.3929 0.6779 0.8233
No log 9.0370 488 0.6913 0.3274 0.6913 0.8315
No log 9.0741 490 0.7178 0.3684 0.7178 0.8472
No log 9.1111 492 0.7447 0.3778 0.7447 0.8630
No log 9.1481 494 0.8016 0.3448 0.8016 0.8953
No log 9.1852 496 0.8558 0.3036 0.8558 0.9251
No log 9.2222 498 0.8768 0.2698 0.8768 0.9364
0.4239 9.2593 500 0.8658 0.2698 0.8658 0.9305
0.4239 9.2963 502 0.8451 0.2727 0.8451 0.9193
0.4239 9.3333 504 0.8284 0.3080 0.8284 0.9102
0.4239 9.3704 506 0.7939 0.3128 0.7939 0.8910
0.4239 9.4074 508 0.7612 0.3391 0.7612 0.8725
0.4239 9.4444 510 0.7458 0.2961 0.7458 0.8636
0.4239 9.4815 512 0.7414 0.2961 0.7414 0.8610
0.4239 9.5185 514 0.7542 0.2961 0.7542 0.8684
0.4239 9.5556 516 0.7691 0.2961 0.7691 0.8770
0.4239 9.5926 518 0.7889 0.3391 0.7889 0.8882
0.4239 9.6296 520 0.8177 0.3448 0.8177 0.9043
0.4239 9.6667 522 0.8363 0.3080 0.8363 0.9145
0.4239 9.7037 524 0.8506 0.3080 0.8506 0.9223
0.4239 9.7407 526 0.8612 0.3080 0.8612 0.9280
0.4239 9.7778 528 0.8642 0.3080 0.8642 0.9296
0.4239 9.8148 530 0.8647 0.3080 0.8647 0.9299
0.4239 9.8519 532 0.8584 0.3080 0.8584 0.9265
0.4239 9.8889 534 0.8484 0.3080 0.8484 0.9211
0.4239 9.9259 536 0.8408 0.3080 0.8408 0.9170
0.4239 9.9630 538 0.8358 0.3080 0.8358 0.9142
0.4239 10.0 540 0.8348 0.3080 0.8348 0.9136

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
3
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits5_FineTuningAraBERT_run1_AugV5_k9_task3_organization

Finetuned
(4023)
this model