ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k15_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7492
  • Qwk: 0.4352
  • Mse: 0.7492
  • Rmse: 0.8656

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0260 2 4.2755 0.0182 4.2755 2.0677
No log 0.0519 4 2.4329 0.0620 2.4329 1.5598
No log 0.0779 6 1.9397 0.0902 1.9397 1.3927
No log 0.1039 8 1.4443 0.1387 1.4443 1.2018
No log 0.1299 10 0.9560 0.2921 0.9560 0.9777
No log 0.1558 12 1.0948 0.1589 1.0948 1.0463
No log 0.1818 14 1.1206 0.1189 1.1206 1.0586
No log 0.2078 16 1.0662 0.1935 1.0662 1.0326
No log 0.2338 18 1.0922 0.0981 1.0922 1.0451
No log 0.2597 20 1.1381 0.0510 1.1381 1.0668
No log 0.2857 22 1.1532 0.1601 1.1532 1.0739
No log 0.3117 24 1.2032 0.1874 1.2032 1.0969
No log 0.3377 26 1.2410 0.0431 1.2410 1.1140
No log 0.3636 28 1.1670 0.1848 1.1670 1.0803
No log 0.3896 30 1.0835 0.1629 1.0835 1.0409
No log 0.4156 32 1.0949 0.1226 1.0949 1.0464
No log 0.4416 34 1.1309 0.1573 1.1309 1.0634
No log 0.4675 36 1.1457 0.2008 1.1457 1.0704
No log 0.4935 38 1.1285 0.2015 1.1285 1.0623
No log 0.5195 40 1.2605 0.0672 1.2605 1.1227
No log 0.5455 42 1.4870 -0.0709 1.4870 1.2194
No log 0.5714 44 1.4431 -0.0296 1.4431 1.2013
No log 0.5974 46 1.1561 0.1711 1.1561 1.0752
No log 0.6234 48 0.9640 0.2991 0.9640 0.9818
No log 0.6494 50 1.0064 0.2130 1.0064 1.0032
No log 0.6753 52 1.1276 0.2446 1.1276 1.0619
No log 0.7013 54 1.1249 0.2446 1.1249 1.0606
No log 0.7273 56 1.0293 0.3546 1.0293 1.0145
No log 0.7532 58 0.9167 0.3326 0.9167 0.9574
No log 0.7792 60 0.9290 0.4089 0.9290 0.9638
No log 0.8052 62 1.1358 0.3504 1.1358 1.0657
No log 0.8312 64 1.4814 0.2314 1.4814 1.2171
No log 0.8571 66 1.7441 0.1264 1.7441 1.3207
No log 0.8831 68 1.7215 0.1756 1.7215 1.3121
No log 0.9091 70 1.2726 0.2641 1.2726 1.1281
No log 0.9351 72 1.0600 0.3549 1.0600 1.0296
No log 0.9610 74 1.1153 0.3418 1.1153 1.0561
No log 0.9870 76 1.5309 0.2359 1.5309 1.2373
No log 1.0130 78 1.8380 0.0864 1.8380 1.3557
No log 1.0390 80 1.6708 -0.0192 1.6708 1.2926
No log 1.0649 82 1.4096 0.0811 1.4096 1.1873
No log 1.0909 84 1.1332 0.2310 1.1332 1.0645
No log 1.1169 86 1.0609 0.3077 1.0609 1.0300
No log 1.1429 88 0.9474 0.2960 0.9474 0.9733
No log 1.1688 90 0.9558 0.2494 0.9558 0.9776
No log 1.1948 92 1.1766 0.3478 1.1766 1.0847
No log 1.2208 94 1.6288 0.2441 1.6288 1.2763
No log 1.2468 96 1.6733 0.1531 1.6733 1.2936
No log 1.2727 98 1.3994 0.2952 1.3994 1.1830
No log 1.2987 100 1.2002 0.3383 1.2002 1.0955
No log 1.3247 102 0.9684 0.3527 0.9684 0.9841
No log 1.3506 104 0.8842 0.3629 0.8842 0.9403
No log 1.3766 106 0.8720 0.3629 0.8720 0.9338
No log 1.4026 108 0.9328 0.3596 0.9328 0.9658
No log 1.4286 110 1.0989 0.3409 1.0989 1.0483
No log 1.4545 112 1.0353 0.3409 1.0353 1.0175
No log 1.4805 114 0.8773 0.4342 0.8773 0.9367
No log 1.5065 116 0.8600 0.4465 0.8600 0.9274
No log 1.5325 118 0.8183 0.4696 0.8183 0.9046
No log 1.5584 120 0.8851 0.4681 0.8851 0.9408
No log 1.5844 122 1.0012 0.3881 1.0012 1.0006
No log 1.6104 124 0.8238 0.4894 0.8238 0.9076
No log 1.6364 126 0.6873 0.4981 0.6873 0.8291
No log 1.6623 128 0.6709 0.5735 0.6709 0.8191
No log 1.6883 130 0.6986 0.5306 0.6986 0.8358
No log 1.7143 132 0.8057 0.5219 0.8057 0.8976
No log 1.7403 134 0.7879 0.5020 0.7879 0.8876
No log 1.7662 136 0.6977 0.5098 0.6977 0.8353
No log 1.7922 138 0.6755 0.4537 0.6755 0.8219
No log 1.8182 140 0.6755 0.4626 0.6755 0.8219
No log 1.8442 142 0.7583 0.5279 0.7583 0.8708
No log 1.8701 144 0.8132 0.4783 0.8132 0.9018
No log 1.8961 146 0.7783 0.5360 0.7783 0.8822
No log 1.9221 148 0.7921 0.4789 0.7921 0.8900
No log 1.9481 150 0.8776 0.4548 0.8776 0.9368
No log 1.9740 152 1.0216 0.3967 1.0216 1.0107
No log 2.0 154 0.9299 0.3216 0.9299 0.9643
No log 2.0260 156 0.7856 0.2773 0.7856 0.8863
No log 2.0519 158 0.7521 0.3063 0.7521 0.8672
No log 2.0779 160 0.7867 0.3511 0.7867 0.8870
No log 2.1039 162 0.7645 0.5167 0.7645 0.8743
No log 2.1299 164 0.7882 0.5572 0.7882 0.8878
No log 2.1558 166 0.8242 0.4810 0.8242 0.9078
No log 2.1818 168 0.7890 0.5057 0.7890 0.8883
No log 2.2078 170 0.8410 0.4050 0.8410 0.9171
No log 2.2338 172 0.8229 0.3948 0.8229 0.9071
No log 2.2597 174 0.8427 0.4169 0.8427 0.9180
No log 2.2857 176 0.9367 0.3913 0.9367 0.9678
No log 2.3117 178 1.0126 0.4130 1.0126 1.0063
No log 2.3377 180 1.0049 0.4141 1.0049 1.0024
No log 2.3636 182 1.0429 0.4012 1.0429 1.0212
No log 2.3896 184 1.0136 0.4012 1.0136 1.0068
No log 2.4156 186 0.8769 0.4681 0.8769 0.9364
No log 2.4416 188 0.7659 0.4063 0.7659 0.8751
No log 2.4675 190 0.7603 0.4606 0.7603 0.8719
No log 2.4935 192 0.7713 0.4606 0.7713 0.8782
No log 2.5195 194 0.7538 0.4428 0.7538 0.8682
No log 2.5455 196 0.8603 0.4821 0.8603 0.9275
No log 2.5714 198 0.9210 0.4801 0.9210 0.9597
No log 2.5974 200 0.9223 0.4570 0.9223 0.9604
No log 2.6234 202 0.8189 0.3780 0.8189 0.9050
No log 2.6494 204 0.8064 0.2787 0.8064 0.8980
No log 2.6753 206 0.7977 0.3200 0.7977 0.8932
No log 2.7013 208 0.8339 0.4318 0.8339 0.9132
No log 2.7273 210 0.9157 0.4932 0.9157 0.9569
No log 2.7532 212 0.9265 0.4926 0.9265 0.9626
No log 2.7792 214 0.9004 0.4937 0.9004 0.9489
No log 2.8052 216 0.8662 0.4937 0.8662 0.9307
No log 2.8312 218 0.7990 0.3521 0.7990 0.8939
No log 2.8571 220 0.8105 0.5062 0.8105 0.9003
No log 2.8831 222 0.8743 0.4696 0.8743 0.9350
No log 2.9091 224 0.9918 0.3969 0.9918 0.9959
No log 2.9351 226 0.9124 0.4536 0.9124 0.9552
No log 2.9610 228 0.8102 0.3902 0.8102 0.9001
No log 2.9870 230 0.7748 0.3959 0.7748 0.8802
No log 3.0130 232 0.8060 0.4712 0.8060 0.8978
No log 3.0390 234 0.9473 0.4775 0.9473 0.9733
No log 3.0649 236 1.0399 0.4444 1.0399 1.0198
No log 3.0909 238 0.9266 0.5013 0.9266 0.9626
No log 3.1169 240 0.7789 0.4836 0.7789 0.8826
No log 3.1429 242 0.7458 0.3821 0.7458 0.8636
No log 3.1688 244 0.7744 0.4180 0.7744 0.8800
No log 3.1948 246 0.8025 0.5042 0.8025 0.8958
No log 3.2208 248 0.8421 0.4681 0.8421 0.9176
No log 3.2468 250 0.8198 0.5443 0.8198 0.9054
No log 3.2727 252 0.8836 0.5115 0.8836 0.9400
No log 3.2987 254 0.9017 0.4885 0.9017 0.9496
No log 3.3247 256 0.8162 0.3989 0.8162 0.9034
No log 3.3506 258 0.7797 0.3590 0.7797 0.8830
No log 3.3766 260 0.7575 0.3922 0.7575 0.8704
No log 3.4026 262 0.7203 0.4133 0.7203 0.8487
No log 3.4286 264 0.7064 0.4691 0.7064 0.8405
No log 3.4545 266 0.7178 0.5186 0.7178 0.8472
No log 3.4805 268 0.8311 0.5027 0.8311 0.9117
No log 3.5065 270 0.8571 0.4902 0.8571 0.9258
No log 3.5325 272 0.7545 0.4450 0.7545 0.8686
No log 3.5584 274 0.7213 0.4204 0.7213 0.8493
No log 3.5844 276 0.7512 0.4078 0.7512 0.8667
No log 3.6104 278 0.8043 0.4937 0.8043 0.8968
No log 3.6364 280 0.8561 0.4555 0.8561 0.9253
No log 3.6623 282 0.8919 0.4444 0.8919 0.9444
No log 3.6883 284 0.8173 0.4801 0.8173 0.9040
No log 3.7143 286 0.7557 0.4836 0.7557 0.8693
No log 3.7403 288 0.7025 0.5552 0.7025 0.8381
No log 3.7662 290 0.7091 0.5552 0.7091 0.8421
No log 3.7922 292 0.7722 0.4932 0.7722 0.8788
No log 3.8182 294 0.8212 0.4917 0.8212 0.9062
No log 3.8442 296 0.7846 0.3902 0.7846 0.8858
No log 3.8701 298 0.7281 0.3797 0.7281 0.8533
No log 3.8961 300 0.7099 0.4624 0.7099 0.8425
No log 3.9221 302 0.7087 0.4516 0.7087 0.8418
No log 3.9481 304 0.7140 0.4391 0.7140 0.8450
No log 3.9740 306 0.7106 0.4659 0.7106 0.8430
No log 4.0 308 0.7126 0.4887 0.7126 0.8441
No log 4.0260 310 0.7227 0.4353 0.7227 0.8501
No log 4.0519 312 0.7221 0.4524 0.7221 0.8498
No log 4.0779 314 0.8074 0.4866 0.8074 0.8985
No log 4.1039 316 0.8533 0.5020 0.8533 0.9238
No log 4.1299 318 0.7854 0.4935 0.7854 0.8862
No log 4.1558 320 0.7694 0.3942 0.7694 0.8772
No log 4.1818 322 0.7679 0.3797 0.7679 0.8763
No log 4.2078 324 0.7504 0.4241 0.7504 0.8663
No log 4.2338 326 0.7365 0.4378 0.7365 0.8582
No log 4.2597 328 0.7064 0.4645 0.7064 0.8405
No log 4.2857 330 0.6819 0.4789 0.6819 0.8258
No log 4.3117 332 0.6708 0.5234 0.6708 0.8190
No log 4.3377 334 0.7325 0.5591 0.7325 0.8559
No log 4.3636 336 0.7703 0.5591 0.7703 0.8777
No log 4.3896 338 0.7377 0.5610 0.7377 0.8589
No log 4.4156 340 0.7352 0.5642 0.7352 0.8574
No log 4.4416 342 0.7122 0.5220 0.7122 0.8439
No log 4.4675 344 0.7239 0.5199 0.7239 0.8508
No log 4.4935 346 0.7855 0.5291 0.7855 0.8863
No log 4.5195 348 0.7972 0.4711 0.7972 0.8929
No log 4.5455 350 0.7743 0.4712 0.7743 0.8800
No log 4.5714 352 0.7498 0.3403 0.7498 0.8659
No log 4.5974 354 0.7633 0.3243 0.7633 0.8737
No log 4.6234 356 0.7727 0.3652 0.7727 0.8790
No log 4.6494 358 0.9001 0.3310 0.9001 0.9487
No log 4.6753 360 1.0103 0.3744 1.0103 1.0051
No log 4.7013 362 0.9372 0.3985 0.9372 0.9681
No log 4.7273 364 0.8051 0.4568 0.8051 0.8972
No log 4.7532 366 0.7516 0.5065 0.7516 0.8670
No log 4.7792 368 0.7626 0.5147 0.7626 0.8732
No log 4.8052 370 0.8003 0.5147 0.8003 0.8946
No log 4.8312 372 0.7985 0.5033 0.7985 0.8936
No log 4.8571 374 0.7508 0.5540 0.7508 0.8665
No log 4.8831 376 0.7469 0.4576 0.7469 0.8642
No log 4.9091 378 0.7747 0.4576 0.7747 0.8802
No log 4.9351 380 0.7901 0.4576 0.7901 0.8889
No log 4.9610 382 0.8967 0.3985 0.8967 0.9469
No log 4.9870 384 0.9244 0.4212 0.9244 0.9615
No log 5.0130 386 0.9143 0.4212 0.9143 0.9562
No log 5.0390 388 0.8668 0.3695 0.8668 0.9310
No log 5.0649 390 0.8832 0.4180 0.8832 0.9398
No log 5.0909 392 0.9889 0.4192 0.9889 0.9944
No log 5.1169 394 0.9882 0.4192 0.9882 0.9941
No log 5.1429 396 0.9804 0.4302 0.9804 0.9902
No log 5.1688 398 0.9548 0.4284 0.9548 0.9771
No log 5.1948 400 0.9379 0.3766 0.9379 0.9685
No log 5.2208 402 0.8581 0.3063 0.8581 0.9263
No log 5.2468 404 0.8348 0.3063 0.8348 0.9137
No log 5.2727 406 0.7774 0.3351 0.7774 0.8817
No log 5.2987 408 0.7048 0.4746 0.7048 0.8395
No log 5.3247 410 0.6706 0.5138 0.6706 0.8189
No log 5.3506 412 0.6349 0.5663 0.6349 0.7968
No log 5.3766 414 0.6123 0.6117 0.6123 0.7825
No log 5.4026 416 0.6049 0.6750 0.6049 0.7778
No log 5.4286 418 0.6298 0.6469 0.6298 0.7936
No log 5.4545 420 0.6521 0.6614 0.6521 0.8075
No log 5.4805 422 0.6230 0.6325 0.6230 0.7893
No log 5.5065 424 0.6244 0.6291 0.6244 0.7902
No log 5.5325 426 0.6413 0.6198 0.6413 0.8008
No log 5.5584 428 0.6360 0.6510 0.6360 0.7975
No log 5.5844 430 0.6304 0.5614 0.6304 0.7940
No log 5.6104 432 0.6298 0.5820 0.6298 0.7936
No log 5.6364 434 0.6494 0.5923 0.6494 0.8058
No log 5.6623 436 0.6433 0.4883 0.6433 0.8020
No log 5.6883 438 0.6513 0.5407 0.6513 0.8070
No log 5.7143 440 0.6822 0.5607 0.6822 0.8260
No log 5.7403 442 0.6759 0.4888 0.6759 0.8221
No log 5.7662 444 0.6732 0.5536 0.6732 0.8205
No log 5.7922 446 0.6882 0.4868 0.6882 0.8296
No log 5.8182 448 0.7340 0.4903 0.7340 0.8567
No log 5.8442 450 0.7088 0.4771 0.7088 0.8419
No log 5.8701 452 0.6609 0.5063 0.6609 0.8130
No log 5.8961 454 0.6952 0.5823 0.6952 0.8338
No log 5.9221 456 0.7288 0.6143 0.7288 0.8537
No log 5.9481 458 0.7484 0.5992 0.7484 0.8651
No log 5.9740 460 0.8202 0.5253 0.8202 0.9056
No log 6.0 462 0.8116 0.5675 0.8116 0.9009
No log 6.0260 464 0.7474 0.6455 0.7474 0.8645
No log 6.0519 466 0.6959 0.6272 0.6959 0.8342
No log 6.0779 468 0.7158 0.6272 0.7158 0.8461
No log 6.1039 470 0.8028 0.4794 0.8028 0.8960
No log 6.1299 472 0.9672 0.4655 0.9672 0.9835
No log 6.1558 474 0.9862 0.4655 0.9862 0.9931
No log 6.1818 476 0.8956 0.4554 0.8956 0.9464
No log 6.2078 478 0.7666 0.5048 0.7666 0.8756
No log 6.2338 480 0.7020 0.4745 0.7020 0.8378
No log 6.2597 482 0.6843 0.4760 0.6843 0.8272
No log 6.2857 484 0.6862 0.4903 0.6862 0.8283
No log 6.3117 486 0.7098 0.4745 0.7098 0.8425
No log 6.3377 488 0.7131 0.5459 0.7131 0.8445
No log 6.3636 490 0.7281 0.6188 0.7281 0.8533
No log 6.3896 492 0.7193 0.6188 0.7193 0.8481
No log 6.4156 494 0.6839 0.5774 0.6839 0.8270
No log 6.4416 496 0.6839 0.5975 0.6839 0.8270
No log 6.4675 498 0.7452 0.5728 0.7452 0.8632
0.3056 6.4935 500 0.7975 0.5137 0.7975 0.8930
0.3056 6.5195 502 0.8379 0.4799 0.8379 0.9154
0.3056 6.5455 504 0.8336 0.4098 0.8336 0.9130
0.3056 6.5714 506 0.8171 0.3551 0.8171 0.9040
0.3056 6.5974 508 0.8049 0.4455 0.8049 0.8972
0.3056 6.6234 510 0.7492 0.4352 0.7492 0.8656

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k15_task5_organization

Finetuned
(4019)
this model