ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k6_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5686
  • Qwk: 0.6275
  • Mse: 0.5686
  • Rmse: 0.7540

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0667 2 3.9682 -0.0297 3.9682 1.9920
No log 0.1333 4 2.0966 0.0970 2.0966 1.4480
No log 0.2 6 2.2394 0.0878 2.2394 1.4964
No log 0.2667 8 1.7856 0.0909 1.7856 1.3363
No log 0.3333 10 1.1180 0.1644 1.1180 1.0574
No log 0.4 12 1.0235 0.2441 1.0235 1.0117
No log 0.4667 14 0.9625 0.2094 0.9625 0.9811
No log 0.5333 16 1.0704 0.2065 1.0704 1.0346
No log 0.6 18 1.1501 0.2359 1.1501 1.0724
No log 0.6667 20 0.9310 0.3583 0.9310 0.9649
No log 0.7333 22 0.7030 0.5862 0.7030 0.8385
No log 0.8 24 0.6837 0.5822 0.6837 0.8269
No log 0.8667 26 0.7342 0.5085 0.7342 0.8568
No log 0.9333 28 1.1201 0.3863 1.1201 1.0584
No log 1.0 30 1.0594 0.4379 1.0594 1.0293
No log 1.0667 32 0.7297 0.5716 0.7297 0.8542
No log 1.1333 34 0.6543 0.6352 0.6543 0.8089
No log 1.2 36 0.7204 0.6372 0.7204 0.8488
No log 1.2667 38 0.8955 0.6296 0.8955 0.9463
No log 1.3333 40 1.0728 0.4476 1.0728 1.0357
No log 1.4 42 0.9037 0.5770 0.9037 0.9506
No log 1.4667 44 0.6705 0.5807 0.6705 0.8188
No log 1.5333 46 0.7014 0.6993 0.7014 0.8375
No log 1.6 48 0.7129 0.6961 0.7129 0.8443
No log 1.6667 50 0.6423 0.7050 0.6423 0.8015
No log 1.7333 52 0.6768 0.6245 0.6768 0.8227
No log 1.8 54 0.6963 0.6371 0.6963 0.8344
No log 1.8667 56 0.6116 0.5910 0.6116 0.7821
No log 1.9333 58 0.7241 0.5563 0.7241 0.8510
No log 2.0 60 1.0037 0.5448 1.0037 1.0019
No log 2.0667 62 1.2672 0.4236 1.2672 1.1257
No log 2.1333 64 0.9991 0.5436 0.9991 0.9995
No log 2.2 66 0.7289 0.6469 0.7289 0.8538
No log 2.2667 68 0.6563 0.6148 0.6563 0.8101
No log 2.3333 70 0.7095 0.6513 0.7095 0.8423
No log 2.4 72 0.7393 0.6312 0.7393 0.8598
No log 2.4667 74 0.7195 0.6165 0.7195 0.8483
No log 2.5333 76 0.6516 0.6298 0.6516 0.8072
No log 2.6 78 0.9261 0.5432 0.9261 0.9624
No log 2.6667 80 1.0402 0.4571 1.0402 1.0199
No log 2.7333 82 0.8377 0.5801 0.8377 0.9153
No log 2.8 84 0.6500 0.5945 0.6500 0.8062
No log 2.8667 86 0.6505 0.5574 0.6505 0.8065
No log 2.9333 88 0.6556 0.5747 0.6556 0.8097
No log 3.0 90 0.6647 0.6205 0.6647 0.8153
No log 3.0667 92 0.6775 0.6179 0.6775 0.8231
No log 3.1333 94 0.6752 0.5911 0.6752 0.8217
No log 3.2 96 0.6500 0.5830 0.6500 0.8062
No log 3.2667 98 0.6379 0.6011 0.6379 0.7987
No log 3.3333 100 0.6546 0.6052 0.6546 0.8091
No log 3.4 102 0.6259 0.6263 0.6259 0.7912
No log 3.4667 104 0.6132 0.6343 0.6132 0.7830
No log 3.5333 106 0.6321 0.5941 0.6321 0.7951
No log 3.6 108 0.7360 0.6570 0.7360 0.8579
No log 3.6667 110 0.6732 0.6733 0.6732 0.8205
No log 3.7333 112 0.6003 0.6468 0.6003 0.7748
No log 3.8 114 0.6327 0.6419 0.6327 0.7954
No log 3.8667 116 0.6131 0.6357 0.6131 0.7830
No log 3.9333 118 0.6061 0.6672 0.6061 0.7785
No log 4.0 120 0.6867 0.6453 0.6867 0.8287
No log 4.0667 122 0.6935 0.6595 0.6935 0.8328
No log 4.1333 124 0.6560 0.6775 0.6560 0.8099
No log 4.2 126 0.6275 0.6187 0.6275 0.7921
No log 4.2667 128 0.6192 0.6503 0.6192 0.7869
No log 4.3333 130 0.6226 0.6570 0.6226 0.7891
No log 4.4 132 0.6339 0.6047 0.6339 0.7962
No log 4.4667 134 0.6284 0.6314 0.6284 0.7927
No log 4.5333 136 0.6665 0.5945 0.6665 0.8164
No log 4.6 138 0.6993 0.5828 0.6993 0.8363
No log 4.6667 140 0.6772 0.6094 0.6772 0.8229
No log 4.7333 142 0.7068 0.6107 0.7068 0.8407
No log 4.8 144 0.7302 0.6107 0.7302 0.8545
No log 4.8667 146 0.6879 0.6107 0.6879 0.8294
No log 4.9333 148 0.6892 0.5930 0.6892 0.8302
No log 5.0 150 0.7080 0.6023 0.7080 0.8414
No log 5.0667 152 0.6438 0.6275 0.6438 0.8024
No log 5.1333 154 0.6364 0.5819 0.6364 0.7977
No log 5.2 156 0.6577 0.6246 0.6577 0.8110
No log 5.2667 158 0.6696 0.6328 0.6696 0.8183
No log 5.3333 160 0.6646 0.6269 0.6646 0.8152
No log 5.4 162 0.8611 0.6235 0.8611 0.9280
No log 5.4667 164 0.9759 0.5503 0.9759 0.9879
No log 5.5333 166 0.8883 0.5971 0.8883 0.9425
No log 5.6 168 0.6870 0.5963 0.6870 0.8288
No log 5.6667 170 0.6656 0.6259 0.6656 0.8158
No log 5.7333 172 0.6672 0.6278 0.6672 0.8168
No log 5.8 174 0.7938 0.5628 0.7938 0.8910
No log 5.8667 176 0.9756 0.4346 0.9756 0.9877
No log 5.9333 178 0.9268 0.5068 0.9268 0.9627
No log 6.0 180 0.7272 0.6071 0.7272 0.8528
No log 6.0667 182 0.6279 0.6684 0.6279 0.7924
No log 6.1333 184 0.6180 0.6365 0.6180 0.7861
No log 6.2 186 0.6170 0.6110 0.6170 0.7855
No log 6.2667 188 0.6238 0.6157 0.6238 0.7898
No log 6.3333 190 0.6173 0.5774 0.6173 0.7857
No log 6.4 192 0.6234 0.5774 0.6234 0.7896
No log 6.4667 194 0.5993 0.6311 0.5993 0.7741
No log 6.5333 196 0.6085 0.6519 0.6085 0.7800
No log 6.6 198 0.6510 0.6245 0.6510 0.8068
No log 6.6667 200 0.6391 0.6519 0.6391 0.7994
No log 6.7333 202 0.6456 0.6337 0.6456 0.8035
No log 6.8 204 0.6819 0.6228 0.6819 0.8258
No log 6.8667 206 0.6917 0.6169 0.6917 0.8317
No log 6.9333 208 0.6909 0.6142 0.6909 0.8312
No log 7.0 210 0.6667 0.5990 0.6667 0.8165
No log 7.0667 212 0.6715 0.6085 0.6715 0.8194
No log 7.1333 214 0.6787 0.6092 0.6787 0.8239
No log 7.2 216 0.8185 0.5305 0.8185 0.9047
No log 7.2667 218 1.0974 0.4577 1.0974 1.0476
No log 7.3333 220 1.1207 0.4177 1.1207 1.0586
No log 7.4 222 0.9334 0.5142 0.9334 0.9661
No log 7.4667 224 0.6860 0.6245 0.6860 0.8282
No log 7.5333 226 0.6087 0.6227 0.6087 0.7802
No log 7.6 228 0.6326 0.5926 0.6326 0.7954
No log 7.6667 230 0.6211 0.6242 0.6211 0.7881
No log 7.7333 232 0.6321 0.6828 0.6321 0.7950
No log 7.8 234 0.6331 0.6479 0.6331 0.7956
No log 7.8667 236 0.6883 0.6822 0.6883 0.8296
No log 7.9333 238 0.7350 0.6530 0.7350 0.8573
No log 8.0 240 0.6419 0.6354 0.6419 0.8012
No log 8.0667 242 0.5927 0.6782 0.5927 0.7699
No log 8.1333 244 0.6167 0.7128 0.6167 0.7853
No log 8.2 246 0.6718 0.6430 0.6718 0.8196
No log 8.2667 248 0.7106 0.6312 0.7106 0.8429
No log 8.3333 250 0.7021 0.6189 0.7021 0.8379
No log 8.4 252 0.6533 0.6528 0.6533 0.8083
No log 8.4667 254 0.6316 0.6397 0.6316 0.7947
No log 8.5333 256 0.6230 0.6104 0.6230 0.7893
No log 8.6 258 0.6178 0.5863 0.6178 0.7860
No log 8.6667 260 0.6250 0.5660 0.6250 0.7905
No log 8.7333 262 0.6374 0.6100 0.6374 0.7984
No log 8.8 264 0.6436 0.6047 0.6436 0.8022
No log 8.8667 266 0.6093 0.6721 0.6093 0.7806
No log 8.9333 268 0.6172 0.6701 0.6172 0.7856
No log 9.0 270 0.6106 0.6701 0.6106 0.7814
No log 9.0667 272 0.5875 0.6510 0.5875 0.7665
No log 9.1333 274 0.6142 0.6007 0.6142 0.7837
No log 9.2 276 0.6608 0.5809 0.6608 0.8129
No log 9.2667 278 0.7268 0.6115 0.7268 0.8526
No log 9.3333 280 0.7489 0.6198 0.7489 0.8654
No log 9.4 282 0.7187 0.6026 0.7187 0.8478
No log 9.4667 284 0.6687 0.6435 0.6687 0.8177
No log 9.5333 286 0.6779 0.5905 0.6779 0.8233
No log 9.6 288 0.6853 0.6080 0.6853 0.8278
No log 9.6667 290 0.6520 0.6302 0.6520 0.8074
No log 9.7333 292 0.6191 0.6364 0.6191 0.7868
No log 9.8 294 0.6328 0.5954 0.6328 0.7955
No log 9.8667 296 0.6705 0.6038 0.6705 0.8188
No log 9.9333 298 0.7082 0.5653 0.7082 0.8415
No log 10.0 300 0.6591 0.6289 0.6591 0.8119
No log 10.0667 302 0.5826 0.6634 0.5826 0.7633
No log 10.1333 304 0.5844 0.5982 0.5844 0.7645
No log 10.2 306 0.6055 0.5993 0.6055 0.7781
No log 10.2667 308 0.6807 0.6338 0.6807 0.8250
No log 10.3333 310 0.7168 0.6274 0.7168 0.8466
No log 10.4 312 0.6341 0.6695 0.6341 0.7963
No log 10.4667 314 0.5836 0.7174 0.5836 0.7640
No log 10.5333 316 0.6411 0.6662 0.6411 0.8007
No log 10.6 318 0.6769 0.6717 0.6769 0.8227
No log 10.6667 320 0.6506 0.6092 0.6506 0.8066
No log 10.7333 322 0.6267 0.6014 0.6267 0.7916
No log 10.8 324 0.6056 0.5570 0.6056 0.7782
No log 10.8667 326 0.6233 0.6028 0.6233 0.7895
No log 10.9333 328 0.6393 0.6696 0.6393 0.7995
No log 11.0 330 0.6979 0.6655 0.6979 0.8354
No log 11.0667 332 0.6725 0.6481 0.6725 0.8201
No log 11.1333 334 0.6285 0.6815 0.6285 0.7928
No log 11.2 336 0.5875 0.6133 0.5875 0.7665
No log 11.2667 338 0.5851 0.6548 0.5851 0.7649
No log 11.3333 340 0.6463 0.6926 0.6463 0.8039
No log 11.4 342 0.7426 0.6080 0.7426 0.8617
No log 11.4667 344 0.6879 0.6749 0.6879 0.8294
No log 11.5333 346 0.5920 0.6672 0.5920 0.7694
No log 11.6 348 0.5847 0.6284 0.5847 0.7647
No log 11.6667 350 0.5826 0.6284 0.5826 0.7633
No log 11.7333 352 0.5918 0.6602 0.5918 0.7693
No log 11.8 354 0.6395 0.6581 0.6395 0.7997
No log 11.8667 356 0.6519 0.6275 0.6519 0.8074
No log 11.9333 358 0.6014 0.6695 0.6014 0.7755
No log 12.0 360 0.5824 0.6284 0.5824 0.7631
No log 12.0667 362 0.6055 0.6779 0.6055 0.7782
No log 12.1333 364 0.5890 0.6835 0.5890 0.7675
No log 12.2 366 0.5898 0.6365 0.5898 0.7680
No log 12.2667 368 0.6130 0.6479 0.6130 0.7830
No log 12.3333 370 0.6051 0.6479 0.6051 0.7779
No log 12.4 372 0.6025 0.6380 0.6025 0.7762
No log 12.4667 374 0.5980 0.6335 0.5980 0.7733
No log 12.5333 376 0.6090 0.6133 0.6090 0.7804
No log 12.6 378 0.6332 0.6404 0.6332 0.7957
No log 12.6667 380 0.6301 0.6705 0.6301 0.7938
No log 12.7333 382 0.6131 0.6188 0.6131 0.7830
No log 12.8 384 0.6084 0.5606 0.6084 0.7800
No log 12.8667 386 0.6182 0.5626 0.6182 0.7863
No log 12.9333 388 0.6391 0.5847 0.6391 0.7994
No log 13.0 390 0.7058 0.5442 0.7058 0.8401
No log 13.0667 392 0.7595 0.4932 0.7595 0.8715
No log 13.1333 394 0.7487 0.5255 0.7487 0.8653
No log 13.2 396 0.6889 0.5909 0.6889 0.8300
No log 13.2667 398 0.6365 0.6217 0.6365 0.7978
No log 13.3333 400 0.6365 0.6176 0.6365 0.7978
No log 13.4 402 0.6513 0.5669 0.6513 0.8070
No log 13.4667 404 0.6405 0.6035 0.6405 0.8003
No log 13.5333 406 0.6245 0.6310 0.6245 0.7902
No log 13.6 408 0.6092 0.6011 0.6092 0.7805
No log 13.6667 410 0.6074 0.5602 0.6074 0.7794
No log 13.7333 412 0.6132 0.5690 0.6132 0.7831
No log 13.8 414 0.6237 0.6133 0.6237 0.7897
No log 13.8667 416 0.6460 0.6133 0.6460 0.8037
No log 13.9333 418 0.6306 0.6133 0.6306 0.7941
No log 14.0 420 0.5998 0.5808 0.5998 0.7745
No log 14.0667 422 0.5954 0.6022 0.5954 0.7716
No log 14.1333 424 0.6224 0.5969 0.6224 0.7889
No log 14.2 426 0.6310 0.6556 0.6310 0.7944
No log 14.2667 428 0.6174 0.6549 0.6174 0.7857
No log 14.3333 430 0.6098 0.6318 0.6098 0.7809
No log 14.4 432 0.6113 0.6102 0.6113 0.7818
No log 14.4667 434 0.6175 0.6634 0.6175 0.7858
No log 14.5333 436 0.6202 0.6104 0.6202 0.7876
No log 14.6 438 0.6135 0.6057 0.6135 0.7832
No log 14.6667 440 0.6134 0.5523 0.6134 0.7832
No log 14.7333 442 0.6080 0.5724 0.6080 0.7798
No log 14.8 444 0.5940 0.5381 0.5940 0.7707
No log 14.8667 446 0.5996 0.5865 0.5996 0.7743
No log 14.9333 448 0.5984 0.5865 0.5984 0.7736
No log 15.0 450 0.5905 0.5650 0.5905 0.7684
No log 15.0667 452 0.5950 0.5650 0.5950 0.7714
No log 15.1333 454 0.5945 0.5630 0.5945 0.7710
No log 15.2 456 0.5998 0.5982 0.5998 0.7745
No log 15.2667 458 0.5866 0.5806 0.5866 0.7659
No log 15.3333 460 0.5743 0.6464 0.5743 0.7578
No log 15.4 462 0.5844 0.6380 0.5844 0.7644
No log 15.4667 464 0.5805 0.6046 0.5805 0.7619
No log 15.5333 466 0.5885 0.5845 0.5885 0.7671
No log 15.6 468 0.6073 0.5480 0.6073 0.7793
No log 15.6667 470 0.6040 0.6249 0.6040 0.7772
No log 15.7333 472 0.5991 0.6330 0.5991 0.7740
No log 15.8 474 0.6046 0.6345 0.6046 0.7775
No log 15.8667 476 0.5976 0.6206 0.5976 0.7730
No log 15.9333 478 0.5936 0.6433 0.5936 0.7704
No log 16.0 480 0.5983 0.6433 0.5983 0.7735
No log 16.0667 482 0.6267 0.6519 0.6267 0.7917
No log 16.1333 484 0.6497 0.6199 0.6497 0.8060
No log 16.2 486 0.6481 0.6529 0.6481 0.8050
No log 16.2667 488 0.6347 0.5585 0.6347 0.7967
No log 16.3333 490 0.6147 0.5960 0.6147 0.7840
No log 16.4 492 0.5888 0.5735 0.5888 0.7673
No log 16.4667 494 0.5586 0.5831 0.5586 0.7474
No log 16.5333 496 0.5363 0.5898 0.5363 0.7323
No log 16.6 498 0.5329 0.6297 0.5329 0.7300
0.2407 16.6667 500 0.5507 0.6398 0.5507 0.7421
0.2407 16.7333 502 0.5946 0.7050 0.5946 0.7711
0.2407 16.8 504 0.6004 0.6743 0.6004 0.7748
0.2407 16.8667 506 0.5576 0.6732 0.5576 0.7467
0.2407 16.9333 508 0.5241 0.6680 0.5241 0.7239
0.2407 17.0 510 0.5364 0.6536 0.5364 0.7324
0.2407 17.0667 512 0.5724 0.6194 0.5724 0.7566
0.2407 17.1333 514 0.5866 0.6194 0.5866 0.7659
0.2407 17.2 516 0.5536 0.6288 0.5536 0.7441
0.2407 17.2667 518 0.5352 0.6339 0.5352 0.7316
0.2407 17.3333 520 0.5404 0.6720 0.5404 0.7351
0.2407 17.4 522 0.5702 0.6584 0.5702 0.7551
0.2407 17.4667 524 0.6158 0.6537 0.6158 0.7848
0.2407 17.5333 526 0.6315 0.6479 0.6315 0.7947
0.2407 17.6 528 0.6092 0.6380 0.6092 0.7805
0.2407 17.6667 530 0.5874 0.6780 0.5874 0.7664
0.2407 17.7333 532 0.5905 0.6406 0.5905 0.7684
0.2407 17.8 534 0.5875 0.6406 0.5875 0.7665
0.2407 17.8667 536 0.5802 0.6310 0.5802 0.7617
0.2407 17.9333 538 0.5811 0.6491 0.5811 0.7623
0.2407 18.0 540 0.5789 0.6105 0.5789 0.7609
0.2407 18.0667 542 0.5799 0.6164 0.5799 0.7615
0.2407 18.1333 544 0.5875 0.6089 0.5875 0.7665
0.2407 18.2 546 0.5778 0.6134 0.5778 0.7602
0.2407 18.2667 548 0.5686 0.6275 0.5686 0.7540

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k6_task5_organization

Finetuned
(4019)
this model