ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k8_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7658
  • Qwk: 0.5045
  • Mse: 0.7658
  • Rmse: 0.8751

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.05 2 4.3560 0.0062 4.3560 2.0871
No log 0.1 4 2.2327 0.0189 2.2327 1.4942
No log 0.15 6 1.4606 0.0 1.4606 1.2085
No log 0.2 8 1.1168 0.2465 1.1168 1.0568
No log 0.25 10 1.0066 0.2643 1.0066 1.0033
No log 0.3 12 1.2747 0.0682 1.2747 1.1290
No log 0.35 14 1.1443 0.2448 1.1443 1.0697
No log 0.4 16 1.1786 0.1443 1.1786 1.0857
No log 0.45 18 1.0046 0.3222 1.0046 1.0023
No log 0.5 20 0.9255 0.3693 0.9255 0.9620
No log 0.55 22 0.8801 0.3996 0.8801 0.9382
No log 0.6 24 1.0609 0.3776 1.0609 1.0300
No log 0.65 26 1.0509 0.4241 1.0509 1.0251
No log 0.7 28 0.8259 0.4553 0.8259 0.9088
No log 0.75 30 0.7378 0.5304 0.7378 0.8590
No log 0.8 32 0.6772 0.5969 0.6772 0.8229
No log 0.85 34 0.6877 0.6005 0.6877 0.8293
No log 0.9 36 0.8198 0.5248 0.8198 0.9054
No log 0.95 38 0.6929 0.6260 0.6929 0.8324
No log 1.0 40 0.6721 0.6424 0.6721 0.8198
No log 1.05 42 0.8026 0.6183 0.8026 0.8959
No log 1.1 44 1.0044 0.5455 1.0044 1.0022
No log 1.15 46 0.7331 0.6295 0.7331 0.8562
No log 1.2 48 0.7349 0.6549 0.7349 0.8573
No log 1.25 50 0.7343 0.6388 0.7343 0.8569
No log 1.3 52 0.8056 0.6457 0.8056 0.8976
No log 1.35 54 0.7710 0.6402 0.7710 0.8781
No log 1.4 56 0.7133 0.6681 0.7133 0.8446
No log 1.45 58 0.7254 0.6697 0.7254 0.8517
No log 1.5 60 0.6371 0.6408 0.6371 0.7982
No log 1.55 62 0.6468 0.6144 0.6468 0.8042
No log 1.6 64 1.0127 0.4802 1.0127 1.0064
No log 1.65 66 1.2197 0.3936 1.2197 1.1044
No log 1.7 68 0.9416 0.5295 0.9416 0.9703
No log 1.75 70 0.6112 0.6796 0.6112 0.7818
No log 1.8 72 0.6481 0.6314 0.6481 0.8050
No log 1.85 74 0.6396 0.6644 0.6396 0.7997
No log 1.9 76 0.6594 0.6224 0.6594 0.8120
No log 1.95 78 0.6491 0.6578 0.6491 0.8057
No log 2.0 80 0.7198 0.6473 0.7198 0.8484
No log 2.05 82 0.6643 0.6368 0.6643 0.8151
No log 2.1 84 0.6413 0.6581 0.6413 0.8008
No log 2.15 86 0.6706 0.6174 0.6706 0.8189
No log 2.2 88 1.0719 0.4640 1.0719 1.0353
No log 2.25 90 1.2704 0.3754 1.2704 1.1271
No log 2.3 92 0.9022 0.5165 0.9022 0.9499
No log 2.35 94 0.6557 0.6526 0.6557 0.8098
No log 2.4 96 0.7436 0.5788 0.7436 0.8623
No log 2.45 98 0.7004 0.6520 0.7004 0.8369
No log 2.5 100 0.7576 0.5629 0.7576 0.8704
No log 2.55 102 0.8945 0.5054 0.8945 0.9458
No log 2.6 104 0.7731 0.5123 0.7731 0.8793
No log 2.65 106 0.6833 0.6310 0.6833 0.8266
No log 2.7 108 0.6831 0.6406 0.6831 0.8265
No log 2.75 110 0.6621 0.6638 0.6621 0.8137
No log 2.8 112 0.6641 0.6804 0.6641 0.8149
No log 2.85 114 0.6437 0.6683 0.6437 0.8023
No log 2.9 116 0.6221 0.6659 0.6221 0.7887
No log 2.95 118 0.6564 0.6845 0.6564 0.8102
No log 3.0 120 0.6360 0.6758 0.6360 0.7975
No log 3.05 122 0.6303 0.6246 0.6303 0.7939
No log 3.1 124 0.8275 0.5911 0.8275 0.9097
No log 3.15 126 0.9359 0.5442 0.9359 0.9674
No log 3.2 128 0.7436 0.6411 0.7436 0.8623
No log 3.25 130 0.6209 0.7034 0.6209 0.7880
No log 3.3 132 0.6096 0.7005 0.6096 0.7808
No log 3.35 134 0.5884 0.6518 0.5884 0.7671
No log 3.4 136 0.5957 0.6049 0.5957 0.7718
No log 3.45 138 0.6012 0.6246 0.6012 0.7754
No log 3.5 140 0.5991 0.6830 0.5991 0.7740
No log 3.55 142 0.6050 0.6932 0.6050 0.7778
No log 3.6 144 0.6045 0.7003 0.6045 0.7775
No log 3.65 146 0.6208 0.6733 0.6208 0.7879
No log 3.7 148 0.6234 0.6733 0.6234 0.7896
No log 3.75 150 0.6609 0.6386 0.6609 0.8130
No log 3.8 152 0.6559 0.6025 0.6559 0.8099
No log 3.85 154 0.6405 0.6830 0.6405 0.8003
No log 3.9 156 0.6777 0.6064 0.6777 0.8232
No log 3.95 158 0.7112 0.6236 0.7112 0.8433
No log 4.0 160 0.6488 0.6354 0.6488 0.8055
No log 4.05 162 0.6540 0.6164 0.6540 0.8087
No log 4.1 164 0.6743 0.5959 0.6743 0.8212
No log 4.15 166 0.7030 0.6189 0.7030 0.8384
No log 4.2 168 0.6313 0.6320 0.6313 0.7945
No log 4.25 170 0.6925 0.6116 0.6925 0.8322
No log 4.3 172 0.6887 0.6479 0.6887 0.8299
No log 4.35 174 0.6275 0.6460 0.6275 0.7922
No log 4.4 176 0.6863 0.5708 0.6863 0.8284
No log 4.45 178 0.6881 0.5410 0.6881 0.8295
No log 4.5 180 0.6786 0.5928 0.6786 0.8238
No log 4.55 182 0.6803 0.6195 0.6803 0.8248
No log 4.6 184 0.7209 0.5541 0.7209 0.8490
No log 4.65 186 0.7278 0.6524 0.7278 0.8531
No log 4.7 188 0.6632 0.6999 0.6632 0.8144
No log 4.75 190 0.7424 0.6141 0.7424 0.8617
No log 4.8 192 0.8203 0.5653 0.8203 0.9057
No log 4.85 194 0.6814 0.6217 0.6814 0.8254
No log 4.9 196 0.6002 0.7231 0.6002 0.7747
No log 4.95 198 0.9221 0.5273 0.9221 0.9603
No log 5.0 200 1.0504 0.5205 1.0504 1.0249
No log 5.05 202 0.8514 0.5273 0.8514 0.9227
No log 5.1 204 0.6165 0.6362 0.6165 0.7852
No log 5.15 206 0.5646 0.7095 0.5646 0.7514
No log 5.2 208 0.5795 0.6732 0.5795 0.7613
No log 5.25 210 0.5389 0.6830 0.5389 0.7341
No log 5.3 212 0.5879 0.5849 0.5879 0.7668
No log 5.35 214 0.6372 0.6445 0.6372 0.7982
No log 5.4 216 0.5637 0.6721 0.5637 0.7508
No log 5.45 218 0.5280 0.6932 0.5280 0.7267
No log 5.5 220 0.5398 0.7301 0.5398 0.7347
No log 5.55 222 0.5301 0.7064 0.5301 0.7281
No log 5.6 224 0.5402 0.7056 0.5402 0.7350
No log 5.65 226 0.5622 0.6795 0.5622 0.7498
No log 5.7 228 0.5894 0.6488 0.5894 0.7677
No log 5.75 230 0.5976 0.7099 0.5976 0.7730
No log 5.8 232 0.5780 0.6988 0.5780 0.7603
No log 5.85 234 0.5456 0.7423 0.5456 0.7386
No log 5.9 236 0.5518 0.7056 0.5518 0.7428
No log 5.95 238 0.5530 0.6876 0.5530 0.7437
No log 6.0 240 0.5618 0.6876 0.5618 0.7496
No log 6.05 242 0.5620 0.6705 0.5620 0.7497
No log 6.1 244 0.5613 0.6297 0.5613 0.7492
No log 6.15 246 0.5767 0.6102 0.5767 0.7594
No log 6.2 248 0.5547 0.6623 0.5547 0.7448
No log 6.25 250 0.5472 0.6392 0.5472 0.7398
No log 6.3 252 0.6453 0.6515 0.6453 0.8033
No log 6.35 254 0.6601 0.6515 0.6601 0.8125
No log 6.4 256 0.5685 0.6570 0.5685 0.7540
No log 6.45 258 0.5843 0.6900 0.5843 0.7644
No log 6.5 260 0.6538 0.5860 0.6538 0.8086
No log 6.55 262 0.6447 0.6071 0.6447 0.8029
No log 6.6 264 0.6097 0.6209 0.6097 0.7808
No log 6.65 266 0.6162 0.6070 0.6162 0.7850
No log 6.7 268 0.6285 0.6009 0.6285 0.7928
No log 6.75 270 0.5791 0.6869 0.5791 0.7610
No log 6.8 272 0.5953 0.6601 0.5953 0.7716
No log 6.85 274 0.6421 0.6746 0.6421 0.8013
No log 6.9 276 0.6641 0.6860 0.6641 0.8149
No log 6.95 278 0.6344 0.6705 0.6344 0.7965
No log 7.0 280 0.5755 0.6365 0.5755 0.7586
No log 7.05 282 0.5974 0.5876 0.5974 0.7729
No log 7.1 284 0.6826 0.6150 0.6826 0.8262
No log 7.15 286 0.6761 0.5951 0.6761 0.8222
No log 7.2 288 0.6139 0.6215 0.6139 0.7835
No log 7.25 290 0.5691 0.6644 0.5691 0.7544
No log 7.3 292 0.5494 0.6697 0.5494 0.7412
No log 7.35 294 0.5408 0.6861 0.5408 0.7354
No log 7.4 296 0.5506 0.6499 0.5506 0.7420
No log 7.45 298 0.6111 0.5770 0.6111 0.7817
No log 7.5 300 0.5848 0.5536 0.5848 0.7647
No log 7.55 302 0.5506 0.6606 0.5506 0.7420
No log 7.6 304 0.5300 0.7308 0.5300 0.7280
No log 7.65 306 0.5334 0.6756 0.5334 0.7304
No log 7.7 308 0.5416 0.6689 0.5416 0.7359
No log 7.75 310 0.5539 0.6796 0.5539 0.7442
No log 7.8 312 0.5717 0.6217 0.5717 0.7561
No log 7.85 314 0.6028 0.6424 0.6028 0.7764
No log 7.9 316 0.6168 0.6424 0.6168 0.7854
No log 7.95 318 0.6221 0.6424 0.6221 0.7887
No log 8.0 320 0.6121 0.6354 0.6121 0.7824
No log 8.05 322 0.6005 0.6584 0.6005 0.7749
No log 8.1 324 0.6163 0.6184 0.6163 0.7850
No log 8.15 326 0.6402 0.6461 0.6402 0.8001
No log 8.2 328 0.6205 0.6713 0.6205 0.7877
No log 8.25 330 0.6313 0.6788 0.6313 0.7945
No log 8.3 332 0.6517 0.6365 0.6517 0.8073
No log 8.35 334 0.6481 0.6398 0.6481 0.8051
No log 8.4 336 0.6528 0.6377 0.6528 0.8080
No log 8.45 338 0.7256 0.5625 0.7256 0.8518
No log 8.5 340 0.7220 0.5443 0.7220 0.8497
No log 8.55 342 0.6468 0.6099 0.6468 0.8042
No log 8.6 344 0.6238 0.6518 0.6238 0.7898
No log 8.65 346 0.6661 0.6249 0.6661 0.8162
No log 8.7 348 0.6745 0.5763 0.6745 0.8213
No log 8.75 350 0.6374 0.5787 0.6374 0.7984
No log 8.8 352 0.6116 0.5432 0.6116 0.7820
No log 8.85 354 0.5954 0.5656 0.5954 0.7716
No log 8.9 356 0.5792 0.6588 0.5792 0.7610
No log 8.95 358 0.5736 0.5959 0.5736 0.7574
No log 9.0 360 0.5558 0.6455 0.5558 0.7455
No log 9.05 362 0.5391 0.6729 0.5391 0.7342
No log 9.1 364 0.6127 0.5656 0.6127 0.7827
No log 9.15 366 0.7147 0.5387 0.7147 0.8454
No log 9.2 368 0.6736 0.5389 0.6736 0.8207
No log 9.25 370 0.5594 0.6383 0.5594 0.7479
No log 9.3 372 0.5797 0.6082 0.5797 0.7614
No log 9.35 374 0.6374 0.6596 0.6374 0.7984
No log 9.4 376 0.6121 0.6413 0.6121 0.7824
No log 9.45 378 0.5777 0.6704 0.5777 0.7600
No log 9.5 380 0.5811 0.5865 0.5811 0.7623
No log 9.55 382 0.5854 0.6578 0.5854 0.7651
No log 9.6 384 0.5718 0.6543 0.5718 0.7562
No log 9.65 386 0.5690 0.6966 0.5690 0.7543
No log 9.7 388 0.5714 0.6888 0.5714 0.7559
No log 9.75 390 0.5806 0.6772 0.5806 0.7620
No log 9.8 392 0.5810 0.6699 0.5810 0.7622
No log 9.85 394 0.5895 0.6078 0.5895 0.7678
No log 9.9 396 0.6099 0.5865 0.6099 0.7810
No log 9.95 398 0.6157 0.5865 0.6157 0.7847
No log 10.0 400 0.5913 0.6078 0.5913 0.7689
No log 10.05 402 0.6199 0.5876 0.6199 0.7874
No log 10.1 404 0.6382 0.5442 0.6382 0.7989
No log 10.15 406 0.6024 0.6278 0.6024 0.7761
No log 10.2 408 0.5950 0.6383 0.5950 0.7714
No log 10.25 410 0.5970 0.6383 0.5970 0.7727
No log 10.3 412 0.6029 0.6400 0.6029 0.7764
No log 10.35 414 0.6125 0.6207 0.6125 0.7827
No log 10.4 416 0.6233 0.6207 0.6233 0.7895
No log 10.45 418 0.6105 0.5650 0.6105 0.7813
No log 10.5 420 0.5944 0.6584 0.5944 0.7710
No log 10.55 422 0.5911 0.6830 0.5911 0.7688
No log 10.6 424 0.6004 0.6830 0.6004 0.7749
No log 10.65 426 0.6246 0.6128 0.6246 0.7903
No log 10.7 428 0.6285 0.6128 0.6285 0.7928
No log 10.75 430 0.6373 0.5820 0.6373 0.7983
No log 10.8 432 0.6749 0.4953 0.6749 0.8216
No log 10.85 434 0.6892 0.4953 0.6892 0.8302
No log 10.9 436 0.6671 0.4388 0.6671 0.8167
No log 10.95 438 0.6573 0.5959 0.6573 0.8108
No log 11.0 440 0.6441 0.6038 0.6441 0.8026
No log 11.05 442 0.6284 0.6447 0.6284 0.7927
No log 11.1 444 0.6258 0.6284 0.6258 0.7911
No log 11.15 446 0.6264 0.6124 0.6264 0.7915
No log 11.2 448 0.6501 0.5690 0.6501 0.8063
No log 11.25 450 0.7082 0.5698 0.7082 0.8415
No log 11.3 452 0.7334 0.5088 0.7334 0.8564
No log 11.35 454 0.7101 0.5113 0.7101 0.8427
No log 11.4 456 0.6706 0.4540 0.6706 0.8189
No log 11.45 458 0.6484 0.5202 0.6484 0.8053
No log 11.5 460 0.6319 0.5425 0.6319 0.7949
No log 11.55 462 0.6136 0.5950 0.6136 0.7833
No log 11.6 464 0.6047 0.6239 0.6047 0.7776
No log 11.65 466 0.6011 0.6167 0.6011 0.7753
No log 11.7 468 0.6011 0.6511 0.6011 0.7753
No log 11.75 470 0.5887 0.6511 0.5887 0.7672
No log 11.8 472 0.5910 0.6337 0.5910 0.7688
No log 11.85 474 0.5905 0.6371 0.5905 0.7684
No log 11.9 476 0.5822 0.6229 0.5822 0.7630
No log 11.95 478 0.5884 0.6262 0.5884 0.7671
No log 12.0 480 0.6048 0.6198 0.6048 0.7777
No log 12.05 482 0.6490 0.5220 0.6490 0.8056
No log 12.1 484 0.6895 0.5663 0.6895 0.8304
No log 12.15 486 0.6756 0.5787 0.6756 0.8220
No log 12.2 488 0.6314 0.6043 0.6314 0.7946
No log 12.25 490 0.6027 0.6296 0.6027 0.7763
No log 12.3 492 0.6155 0.5315 0.6155 0.7845
No log 12.35 494 0.6544 0.5005 0.6544 0.8089
No log 12.4 496 0.6559 0.5005 0.6559 0.8099
No log 12.45 498 0.6115 0.5759 0.6115 0.7820
0.2359 12.5 500 0.6074 0.5722 0.6074 0.7793
0.2359 12.55 502 0.6518 0.5858 0.6518 0.8074
0.2359 12.6 504 0.6762 0.5663 0.6762 0.8223
0.2359 12.65 506 0.6875 0.5663 0.6875 0.8291
0.2359 12.7 508 0.6651 0.5688 0.6651 0.8155
0.2359 12.75 510 0.6399 0.5917 0.6399 0.7999
0.2359 12.8 512 0.6269 0.6018 0.6269 0.7918
0.2359 12.85 514 0.6140 0.6227 0.6140 0.7836
0.2359 12.9 516 0.6096 0.6028 0.6096 0.7807
0.2359 12.95 518 0.6186 0.6046 0.6186 0.7865
0.2359 13.0 520 0.6286 0.6046 0.6286 0.7928
0.2359 13.05 522 0.6266 0.6249 0.6266 0.7916
0.2359 13.1 524 0.6140 0.6584 0.6140 0.7836
0.2359 13.15 526 0.6035 0.6259 0.6035 0.7768
0.2359 13.2 528 0.6084 0.5759 0.6084 0.7800
0.2359 13.25 530 0.6487 0.5667 0.6487 0.8054
0.2359 13.3 532 0.7047 0.6026 0.7047 0.8395
0.2359 13.35 534 0.6676 0.5843 0.6676 0.8171
0.2359 13.4 536 0.5946 0.5536 0.5946 0.7711
0.2359 13.45 538 0.5848 0.6249 0.5848 0.7647
0.2359 13.5 540 0.6121 0.6241 0.6121 0.7824
0.2359 13.55 542 0.6100 0.6122 0.6100 0.7810
0.2359 13.6 544 0.5695 0.6597 0.5695 0.7547
0.2359 13.65 546 0.5665 0.6108 0.5665 0.7527
0.2359 13.7 548 0.6272 0.5816 0.6272 0.7920
0.2359 13.75 550 0.6823 0.6102 0.6823 0.8260
0.2359 13.8 552 0.6203 0.6550 0.6203 0.7876
0.2359 13.85 554 0.5480 0.6659 0.5480 0.7403
0.2359 13.9 556 0.6187 0.6263 0.6187 0.7866
0.2359 13.95 558 0.7045 0.5686 0.7045 0.8394
0.2359 14.0 560 0.7303 0.5527 0.7303 0.8545
0.2359 14.05 562 0.7551 0.4421 0.7551 0.8690
0.2359 14.1 564 0.7637 0.4681 0.7637 0.8739
0.2359 14.15 566 0.7658 0.5045 0.7658 0.8751

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k8_task5_organization

Finetuned
(4019)
this model