ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k10_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5416
  • Qwk: 0.5644
  • Mse: 0.5416
  • Rmse: 0.7359

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.04 2 3.9310 -0.0387 3.9310 1.9827
No log 0.08 4 2.3112 0.0454 2.3112 1.5203
No log 0.12 6 1.4877 0.0362 1.4877 1.2197
No log 0.16 8 1.0467 0.2214 1.0467 1.0231
No log 0.2 10 1.0524 0.1962 1.0524 1.0259
No log 0.24 12 1.1484 0.1057 1.1484 1.0716
No log 0.28 14 1.1795 0.1460 1.1795 1.0860
No log 0.32 16 0.9398 0.3604 0.9398 0.9694
No log 0.36 18 0.9830 0.2850 0.9830 0.9915
No log 0.4 20 1.1579 0.1460 1.1579 1.0761
No log 0.44 22 1.0405 0.1835 1.0405 1.0201
No log 0.48 24 0.9368 0.4120 0.9368 0.9679
No log 0.52 26 0.7973 0.5004 0.7973 0.8929
No log 0.56 28 0.6969 0.5872 0.6969 0.8348
No log 0.6 30 1.0350 0.4422 1.0350 1.0173
No log 0.64 32 0.9401 0.4872 0.9401 0.9696
No log 0.68 34 0.7843 0.5922 0.7843 0.8856
No log 0.72 36 1.0590 0.4307 1.0590 1.0291
No log 0.76 38 0.7912 0.5547 0.7912 0.8895
No log 0.8 40 0.6922 0.6406 0.6922 0.8320
No log 0.84 42 0.6789 0.6439 0.6789 0.8239
No log 0.88 44 0.6657 0.6275 0.6657 0.8159
No log 0.92 46 0.6976 0.5409 0.6976 0.8352
No log 0.96 48 0.6742 0.6054 0.6742 0.8211
No log 1.0 50 0.6734 0.6315 0.6734 0.8206
No log 1.04 52 0.6684 0.6699 0.6684 0.8176
No log 1.08 54 0.6485 0.5964 0.6485 0.8053
No log 1.12 56 0.9853 0.5205 0.9853 0.9926
No log 1.16 58 1.1450 0.4493 1.1450 1.0701
No log 1.2 60 0.9591 0.5468 0.9591 0.9793
No log 1.24 62 0.8000 0.5730 0.8000 0.8944
No log 1.28 64 0.7442 0.6059 0.7442 0.8627
No log 1.32 66 0.7052 0.6238 0.7052 0.8398
No log 1.3600 68 0.6813 0.6438 0.6813 0.8254
No log 1.4 70 0.6807 0.6341 0.6807 0.8251
No log 1.44 72 0.7335 0.6109 0.7335 0.8564
No log 1.48 74 0.6393 0.6243 0.6393 0.7996
No log 1.52 76 0.7980 0.6057 0.7980 0.8933
No log 1.56 78 0.7395 0.6285 0.7395 0.8599
No log 1.6 80 0.7050 0.6371 0.7050 0.8396
No log 1.6400 82 0.6277 0.6186 0.6277 0.7923
No log 1.6800 84 0.6836 0.5645 0.6836 0.8268
No log 1.72 86 0.7973 0.6219 0.7973 0.8929
No log 1.76 88 1.0863 0.4855 1.0863 1.0423
No log 1.8 90 1.0548 0.4838 1.0548 1.0270
No log 1.8400 92 0.8415 0.5370 0.8415 0.9173
No log 1.88 94 0.6634 0.5847 0.6634 0.8145
No log 1.92 96 0.6156 0.6932 0.6156 0.7846
No log 1.96 98 0.6671 0.7042 0.6671 0.8168
No log 2.0 100 0.5660 0.6743 0.5660 0.7523
No log 2.04 102 0.9172 0.5571 0.9172 0.9577
No log 2.08 104 1.1138 0.4790 1.1138 1.0553
No log 2.12 106 0.7268 0.6065 0.7268 0.8525
No log 2.16 108 0.6032 0.6488 0.6032 0.7766
No log 2.2 110 0.5447 0.6901 0.5447 0.7381
No log 2.24 112 0.5447 0.6780 0.5447 0.7381
No log 2.2800 114 0.6116 0.6275 0.6116 0.7821
No log 2.32 116 0.8496 0.5885 0.8496 0.9218
No log 2.36 118 0.7266 0.5868 0.7266 0.8524
No log 2.4 120 0.5709 0.6184 0.5709 0.7556
No log 2.44 122 0.5311 0.6911 0.5311 0.7288
No log 2.48 124 0.5432 0.6981 0.5432 0.7370
No log 2.52 126 0.5282 0.7171 0.5282 0.7268
No log 2.56 128 0.5284 0.6896 0.5284 0.7269
No log 2.6 130 0.5733 0.6234 0.5733 0.7572
No log 2.64 132 0.5809 0.6089 0.5809 0.7622
No log 2.68 134 0.5928 0.5672 0.5928 0.7700
No log 2.7200 136 0.5492 0.6690 0.5492 0.7411
No log 2.76 138 0.5555 0.6482 0.5555 0.7453
No log 2.8 140 0.5916 0.6341 0.5916 0.7692
No log 2.84 142 0.7310 0.6340 0.7310 0.8550
No log 2.88 144 0.8603 0.5627 0.8603 0.9275
No log 2.92 146 0.6803 0.6022 0.6803 0.8248
No log 2.96 148 0.5680 0.6940 0.5680 0.7537
No log 3.0 150 0.6043 0.7066 0.6043 0.7774
No log 3.04 152 0.5946 0.6217 0.5946 0.7711
No log 3.08 154 0.6588 0.5898 0.6588 0.8117
No log 3.12 156 0.7111 0.5759 0.7111 0.8433
No log 3.16 158 0.8647 0.5436 0.8647 0.9299
No log 3.2 160 0.8021 0.6005 0.8021 0.8956
No log 3.24 162 0.6006 0.7440 0.6006 0.7750
No log 3.2800 164 0.5707 0.7143 0.5707 0.7554
No log 3.32 166 0.5723 0.7171 0.5723 0.7565
No log 3.36 168 0.5697 0.6713 0.5697 0.7548
No log 3.4 170 0.7734 0.5888 0.7735 0.8795
No log 3.44 172 0.7455 0.5986 0.7455 0.8635
No log 3.48 174 0.6104 0.6344 0.6104 0.7813
No log 3.52 176 0.5580 0.7033 0.5580 0.7470
No log 3.56 178 0.5947 0.6499 0.5947 0.7712
No log 3.6 180 0.7762 0.5516 0.7762 0.8810
No log 3.64 182 0.8487 0.5602 0.8487 0.9212
No log 3.68 184 0.6333 0.6592 0.6333 0.7958
No log 3.7200 186 0.5510 0.7562 0.5510 0.7423
No log 3.76 188 0.5276 0.7428 0.5276 0.7264
No log 3.8 190 0.5727 0.6603 0.5727 0.7568
No log 3.84 192 0.7488 0.5878 0.7488 0.8653
No log 3.88 194 0.7420 0.5681 0.7420 0.8614
No log 3.92 196 0.6176 0.6129 0.6176 0.7859
No log 3.96 198 0.5369 0.6690 0.5369 0.7328
No log 4.0 200 0.6228 0.6727 0.6228 0.7892
No log 4.04 202 0.6296 0.6289 0.6296 0.7934
No log 4.08 204 0.5329 0.6986 0.5329 0.7300
No log 4.12 206 0.6462 0.6861 0.6462 0.8039
No log 4.16 208 1.0600 0.4962 1.0600 1.0296
No log 4.2 210 1.1545 0.4307 1.1545 1.0745
No log 4.24 212 0.8592 0.5294 0.8592 0.9269
No log 4.28 214 0.5824 0.6228 0.5824 0.7632
No log 4.32 216 0.6299 0.6450 0.6299 0.7937
No log 4.36 218 0.7013 0.6107 0.7013 0.8374
No log 4.4 220 0.6326 0.6558 0.6326 0.7954
No log 4.44 222 0.5835 0.6678 0.5835 0.7639
No log 4.48 224 0.5276 0.6874 0.5276 0.7264
No log 4.52 226 0.6109 0.6507 0.6109 0.7816
No log 4.5600 228 0.7255 0.6271 0.7255 0.8517
No log 4.6 230 0.6527 0.6596 0.6527 0.8079
No log 4.64 232 0.5591 0.6780 0.5591 0.7477
No log 4.68 234 0.5306 0.6846 0.5306 0.7284
No log 4.72 236 0.5671 0.7175 0.5671 0.7531
No log 4.76 238 0.6025 0.6539 0.6025 0.7762
No log 4.8 240 0.5976 0.6974 0.5976 0.7730
No log 4.84 242 0.5769 0.6139 0.5769 0.7596
No log 4.88 244 0.5957 0.5328 0.5957 0.7718
No log 4.92 246 0.6266 0.5676 0.6266 0.7916
No log 4.96 248 0.6342 0.5951 0.6342 0.7964
No log 5.0 250 0.5909 0.6733 0.5909 0.7687
No log 5.04 252 0.5329 0.7040 0.5329 0.7300
No log 5.08 254 0.5130 0.6959 0.5130 0.7162
No log 5.12 256 0.5206 0.6988 0.5206 0.7215
No log 5.16 258 0.5358 0.6903 0.5358 0.7320
No log 5.2 260 0.5515 0.6903 0.5515 0.7426
No log 5.24 262 0.5614 0.6196 0.5614 0.7493
No log 5.28 264 0.5972 0.5342 0.5972 0.7728
No log 5.32 266 0.6030 0.5696 0.6030 0.7765
No log 5.36 268 0.5688 0.6644 0.5688 0.7542
No log 5.4 270 0.5386 0.6983 0.5386 0.7339
No log 5.44 272 0.5360 0.6969 0.5360 0.7321
No log 5.48 274 0.5314 0.7198 0.5314 0.7290
No log 5.52 276 0.5153 0.6756 0.5153 0.7179
No log 5.5600 278 0.5220 0.6535 0.5220 0.7225
No log 5.6 280 0.5313 0.6292 0.5313 0.7289
No log 5.64 282 0.5370 0.6841 0.5370 0.7328
No log 5.68 284 0.5446 0.6648 0.5446 0.7379
No log 5.72 286 0.5558 0.6097 0.5558 0.7455
No log 5.76 288 0.6625 0.5270 0.6625 0.8139
No log 5.8 290 0.7315 0.5491 0.7315 0.8553
No log 5.84 292 0.6294 0.5793 0.6294 0.7934
No log 5.88 294 0.5265 0.6699 0.5265 0.7256
No log 5.92 296 0.5384 0.6704 0.5384 0.7338
No log 5.96 298 0.5583 0.7011 0.5583 0.7472
No log 6.0 300 0.5323 0.6833 0.5323 0.7296
No log 6.04 302 0.5501 0.6017 0.5501 0.7417
No log 6.08 304 0.6376 0.6035 0.6376 0.7985
No log 6.12 306 0.6537 0.6018 0.6537 0.8085
No log 6.16 308 0.5524 0.6701 0.5524 0.7433
No log 6.2 310 0.5010 0.7005 0.5010 0.7078
No log 6.24 312 0.6185 0.6661 0.6185 0.7864
No log 6.28 314 0.7123 0.6653 0.7123 0.8440
No log 6.32 316 0.6645 0.6343 0.6645 0.8152
No log 6.36 318 0.5990 0.6900 0.5990 0.7740
No log 6.4 320 0.5407 0.7196 0.5407 0.7353
No log 6.44 322 0.5296 0.7133 0.5296 0.7278
No log 6.48 324 0.5272 0.7278 0.5272 0.7261
No log 6.52 326 0.5278 0.7213 0.5278 0.7265
No log 6.5600 328 0.5193 0.7213 0.5193 0.7206
No log 6.6 330 0.5255 0.7049 0.5255 0.7249
No log 6.64 332 0.5291 0.6259 0.5291 0.7274
No log 6.68 334 0.5304 0.6292 0.5304 0.7283
No log 6.72 336 0.5233 0.6327 0.5233 0.7234
No log 6.76 338 0.5180 0.6195 0.5180 0.7197
No log 6.8 340 0.5514 0.6545 0.5514 0.7426
No log 6.84 342 0.5102 0.6174 0.5102 0.7143
No log 6.88 344 0.4757 0.7154 0.4757 0.6897
No log 6.92 346 0.5036 0.7042 0.5036 0.7096
No log 6.96 348 0.5246 0.6872 0.5246 0.7243
No log 7.0 350 0.5282 0.6872 0.5282 0.7267
No log 7.04 352 0.5002 0.6822 0.5002 0.7072
No log 7.08 354 0.4816 0.6896 0.4816 0.6940
No log 7.12 356 0.4928 0.7078 0.4928 0.7020
No log 7.16 358 0.5558 0.7304 0.5558 0.7455
No log 7.2 360 0.6065 0.7039 0.6065 0.7788
No log 7.24 362 0.5423 0.7304 0.5423 0.7364
No log 7.28 364 0.4761 0.6931 0.4761 0.6900
No log 7.32 366 0.5056 0.7158 0.5056 0.7110
No log 7.36 368 0.5284 0.7444 0.5284 0.7269
No log 7.4 370 0.5048 0.7148 0.5048 0.7105
No log 7.44 372 0.4850 0.7056 0.4850 0.6964
No log 7.48 374 0.4923 0.6940 0.4923 0.7016
No log 7.52 376 0.5044 0.6400 0.5044 0.7102
No log 7.5600 378 0.5276 0.6230 0.5276 0.7263
No log 7.6 380 0.5569 0.6322 0.5569 0.7463
No log 7.64 382 0.5725 0.6673 0.5725 0.7566
No log 7.68 384 0.5480 0.6322 0.5480 0.7403
No log 7.72 386 0.5059 0.6282 0.5059 0.7112
No log 7.76 388 0.4826 0.7122 0.4826 0.6947
No log 7.8 390 0.4879 0.7193 0.4879 0.6985
No log 7.84 392 0.4877 0.7048 0.4877 0.6984
No log 7.88 394 0.4828 0.7115 0.4828 0.6948
No log 7.92 396 0.4952 0.6880 0.4952 0.7037
No log 7.96 398 0.5147 0.6219 0.5147 0.7174
No log 8.0 400 0.5447 0.6217 0.5447 0.7380
No log 8.04 402 0.5675 0.6028 0.5675 0.7533
No log 8.08 404 0.5637 0.6028 0.5637 0.7508
No log 8.12 406 0.5164 0.6813 0.5164 0.7186
No log 8.16 408 0.5062 0.6770 0.5062 0.7115
No log 8.2 410 0.5099 0.6658 0.5099 0.7141
No log 8.24 412 0.5265 0.6517 0.5265 0.7256
No log 8.28 414 0.5260 0.6894 0.5260 0.7252
No log 8.32 416 0.5464 0.6528 0.5464 0.7392
No log 8.36 418 0.5498 0.6013 0.5498 0.7415
No log 8.4 420 0.5497 0.6445 0.5497 0.7414
No log 8.44 422 0.5562 0.5874 0.5562 0.7458
No log 8.48 424 0.5637 0.6118 0.5637 0.7508
No log 8.52 426 0.5706 0.5988 0.5706 0.7554
No log 8.56 428 0.5689 0.5966 0.5689 0.7543
No log 8.6 430 0.5698 0.5859 0.5698 0.7548
No log 8.64 432 0.5650 0.6916 0.5650 0.7516
No log 8.68 434 0.5752 0.6830 0.5752 0.7584
No log 8.72 436 0.6013 0.6830 0.6013 0.7754
No log 8.76 438 0.6336 0.6547 0.6336 0.7960
No log 8.8 440 0.6538 0.6605 0.6538 0.8086
No log 8.84 442 0.6594 0.6396 0.6594 0.8120
No log 8.88 444 0.6666 0.5787 0.6666 0.8164
No log 8.92 446 0.6069 0.6198 0.6069 0.7790
No log 8.96 448 0.5633 0.6097 0.5633 0.7505
No log 9.0 450 0.5579 0.6426 0.5579 0.7469
No log 9.04 452 0.5587 0.6703 0.5587 0.7474
No log 9.08 454 0.5700 0.6583 0.5700 0.7550
No log 9.12 456 0.5738 0.6290 0.5738 0.7575
No log 9.16 458 0.5536 0.6627 0.5536 0.7441
No log 9.2 460 0.5458 0.6610 0.5458 0.7387
No log 9.24 462 0.5525 0.6507 0.5525 0.7433
No log 9.28 464 0.5860 0.5894 0.5860 0.7655
No log 9.32 466 0.6227 0.5566 0.6227 0.7891
No log 9.36 468 0.6771 0.5699 0.6771 0.8229
No log 9.4 470 0.6508 0.5468 0.6508 0.8067
No log 9.44 472 0.5868 0.5672 0.5868 0.7660
No log 9.48 474 0.5632 0.6272 0.5632 0.7505
No log 9.52 476 0.5765 0.6584 0.5765 0.7593
No log 9.56 478 0.5571 0.6500 0.5571 0.7464
No log 9.6 480 0.5411 0.6770 0.5411 0.7356
No log 9.64 482 0.5438 0.6575 0.5438 0.7374
No log 9.68 484 0.5492 0.6390 0.5492 0.7411
No log 9.72 486 0.5620 0.6498 0.5620 0.7497
No log 9.76 488 0.5709 0.6397 0.5709 0.7556
No log 9.8 490 0.5708 0.5874 0.5708 0.7555
No log 9.84 492 0.5619 0.6096 0.5619 0.7496
No log 9.88 494 0.5548 0.6128 0.5548 0.7448
No log 9.92 496 0.5545 0.5988 0.5545 0.7447
No log 9.96 498 0.5572 0.6207 0.5572 0.7464
0.2617 10.0 500 0.5632 0.5879 0.5632 0.7504
0.2617 10.04 502 0.5634 0.6402 0.5634 0.7506
0.2617 10.08 504 0.5479 0.6402 0.5479 0.7402
0.2617 10.12 506 0.5378 0.6869 0.5378 0.7333
0.2617 10.16 508 0.5453 0.6869 0.5453 0.7385
0.2617 10.2 510 0.5670 0.6452 0.5670 0.7530
0.2617 10.24 512 0.5832 0.6275 0.5832 0.7637
0.2617 10.28 514 0.5638 0.6377 0.5638 0.7508
0.2617 10.32 516 0.5498 0.6006 0.5498 0.7415
0.2617 10.36 518 0.5591 0.6017 0.5591 0.7477
0.2617 10.4 520 0.5878 0.5342 0.5878 0.7667
0.2617 10.44 522 0.5893 0.5902 0.5893 0.7677
0.2617 10.48 524 0.5373 0.6174 0.5373 0.7330
0.2617 10.52 526 0.5274 0.6107 0.5274 0.7262
0.2617 10.56 528 0.5245 0.6508 0.5245 0.7242
0.2617 10.6 530 0.5224 0.6675 0.5224 0.7228
0.2617 10.64 532 0.5341 0.6317 0.5341 0.7308
0.2617 10.68 534 0.5495 0.6317 0.5495 0.7413
0.2617 10.72 536 0.5775 0.6539 0.5775 0.7599
0.2617 10.76 538 0.5936 0.6719 0.5936 0.7704
0.2617 10.8 540 0.5724 0.6539 0.5724 0.7566
0.2617 10.84 542 0.5471 0.5894 0.5471 0.7397
0.2617 10.88 544 0.5587 0.6322 0.5587 0.7474
0.2617 10.92 546 0.5866 0.6090 0.5866 0.7659
0.2617 10.96 548 0.5746 0.6090 0.5746 0.7580
0.2617 11.0 550 0.5413 0.6139 0.5413 0.7357
0.2617 11.04 552 0.5388 0.6433 0.5388 0.7340
0.2617 11.08 554 0.6152 0.6807 0.6152 0.7844
0.2617 11.12 556 0.6755 0.6661 0.6755 0.8219
0.2617 11.16 558 0.6716 0.6624 0.6716 0.8195
0.2617 11.2 560 0.6576 0.6624 0.6576 0.8110
0.2617 11.24 562 0.5878 0.6748 0.5878 0.7667
0.2617 11.28 564 0.5466 0.6602 0.5466 0.7393
0.2617 11.32 566 0.5299 0.6830 0.5299 0.7279
0.2617 11.36 568 0.5321 0.6720 0.5321 0.7295
0.2617 11.4 570 0.5459 0.6916 0.5459 0.7389
0.2617 11.44 572 0.6004 0.6687 0.6004 0.7748
0.2617 11.48 574 0.6288 0.6464 0.6288 0.7930
0.2617 11.52 576 0.6516 0.6539 0.6516 0.8072
0.2617 11.56 578 0.6330 0.6187 0.6330 0.7956
0.2617 11.6 580 0.5975 0.5640 0.5975 0.7730
0.2617 11.64 582 0.5799 0.5759 0.5799 0.7615
0.2617 11.68 584 0.5613 0.5644 0.5613 0.7492
0.2617 11.72 586 0.5532 0.5644 0.5532 0.7438
0.2617 11.76 588 0.5416 0.5644 0.5416 0.7359

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k10_task5_organization

Finetuned
(4019)
this model