ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k15_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8403
  • Qwk: 0.3879
  • Mse: 0.8403
  • Rmse: 0.9167

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0417 2 4.1548 -0.0255 4.1548 2.0383
No log 0.0833 4 2.3958 0.0782 2.3958 1.5478
No log 0.125 6 1.7485 -0.0277 1.7485 1.3223
No log 0.1667 8 1.4278 0.0143 1.4278 1.1949
No log 0.2083 10 1.6207 -0.0078 1.6207 1.2731
No log 0.25 12 1.4777 0.0143 1.4777 1.2156
No log 0.2917 14 1.1192 0.0849 1.1192 1.0579
No log 0.3333 16 1.0639 0.1805 1.0639 1.0315
No log 0.375 18 1.2032 0.1471 1.2032 1.0969
No log 0.4167 20 1.3116 0.0 1.3116 1.1452
No log 0.4583 22 1.2854 0.0380 1.2854 1.1338
No log 0.5 24 1.3565 0.0 1.3565 1.1647
No log 0.5417 26 1.2364 0.1525 1.2364 1.1119
No log 0.5833 28 1.0653 0.1832 1.0653 1.0321
No log 0.625 30 1.0585 0.2221 1.0585 1.0288
No log 0.6667 32 1.0449 0.1805 1.0449 1.0222
No log 0.7083 34 1.0864 0.2171 1.0864 1.0423
No log 0.75 36 1.1779 0.0909 1.1779 1.0853
No log 0.7917 38 1.1656 0.1057 1.1656 1.0796
No log 0.8333 40 1.1232 0.1057 1.1232 1.0598
No log 0.875 42 1.0971 0.1444 1.0971 1.0474
No log 0.9167 44 1.1081 0.1564 1.1081 1.0527
No log 0.9583 46 1.1087 0.1268 1.1087 1.0529
No log 1.0 48 1.0430 0.2539 1.0430 1.0213
No log 1.0417 50 1.0234 0.2391 1.0234 1.0116
No log 1.0833 52 0.9821 0.2391 0.9821 0.9910
No log 1.125 54 0.9673 0.2416 0.9673 0.9835
No log 1.1667 56 0.9211 0.2515 0.9211 0.9597
No log 1.2083 58 0.8952 0.3562 0.8952 0.9462
No log 1.25 60 0.9190 0.2441 0.9190 0.9587
No log 1.2917 62 0.8756 0.2566 0.8756 0.9357
No log 1.3333 64 0.8658 0.3446 0.8658 0.9305
No log 1.375 66 0.8698 0.2865 0.8698 0.9326
No log 1.4167 68 0.9062 0.3563 0.9062 0.9519
No log 1.4583 70 0.9351 0.3194 0.9351 0.9670
No log 1.5 72 0.9294 0.4156 0.9294 0.9641
No log 1.5417 74 0.9215 0.3514 0.9215 0.9599
No log 1.5833 76 0.9396 0.2599 0.9396 0.9693
No log 1.625 78 0.9561 0.3045 0.9561 0.9778
No log 1.6667 80 0.9262 0.2365 0.9262 0.9624
No log 1.7083 82 0.9044 0.3237 0.9044 0.9510
No log 1.75 84 0.8804 0.3354 0.8804 0.9383
No log 1.7917 86 0.8768 0.4027 0.8768 0.9364
No log 1.8333 88 0.8972 0.3572 0.8972 0.9472
No log 1.875 90 0.9541 0.4474 0.9541 0.9768
No log 1.9167 92 0.9286 0.3786 0.9286 0.9636
No log 1.9583 94 0.9339 0.3276 0.9339 0.9664
No log 2.0 96 1.0007 0.3409 1.0007 1.0004
No log 2.0417 98 1.1665 0.4379 1.1665 1.0801
No log 2.0833 100 1.0247 0.4085 1.0247 1.0123
No log 2.125 102 0.9765 0.4722 0.9765 0.9882
No log 2.1667 104 1.1636 0.4009 1.1636 1.0787
No log 2.2083 106 1.1668 0.4281 1.1668 1.0802
No log 2.25 108 0.9637 0.4474 0.9637 0.9817
No log 2.2917 110 0.8928 0.4935 0.8928 0.9449
No log 2.3333 112 0.8580 0.4015 0.8580 0.9263
No log 2.375 114 0.8747 0.4419 0.8747 0.9352
No log 2.4167 116 1.1410 0.3569 1.1410 1.0682
No log 2.4583 118 1.2761 0.2939 1.2761 1.1297
No log 2.5 120 1.0897 0.3958 1.0897 1.0439
No log 2.5417 122 0.9449 0.3861 0.9449 0.9721
No log 2.5833 124 0.9232 0.4030 0.9232 0.9608
No log 2.625 126 0.9388 0.2933 0.9388 0.9689
No log 2.6667 128 0.9570 0.2865 0.9570 0.9782
No log 2.7083 130 1.0198 0.3418 1.0198 1.0099
No log 2.75 132 0.9547 0.4224 0.9547 0.9771
No log 2.7917 134 0.8685 0.3525 0.8685 0.9319
No log 2.8333 136 0.8741 0.4232 0.8741 0.9349
No log 2.875 138 0.8458 0.3647 0.8458 0.9197
No log 2.9167 140 1.0142 0.4667 1.0142 1.0071
No log 2.9583 142 1.1446 0.3863 1.1446 1.0699
No log 3.0 144 1.0006 0.4681 1.0006 1.0003
No log 3.0417 146 0.9234 0.2625 0.9234 0.9610
No log 3.0833 148 0.9595 0.3340 0.9595 0.9796
No log 3.125 150 0.8430 0.3952 0.8430 0.9182
No log 3.1667 152 0.8278 0.4440 0.8278 0.9098
No log 3.2083 154 0.8536 0.4606 0.8536 0.9239
No log 3.25 156 1.0347 0.4287 1.0347 1.0172
No log 3.2917 158 1.2677 0.3806 1.2677 1.1259
No log 3.3333 160 1.2201 0.3969 1.2201 1.1046
No log 3.375 162 0.9297 0.4341 0.9297 0.9642
No log 3.4167 164 0.7180 0.4869 0.7180 0.8474
No log 3.4583 166 0.7563 0.5600 0.7563 0.8697
No log 3.5 168 0.7580 0.5600 0.7580 0.8706
No log 3.5417 170 0.7306 0.5354 0.7306 0.8547
No log 3.5833 172 0.8614 0.5252 0.8614 0.9281
No log 3.625 174 0.9489 0.4893 0.9489 0.9741
No log 3.6667 176 0.8565 0.5252 0.8565 0.9255
No log 3.7083 178 0.7778 0.4676 0.7778 0.8820
No log 3.75 180 0.7795 0.4407 0.7795 0.8829
No log 3.7917 182 0.8096 0.3631 0.8096 0.8998
No log 3.8333 184 0.8346 0.3891 0.8346 0.9136
No log 3.875 186 0.8647 0.3689 0.8647 0.9299
No log 3.9167 188 0.9317 0.3172 0.9317 0.9653
No log 3.9583 190 0.9565 0.2643 0.9565 0.9780
No log 4.0 192 0.9462 0.2643 0.9462 0.9727
No log 4.0417 194 0.9576 0.4041 0.9576 0.9785
No log 4.0833 196 0.9202 0.3306 0.9202 0.9593
No log 4.125 198 0.9561 0.3782 0.9561 0.9778
No log 4.1667 200 0.9587 0.2922 0.9587 0.9792
No log 4.2083 202 0.9556 0.3474 0.9556 0.9776
No log 4.25 204 1.1300 0.2870 1.1300 1.0630
No log 4.2917 206 1.1312 0.3226 1.1312 1.0636
No log 4.3333 208 0.9576 0.4102 0.9576 0.9786
No log 4.375 210 0.9272 0.3952 0.9272 0.9629
No log 4.4167 212 0.9486 0.3891 0.9486 0.9740
No log 4.4583 214 0.8700 0.3418 0.8700 0.9327
No log 4.5 216 0.9413 0.3424 0.9413 0.9702
No log 4.5417 218 1.1120 0.4540 1.1120 1.0545
No log 4.5833 220 1.1204 0.4654 1.1204 1.0585
No log 4.625 222 0.9963 0.4333 0.9963 0.9982
No log 4.6667 224 0.9049 0.3714 0.9049 0.9513
No log 4.7083 226 0.8515 0.3025 0.8515 0.9228
No log 4.75 228 0.8499 0.2897 0.8499 0.9219
No log 4.7917 230 0.8689 0.3192 0.8689 0.9322
No log 4.8333 232 0.8914 0.3760 0.8914 0.9442
No log 4.875 234 0.9697 0.3322 0.9697 0.9847
No log 4.9167 236 0.9465 0.3322 0.9465 0.9729
No log 4.9583 238 0.8557 0.3883 0.8557 0.9251
No log 5.0 240 0.8114 0.3858 0.8114 0.9008
No log 5.0417 242 0.8086 0.3960 0.8086 0.8992
No log 5.0833 244 0.8653 0.3763 0.8653 0.9302
No log 5.125 246 0.9370 0.3597 0.9370 0.9680
No log 5.1667 248 0.8798 0.3740 0.8798 0.9380
No log 5.2083 250 0.8603 0.3740 0.8603 0.9275
No log 5.25 252 0.8188 0.4168 0.8188 0.9049
No log 5.2917 254 0.8139 0.4456 0.8139 0.9021
No log 5.3333 256 0.8805 0.4560 0.8805 0.9383
No log 5.375 258 0.9266 0.4560 0.9266 0.9626
No log 5.4167 260 1.1049 0.4463 1.1049 1.0511
No log 5.4583 262 1.1034 0.3893 1.1034 1.0504
No log 5.5 264 1.0363 0.3943 1.0363 1.0180
No log 5.5417 266 0.9620 0.4709 0.9620 0.9808
No log 5.5833 268 0.9249 0.4774 0.9249 0.9617
No log 5.625 270 0.9346 0.4245 0.9346 0.9667
No log 5.6667 272 0.9486 0.4066 0.9486 0.9740
No log 5.7083 274 1.0117 0.4107 1.0117 1.0058
No log 5.75 276 1.0098 0.4192 1.0098 1.0049
No log 5.7917 278 0.9511 0.2818 0.9511 0.9753
No log 5.8333 280 0.9663 0.3363 0.9663 0.9830
No log 5.875 282 0.9960 0.3942 0.9960 0.9980
No log 5.9167 284 0.9172 0.3571 0.9172 0.9577
No log 5.9583 286 1.0091 0.4202 1.0091 1.0046
No log 6.0 288 1.1785 0.3243 1.1785 1.0856
No log 6.0417 290 1.0882 0.4216 1.0882 1.0432
No log 6.0833 292 0.8822 0.4428 0.8822 0.9393
No log 6.125 294 0.8384 0.4168 0.8384 0.9156
No log 6.1667 296 0.8490 0.4370 0.8490 0.9214
No log 6.2083 298 0.8154 0.4527 0.8154 0.9030
No log 6.25 300 0.8294 0.3494 0.8294 0.9107
No log 6.2917 302 0.8173 0.4310 0.8173 0.9041
No log 6.3333 304 0.7874 0.4368 0.7874 0.8873
No log 6.375 306 0.8128 0.4450 0.8128 0.9016
No log 6.4167 308 0.8299 0.4433 0.8299 0.9110
No log 6.4583 310 0.7925 0.4192 0.7925 0.8902
No log 6.5 312 0.8308 0.4433 0.8308 0.9115
No log 6.5417 314 0.9068 0.4666 0.9068 0.9522
No log 6.5833 316 0.9417 0.4542 0.9417 0.9704
No log 6.625 318 0.8779 0.4144 0.8779 0.9369
No log 6.6667 320 0.7791 0.3025 0.7791 0.8827
No log 6.7083 322 0.7611 0.3326 0.7611 0.8724
No log 6.75 324 0.7638 0.3339 0.7638 0.8740
No log 6.7917 326 0.8430 0.4174 0.8430 0.9182
No log 6.8333 328 0.9390 0.4773 0.9390 0.9690
No log 6.875 330 0.8920 0.4799 0.8920 0.9445
No log 6.9167 332 0.8459 0.4916 0.8459 0.9197
No log 6.9583 334 0.8064 0.5041 0.8064 0.8980
No log 7.0 336 0.7656 0.5056 0.7656 0.8750
No log 7.0417 338 0.7519 0.5056 0.7519 0.8671
No log 7.0833 340 0.7164 0.4712 0.7164 0.8464
No log 7.125 342 0.7254 0.4562 0.7254 0.8517
No log 7.1667 344 0.7600 0.4809 0.7600 0.8718
No log 7.2083 346 0.8554 0.5229 0.8554 0.9249
No log 7.25 348 0.8250 0.4681 0.8250 0.9083
No log 7.2917 350 0.7833 0.4824 0.7833 0.8850
No log 7.3333 352 0.7662 0.3840 0.7662 0.8753
No log 7.375 354 0.7937 0.3821 0.7937 0.8909
No log 7.4167 356 0.8736 0.4292 0.8736 0.9347
No log 7.4583 358 0.9297 0.4417 0.9297 0.9642
No log 7.5 360 0.9084 0.4549 0.9084 0.9531
No log 7.5417 362 0.8279 0.4839 0.8279 0.9099
No log 7.5833 364 0.8035 0.4428 0.8035 0.8964
No log 7.625 366 0.8155 0.4611 0.8155 0.9031
No log 7.6667 368 0.7815 0.4524 0.7815 0.8840
No log 7.7083 370 0.8037 0.4352 0.8037 0.8965
No log 7.75 372 0.8179 0.4450 0.8179 0.9044
No log 7.7917 374 0.8341 0.4565 0.8341 0.9133
No log 7.8333 376 0.7948 0.4075 0.7948 0.8915
No log 7.875 378 0.7599 0.4628 0.7599 0.8717
No log 7.9167 380 0.7433 0.4371 0.7433 0.8621
No log 7.9583 382 0.7448 0.4483 0.7448 0.8630
No log 8.0 384 0.7730 0.4839 0.7730 0.8792
No log 8.0417 386 0.7958 0.4696 0.7958 0.8921
No log 8.0833 388 0.8024 0.4696 0.8024 0.8958
No log 8.125 390 0.8151 0.4804 0.8151 0.9028
No log 8.1667 392 0.8188 0.4708 0.8188 0.9049
No log 8.2083 394 0.7838 0.4836 0.7838 0.8853
No log 8.25 396 0.7529 0.5208 0.7529 0.8677
No log 8.2917 398 0.7396 0.4759 0.7396 0.8600
No log 8.3333 400 0.7487 0.4110 0.7487 0.8653
No log 8.375 402 0.7470 0.4128 0.7470 0.8643
No log 8.4167 404 0.7336 0.4527 0.7336 0.8565
No log 8.4583 406 0.7299 0.5052 0.7299 0.8544
No log 8.5 408 0.7179 0.5160 0.7179 0.8473
No log 8.5417 410 0.7154 0.4645 0.7154 0.8458
No log 8.5833 412 0.7946 0.4696 0.7946 0.8914
No log 8.625 414 0.7924 0.4711 0.7924 0.8902
No log 8.6667 416 0.7317 0.4361 0.7317 0.8554
No log 8.7083 418 0.7236 0.5040 0.7236 0.8506
No log 8.75 420 0.7264 0.5040 0.7264 0.8523
No log 8.7917 422 0.7566 0.4499 0.7566 0.8698
No log 8.8333 424 0.8262 0.4832 0.8262 0.9089
No log 8.875 426 0.7983 0.4237 0.7983 0.8935
No log 8.9167 428 0.7494 0.4774 0.7494 0.8657
No log 8.9583 430 0.7527 0.5288 0.7527 0.8676
No log 9.0 432 0.7681 0.4515 0.7681 0.8764
No log 9.0417 434 0.8779 0.4832 0.8779 0.9370
No log 9.0833 436 0.9250 0.4470 0.9250 0.9618
No log 9.125 438 0.8619 0.4847 0.8619 0.9284
No log 9.1667 440 0.8151 0.4288 0.8151 0.9029
No log 9.2083 442 0.8085 0.4802 0.8085 0.8991
No log 9.25 444 0.8189 0.3994 0.8189 0.9049
No log 9.2917 446 0.8421 0.4344 0.8421 0.9177
No log 9.3333 448 0.8810 0.3623 0.8810 0.9386
No log 9.375 450 0.8583 0.3474 0.8583 0.9265
No log 9.4167 452 0.8178 0.3556 0.8178 0.9043
No log 9.4583 454 0.7984 0.4030 0.7984 0.8935
No log 9.5 456 0.8025 0.4988 0.8025 0.8958
No log 9.5417 458 0.8656 0.4098 0.8656 0.9304
No log 9.5833 460 0.9848 0.4439 0.9848 0.9924
No log 9.625 462 1.0176 0.4439 1.0176 1.0088
No log 9.6667 464 0.9329 0.4326 0.9329 0.9659
No log 9.7083 466 0.8348 0.4966 0.8348 0.9137
No log 9.75 468 0.7873 0.4531 0.7873 0.8873
No log 9.7917 470 0.7887 0.4772 0.7887 0.8881
No log 9.8333 472 0.8532 0.4590 0.8532 0.9237
No log 9.875 474 0.8682 0.4812 0.8682 0.9318
No log 9.9167 476 0.7911 0.4741 0.7911 0.8894
No log 9.9583 478 0.7564 0.4912 0.7564 0.8697
No log 10.0 480 0.7624 0.5038 0.7624 0.8732
No log 10.0417 482 0.7601 0.4912 0.7601 0.8718
No log 10.0833 484 0.7599 0.5139 0.7599 0.8717
No log 10.125 486 0.7662 0.5133 0.7662 0.8753
No log 10.1667 488 0.7765 0.5316 0.7765 0.8812
No log 10.2083 490 0.8319 0.5140 0.8319 0.9121
No log 10.25 492 0.8213 0.5147 0.8213 0.9062
No log 10.2917 494 0.7815 0.3760 0.7815 0.8840
No log 10.3333 496 0.8002 0.3854 0.8002 0.8946
No log 10.375 498 0.7978 0.3968 0.7978 0.8932
0.277 10.4167 500 0.7946 0.4048 0.7946 0.8914
0.277 10.4583 502 0.8133 0.4378 0.8133 0.9018
0.277 10.5 504 0.8605 0.4613 0.8605 0.9276
0.277 10.5417 506 0.9032 0.4231 0.9032 0.9504
0.277 10.5833 508 0.8842 0.4373 0.8842 0.9403
0.277 10.625 510 0.8728 0.4495 0.8728 0.9343
0.277 10.6667 512 0.8389 0.4968 0.8389 0.9159
0.277 10.7083 514 0.8076 0.5107 0.8076 0.8987
0.277 10.75 516 0.8013 0.5107 0.8013 0.8951
0.277 10.7917 518 0.8269 0.4853 0.8269 0.9093
0.277 10.8333 520 0.8270 0.4853 0.8270 0.9094
0.277 10.875 522 0.7992 0.5107 0.7993 0.8940
0.277 10.9167 524 0.8064 0.5002 0.8064 0.8980
0.277 10.9583 526 0.8172 0.5035 0.8172 0.9040
0.277 11.0 528 0.8188 0.4842 0.8188 0.9049
0.277 11.0417 530 0.8302 0.5065 0.8302 0.9112
0.277 11.0833 532 0.8032 0.5040 0.8032 0.8962
0.277 11.125 534 0.8217 0.5489 0.8217 0.9065
0.277 11.1667 536 0.8653 0.5353 0.8653 0.9302
0.277 11.2083 538 0.8131 0.5267 0.8131 0.9017
0.277 11.25 540 0.7666 0.5551 0.7666 0.8755
0.277 11.2917 542 0.7751 0.5406 0.7751 0.8804
0.277 11.3333 544 0.7695 0.5287 0.7695 0.8772
0.277 11.375 546 0.7788 0.4975 0.7788 0.8825
0.277 11.4167 548 0.8797 0.5046 0.8797 0.9379
0.277 11.4583 550 0.9438 0.4341 0.9438 0.9715
0.277 11.5 552 0.9051 0.3972 0.9051 0.9514
0.277 11.5417 554 0.8319 0.3992 0.8319 0.9121
0.277 11.5833 556 0.8202 0.3556 0.8202 0.9057
0.277 11.625 558 0.8178 0.3859 0.8178 0.9043
0.277 11.6667 560 0.8403 0.3879 0.8403 0.9167

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k15_task5_organization

Finetuned
(4019)
this model