ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k8_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8769
  • Qwk: 0.4489
  • Mse: 0.8769
  • Rmse: 0.9364

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0714 2 3.9664 -0.0076 3.9664 1.9916
No log 0.1429 4 2.1210 0.0539 2.1210 1.4564
No log 0.2143 6 1.5060 -0.0180 1.5060 1.2272
No log 0.2857 8 1.1326 0.1927 1.1326 1.0642
No log 0.3571 10 1.2156 0.0916 1.2156 1.1025
No log 0.4286 12 1.1578 0.1821 1.1578 1.0760
No log 0.5 14 1.3698 0.0714 1.3698 1.1704
No log 0.5714 16 1.2796 0.1224 1.2796 1.1312
No log 0.6429 18 1.0425 0.3059 1.0425 1.0210
No log 0.7143 20 1.0512 0.3104 1.0512 1.0253
No log 0.7857 22 1.1021 0.2903 1.1021 1.0498
No log 0.8571 24 1.1841 0.0318 1.1841 1.0882
No log 0.9286 26 1.1531 0.3229 1.1531 1.0738
No log 1.0 28 0.9933 0.2967 0.9933 0.9966
No log 1.0714 30 0.9777 0.2365 0.9777 0.9888
No log 1.1429 32 0.9964 0.3056 0.9964 0.9982
No log 1.2143 34 1.1607 0.2380 1.1607 1.0774
No log 1.2857 36 1.3253 0.2011 1.3253 1.1512
No log 1.3571 38 1.1250 0.2457 1.1250 1.0607
No log 1.4286 40 1.0607 0.2953 1.0607 1.0299
No log 1.5 42 1.0331 0.3521 1.0331 1.0164
No log 1.5714 44 1.1041 0.2037 1.1041 1.0508
No log 1.6429 46 1.0245 0.2887 1.0245 1.0122
No log 1.7143 48 1.0988 0.2973 1.0988 1.0482
No log 1.7857 50 1.3634 0.2217 1.3634 1.1677
No log 1.8571 52 1.1732 0.4011 1.1732 1.0831
No log 1.9286 54 0.9775 0.2670 0.9775 0.9887
No log 2.0 56 0.9659 0.2643 0.9659 0.9828
No log 2.0714 58 0.9634 0.2416 0.9634 0.9815
No log 2.1429 60 1.0280 0.3202 1.0280 1.0139
No log 2.2143 62 1.0186 0.2871 1.0186 1.0092
No log 2.2857 64 0.9823 0.3017 0.9823 0.9911
No log 2.3571 66 0.9970 0.2910 0.9970 0.9985
No log 2.4286 68 1.0315 0.2910 1.0315 1.0156
No log 2.5 70 1.0214 0.3035 1.0214 1.0106
No log 2.5714 72 1.1504 0.3025 1.1504 1.0726
No log 2.6429 74 1.3487 0.3194 1.3487 1.1613
No log 2.7143 76 1.2128 0.2958 1.2128 1.1013
No log 2.7857 78 0.9724 0.3216 0.9724 0.9861
No log 2.8571 80 0.9799 0.3921 0.9799 0.9899
No log 2.9286 82 1.2442 0.25 1.2442 1.1154
No log 3.0 84 1.2563 0.2704 1.2563 1.1208
No log 3.0714 86 0.9986 0.3760 0.9986 0.9993
No log 3.1429 88 0.8454 0.3713 0.8454 0.9194
No log 3.2143 90 1.0219 0.4104 1.0219 1.0109
No log 3.2857 92 1.1751 0.3538 1.1751 1.0840
No log 3.3571 94 0.9848 0.3459 0.9848 0.9924
No log 3.4286 96 0.8765 0.3576 0.8765 0.9362
No log 3.5 98 0.9907 0.3822 0.9907 0.9954
No log 3.5714 100 0.9574 0.3326 0.9574 0.9785
No log 3.6429 102 0.9662 0.3321 0.9662 0.9829
No log 3.7143 104 1.0412 0.3250 1.0412 1.0204
No log 3.7857 106 0.9477 0.3402 0.9477 0.9735
No log 3.8571 108 1.0077 0.4603 1.0077 1.0038
No log 3.9286 110 1.0284 0.4606 1.0284 1.0141
No log 4.0 112 0.9548 0.3430 0.9548 0.9771
No log 4.0714 114 1.0292 0.3424 1.0292 1.0145
No log 4.1429 116 1.0584 0.3188 1.0584 1.0288
No log 4.2143 118 1.0654 0.3219 1.0654 1.0322
No log 4.2857 120 0.9685 0.5124 0.9685 0.9841
No log 4.3571 122 0.9997 0.4607 0.9997 0.9998
No log 4.4286 124 0.9305 0.4902 0.9305 0.9646
No log 4.5 126 1.1710 0.3210 1.1710 1.0821
No log 4.5714 128 1.2654 0.3361 1.2654 1.1249
No log 4.6429 130 1.0963 0.3606 1.0963 1.0470
No log 4.7143 132 1.0112 0.4166 1.0112 1.0056
No log 4.7857 134 1.0324 0.3552 1.0324 1.0161
No log 4.8571 136 1.0035 0.3877 1.0035 1.0018
No log 4.9286 138 0.9283 0.4540 0.9283 0.9635
No log 5.0 140 0.9576 0.3798 0.9576 0.9786
No log 5.0714 142 0.9555 0.3487 0.9555 0.9775
No log 5.1429 144 0.8853 0.2988 0.8853 0.9409
No log 5.2143 146 0.9301 0.3085 0.9301 0.9644
No log 5.2857 148 0.9530 0.3826 0.9530 0.9762
No log 5.3571 150 0.8961 0.3948 0.8961 0.9466
No log 5.4286 152 0.9174 0.4058 0.9174 0.9578
No log 5.5 154 0.9304 0.4676 0.9304 0.9645
No log 5.5714 156 1.1454 0.3460 1.1454 1.0702
No log 5.6429 158 1.1595 0.3460 1.1595 1.0768
No log 5.7143 160 0.9528 0.4758 0.9528 0.9761
No log 5.7857 162 1.1039 0.4358 1.1039 1.0507
No log 5.8571 164 1.4429 0.2918 1.4429 1.2012
No log 5.9286 166 1.3030 0.3599 1.3030 1.1415
No log 6.0 168 0.9984 0.4136 0.9984 0.9992
No log 6.0714 170 0.9101 0.3979 0.9101 0.9540
No log 6.1429 172 0.9766 0.4356 0.9766 0.9882
No log 6.2143 174 0.9487 0.3560 0.9487 0.9740
No log 6.2857 176 0.9136 0.5022 0.9136 0.9558
No log 6.3571 178 0.9426 0.4089 0.9426 0.9709
No log 6.4286 180 0.9127 0.4816 0.9127 0.9553
No log 6.5 182 0.9549 0.4721 0.9549 0.9772
No log 6.5714 184 0.9115 0.4721 0.9115 0.9547
No log 6.6429 186 0.8572 0.3596 0.8572 0.9258
No log 6.7143 188 1.0225 0.4171 1.0225 1.0112
No log 6.7857 190 1.1125 0.3667 1.1125 1.0547
No log 6.8571 192 0.9600 0.4479 0.9600 0.9798
No log 6.9286 194 0.8400 0.3820 0.8400 0.9165
No log 7.0 196 0.9670 0.4666 0.9670 0.9834
No log 7.0714 198 1.0408 0.4449 1.0408 1.0202
No log 7.1429 200 0.9035 0.4821 0.9035 0.9505
No log 7.2143 202 0.8739 0.4208 0.8739 0.9348
No log 7.2857 204 0.9973 0.3572 0.9973 0.9986
No log 7.3571 206 1.1658 0.4353 1.1658 1.0797
No log 7.4286 208 1.0905 0.4081 1.0905 1.0442
No log 7.5 210 0.9295 0.4644 0.9295 0.9641
No log 7.5714 212 0.9314 0.4863 0.9314 0.9651
No log 7.6429 214 0.9368 0.4863 0.9368 0.9679
No log 7.7143 216 0.9140 0.4603 0.9140 0.9560
No log 7.7857 218 0.9139 0.4617 0.9139 0.9560
No log 7.8571 220 0.9110 0.4852 0.9110 0.9544
No log 7.9286 222 0.8885 0.5339 0.8885 0.9426
No log 8.0 224 0.8897 0.4634 0.8897 0.9432
No log 8.0714 226 0.8824 0.4634 0.8824 0.9394
No log 8.1429 228 0.8854 0.4792 0.8854 0.9410
No log 8.2143 230 0.9023 0.4617 0.9023 0.9499
No log 8.2857 232 0.9238 0.4297 0.9238 0.9611
No log 8.3571 234 0.9903 0.4412 0.9903 0.9951
No log 8.4286 236 0.9639 0.4430 0.9639 0.9818
No log 8.5 238 0.9645 0.4245 0.9645 0.9821
No log 8.5714 240 0.9908 0.4527 0.9908 0.9954
No log 8.6429 242 0.9837 0.4277 0.9837 0.9918
No log 8.7143 244 0.9294 0.4550 0.9294 0.9641
No log 8.7857 246 0.9549 0.4175 0.9549 0.9772
No log 8.8571 248 0.9514 0.4766 0.9514 0.9754
No log 8.9286 250 0.9563 0.4705 0.9563 0.9779
No log 9.0 252 1.0849 0.3617 1.0849 1.0416
No log 9.0714 254 1.2136 0.2783 1.2136 1.1016
No log 9.1429 256 1.1322 0.2443 1.1322 1.0641
No log 9.2143 258 0.9464 0.4393 0.9464 0.9728
No log 9.2857 260 0.9337 0.4143 0.9337 0.9663
No log 9.3571 262 1.0382 0.4468 1.0382 1.0189
No log 9.4286 264 0.9946 0.4581 0.9946 0.9973
No log 9.5 266 0.8843 0.4514 0.8843 0.9404
No log 9.5714 268 0.8780 0.4824 0.8780 0.9370
No log 9.6429 270 0.8823 0.4852 0.8823 0.9393
No log 9.7143 272 0.8678 0.4792 0.8678 0.9315
No log 9.7857 274 0.8635 0.4648 0.8635 0.9293
No log 9.8571 276 0.8853 0.5356 0.8853 0.9409
No log 9.9286 278 0.8139 0.4768 0.8139 0.9022
No log 10.0 280 0.8148 0.4034 0.8148 0.9027
No log 10.0714 282 0.9044 0.4039 0.9044 0.9510
No log 10.1429 284 0.8852 0.3777 0.8852 0.9408
No log 10.2143 286 0.8242 0.4252 0.8242 0.9078
No log 10.2857 288 0.8363 0.4824 0.8363 0.9145
No log 10.3571 290 0.8855 0.5372 0.8855 0.9410
No log 10.4286 292 0.9182 0.5268 0.9182 0.9583
No log 10.5 294 0.9354 0.5143 0.9354 0.9672
No log 10.5714 296 0.9231 0.5550 0.9231 0.9608
No log 10.6429 298 0.8679 0.5702 0.8679 0.9316
No log 10.7143 300 0.8105 0.5215 0.8105 0.9003
No log 10.7857 302 0.7977 0.5304 0.7977 0.8932
No log 10.8571 304 0.8057 0.5186 0.8057 0.8976
No log 10.9286 306 0.8015 0.5226 0.8015 0.8952
No log 11.0 308 0.8495 0.5178 0.8495 0.9217
No log 11.0714 310 0.9291 0.4932 0.9291 0.9639
No log 11.1429 312 0.8919 0.5279 0.8919 0.9444
No log 11.2143 314 0.8505 0.5683 0.8505 0.9222
No log 11.2857 316 0.8463 0.5683 0.8463 0.9200
No log 11.3571 318 0.8570 0.5303 0.8570 0.9258
No log 11.4286 320 0.8350 0.4889 0.8350 0.9138
No log 11.5 322 0.8077 0.5036 0.8077 0.8987
No log 11.5714 324 0.8383 0.3925 0.8383 0.9156
No log 11.6429 326 0.8308 0.4048 0.8308 0.9115
No log 11.7143 328 0.8082 0.3802 0.8082 0.8990
No log 11.7857 330 0.8095 0.4583 0.8095 0.8997
No log 11.8571 332 0.8365 0.4544 0.8365 0.9146
No log 11.9286 334 1.0127 0.4627 1.0127 1.0063
No log 12.0 336 1.1572 0.3677 1.1572 1.0757
No log 12.0714 338 1.0431 0.4215 1.0431 1.0213
No log 12.1429 340 0.8744 0.5053 0.8744 0.9351
No log 12.2143 342 0.8663 0.4202 0.8663 0.9308
No log 12.2857 344 0.8445 0.5340 0.8445 0.9190
No log 12.3571 346 0.8309 0.5250 0.8309 0.9115
No log 12.4286 348 0.8221 0.5275 0.8221 0.9067
No log 12.5 350 0.8024 0.5490 0.8024 0.8957
No log 12.5714 352 0.7715 0.5029 0.7715 0.8784
No log 12.6429 354 0.7387 0.5152 0.7387 0.8595
No log 12.7143 356 0.7199 0.5275 0.7199 0.8484
No log 12.7857 358 0.7129 0.5155 0.7129 0.8443
No log 12.8571 360 0.7197 0.5135 0.7197 0.8484
No log 12.9286 362 0.7253 0.5500 0.7253 0.8516
No log 13.0 364 0.7353 0.5406 0.7353 0.8575
No log 13.0714 366 0.7382 0.5500 0.7382 0.8592
No log 13.1429 368 0.7869 0.5517 0.7869 0.8871
No log 13.2143 370 0.8239 0.4962 0.8239 0.9077
No log 13.2857 372 0.8786 0.4349 0.8786 0.9373
No log 13.3571 374 0.8972 0.4579 0.8972 0.9472
No log 13.4286 376 0.9180 0.4796 0.9180 0.9581
No log 13.5 378 0.8272 0.3861 0.8272 0.9095
No log 13.5714 380 0.8091 0.3615 0.8091 0.8995
No log 13.6429 382 0.7935 0.4776 0.7935 0.8908
No log 13.7143 384 0.8041 0.4261 0.8041 0.8967
No log 13.7857 386 0.8268 0.3502 0.8268 0.9093
No log 13.8571 388 0.8296 0.4645 0.8296 0.9108
No log 13.9286 390 0.8431 0.5174 0.8431 0.9182
No log 14.0 392 0.9341 0.3659 0.9341 0.9665
No log 14.0714 394 0.9898 0.3849 0.9898 0.9949
No log 14.1429 396 0.9340 0.3578 0.9340 0.9664
No log 14.2143 398 0.8413 0.4872 0.8413 0.9172
No log 14.2857 400 0.8042 0.6041 0.8042 0.8968
No log 14.3571 402 0.7964 0.6041 0.7964 0.8924
No log 14.4286 404 0.7897 0.6041 0.7897 0.8887
No log 14.5 406 0.7864 0.4952 0.7864 0.8868
No log 14.5714 408 0.8301 0.4202 0.8301 0.9111
No log 14.6429 410 0.8813 0.4671 0.8813 0.9388
No log 14.7143 412 0.8373 0.4191 0.8373 0.9150
No log 14.7857 414 0.7737 0.5773 0.7737 0.8796
No log 14.8571 416 0.8536 0.5032 0.8536 0.9239
No log 14.9286 418 0.8956 0.4694 0.8956 0.9464
No log 15.0 420 0.8443 0.5244 0.8443 0.9189
No log 15.0714 422 0.7727 0.5887 0.7727 0.8790
No log 15.1429 424 0.7643 0.5898 0.7643 0.8742
No log 15.2143 426 0.7866 0.5774 0.7866 0.8869
No log 15.2857 428 0.8170 0.5168 0.8170 0.9039
No log 15.3571 430 0.8459 0.5374 0.8459 0.9197
No log 15.4286 432 0.8347 0.5892 0.8347 0.9136
No log 15.5 434 0.8046 0.5338 0.8046 0.8970
No log 15.5714 436 0.7864 0.5785 0.7864 0.8868
No log 15.6429 438 0.7985 0.5731 0.7985 0.8936
No log 15.7143 440 0.8376 0.5469 0.8376 0.9152
No log 15.7857 442 0.8715 0.5480 0.8715 0.9336
No log 15.8571 444 0.9275 0.5014 0.9275 0.9631
No log 15.9286 446 0.9461 0.4622 0.9461 0.9727
No log 16.0 448 0.9524 0.4757 0.9524 0.9759
No log 16.0714 450 0.8829 0.5055 0.8829 0.9396
No log 16.1429 452 0.8393 0.5471 0.8393 0.9161
No log 16.2143 454 0.8334 0.4969 0.8334 0.9129
No log 16.2857 456 0.8078 0.5125 0.8078 0.8988
No log 16.3571 458 0.7986 0.5379 0.7986 0.8936
No log 16.4286 460 0.8082 0.5076 0.8082 0.8990
No log 16.5 462 0.8089 0.5379 0.8089 0.8994
No log 16.5714 464 0.8105 0.5263 0.8105 0.9003
No log 16.6429 466 0.8079 0.5807 0.8079 0.8988
No log 16.7143 468 0.8145 0.5807 0.8145 0.9025
No log 16.7857 470 0.7981 0.5712 0.7981 0.8934
No log 16.8571 472 0.7987 0.5807 0.7987 0.8937
No log 16.9286 474 0.8182 0.5465 0.8182 0.9045
No log 17.0 476 0.8252 0.5255 0.8252 0.9084
No log 17.0714 478 0.7831 0.5186 0.7831 0.8849
No log 17.1429 480 0.7266 0.5923 0.7266 0.8524
No log 17.2143 482 0.7389 0.5406 0.7389 0.8596
No log 17.2857 484 0.7797 0.5312 0.7797 0.8830
No log 17.3571 486 0.7481 0.5391 0.7481 0.8649
No log 17.4286 488 0.7491 0.6157 0.7491 0.8655
No log 17.5 490 0.8228 0.5455 0.8228 0.9071
No log 17.5714 492 0.8463 0.5455 0.8463 0.9200
No log 17.6429 494 0.8034 0.5860 0.8034 0.8963
No log 17.7143 496 0.7588 0.5364 0.7588 0.8711
No log 17.7857 498 0.7792 0.5406 0.7792 0.8827
0.2992 17.8571 500 0.7588 0.5406 0.7588 0.8711
0.2992 17.9286 502 0.7292 0.6018 0.7292 0.8539
0.2992 18.0 504 0.7807 0.5798 0.7807 0.8836
0.2992 18.0714 506 0.7964 0.5366 0.7964 0.8924
0.2992 18.1429 508 0.7666 0.5730 0.7666 0.8755
0.2992 18.2143 510 0.7590 0.5552 0.7590 0.8712
0.2992 18.2857 512 0.7590 0.5434 0.7590 0.8712
0.2992 18.3571 514 0.7592 0.5125 0.7592 0.8713
0.2992 18.4286 516 0.7756 0.4975 0.7756 0.8807
0.2992 18.5 518 0.8017 0.5178 0.8017 0.8954
0.2992 18.5714 520 0.8183 0.5064 0.8183 0.9046
0.2992 18.6429 522 0.7945 0.5528 0.7945 0.8914
0.2992 18.7143 524 0.8051 0.5312 0.8051 0.8973
0.2992 18.7857 526 0.8129 0.5312 0.8129 0.9016
0.2992 18.8571 528 0.7839 0.5080 0.7839 0.8854
0.2992 18.9286 530 0.7970 0.4008 0.7970 0.8927
0.2992 19.0 532 0.8769 0.4489 0.8769 0.9364

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k8_task5_organization

Finetuned
(4019)
this model