ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k9_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7293
  • Qwk: 0.3382
  • Mse: 0.7293
  • Rmse: 0.8540

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0645 2 2.5781 -0.0407 2.5781 1.6056
No log 0.1290 4 1.2852 0.0479 1.2852 1.1337
No log 0.1935 6 0.9413 0.0535 0.9413 0.9702
No log 0.2581 8 1.0206 -0.1464 1.0206 1.0102
No log 0.3226 10 1.0062 -0.0831 1.0062 1.0031
No log 0.3871 12 1.0075 -0.0408 1.0075 1.0038
No log 0.4516 14 0.9233 0.0 0.9233 0.9609
No log 0.5161 16 0.8964 0.0 0.8964 0.9468
No log 0.5806 18 0.8970 0.0 0.8970 0.9471
No log 0.6452 20 0.8868 0.0 0.8868 0.9417
No log 0.7097 22 0.8493 0.0 0.8493 0.9216
No log 0.7742 24 0.8199 0.0 0.8199 0.9055
No log 0.8387 26 0.8346 0.0679 0.8346 0.9136
No log 0.9032 28 0.8832 0.1815 0.8832 0.9398
No log 0.9677 30 0.9352 0.2193 0.9352 0.9671
No log 1.0323 32 0.8647 0.1268 0.8647 0.9299
No log 1.0968 34 0.9041 0.0 0.9041 0.9508
No log 1.1613 36 0.9477 0.0 0.9477 0.9735
No log 1.2258 38 0.8550 0.0 0.8550 0.9247
No log 1.2903 40 0.7927 0.0359 0.7927 0.8903
No log 1.3548 42 0.8527 0.1918 0.8527 0.9234
No log 1.4194 44 0.8581 0.1815 0.8581 0.9263
No log 1.4839 46 0.7934 0.2027 0.7934 0.8907
No log 1.5484 48 0.8316 -0.0483 0.8316 0.9119
No log 1.6129 50 0.8412 0.0 0.8412 0.9172
No log 1.6774 52 0.7371 0.0 0.7371 0.8585
No log 1.7419 54 0.7217 0.0 0.7217 0.8495
No log 1.8065 56 0.7414 0.0 0.7414 0.8611
No log 1.8710 58 0.8338 0.0522 0.8338 0.9131
No log 1.9355 60 0.9779 0.1618 0.9779 0.9889
No log 2.0 62 0.9826 0.1304 0.9826 0.9913
No log 2.0645 64 0.8064 0.0481 0.8064 0.8980
No log 2.1290 66 0.7668 -0.0054 0.7668 0.8756
No log 2.1935 68 0.7707 0.0 0.7707 0.8779
No log 2.2581 70 0.8179 0.0937 0.8179 0.9044
No log 2.3226 72 0.8835 0.1352 0.8835 0.9399
No log 2.3871 74 0.8282 0.0937 0.8282 0.9101
No log 2.4516 76 0.7830 0.0 0.7830 0.8849
No log 2.5161 78 0.7480 0.0 0.7480 0.8649
No log 2.5806 80 0.7501 0.2145 0.7501 0.8661
No log 2.6452 82 0.7950 0.1866 0.7950 0.8917
No log 2.7097 84 0.8218 0.2261 0.8218 0.9065
No log 2.7742 86 0.8335 0.1923 0.8335 0.9129
No log 2.8387 88 0.9062 0.1277 0.9062 0.9519
No log 2.9032 90 1.0388 0.2113 1.0388 1.0192
No log 2.9677 92 1.0083 0.1581 1.0083 1.0042
No log 3.0323 94 0.8857 0.1339 0.8857 0.9411
No log 3.0968 96 0.8648 0.2057 0.8648 0.9299
No log 3.1613 98 0.8892 0.1700 0.8892 0.9430
No log 3.2258 100 0.8565 0.1673 0.8565 0.9255
No log 3.2903 102 0.7764 0.0804 0.7764 0.8811
No log 3.3548 104 0.7586 0.0 0.7586 0.8710
No log 3.4194 106 0.7334 0.0428 0.7334 0.8564
No log 3.4839 108 0.7062 0.1456 0.7062 0.8403
No log 3.5484 110 0.8441 0.2904 0.8441 0.9188
No log 3.6129 112 0.8574 0.3105 0.8574 0.9259
No log 3.6774 114 0.6721 0.3465 0.6721 0.8198
No log 3.7419 116 0.6479 0.1184 0.6479 0.8049
No log 3.8065 118 0.6910 0.3040 0.6910 0.8313
No log 3.8710 120 0.7645 0.2555 0.7645 0.8743
No log 3.9355 122 0.9740 0.2627 0.9740 0.9869
No log 4.0 124 0.9808 0.2416 0.9808 0.9904
No log 4.0645 126 0.8194 0.1958 0.8194 0.9052
No log 4.1290 128 0.7216 0.3863 0.7216 0.8495
No log 4.1935 130 0.7432 0.3754 0.7432 0.8621
No log 4.2581 132 0.7385 0.3031 0.7385 0.8594
No log 4.3226 134 0.7386 0.2715 0.7386 0.8594
No log 4.3871 136 0.7664 0.3551 0.7664 0.8754
No log 4.4516 138 0.7803 0.2652 0.7803 0.8834
No log 4.5161 140 0.8241 0.3225 0.8241 0.9078
No log 4.5806 142 0.9734 0.2166 0.9734 0.9866
No log 4.6452 144 1.0714 0.2998 1.0714 1.0351
No log 4.7097 146 1.0147 0.2026 1.0147 1.0073
No log 4.7742 148 0.8747 0.3433 0.8747 0.9352
No log 4.8387 150 0.8548 0.2914 0.8548 0.9246
No log 4.9032 152 0.8417 0.2962 0.8417 0.9174
No log 4.9677 154 0.8133 0.3841 0.8133 0.9018
No log 5.0323 156 0.7561 0.2530 0.7561 0.8695
No log 5.0968 158 0.7585 0.2239 0.7585 0.8709
No log 5.1613 160 0.7573 0.2239 0.7573 0.8702
No log 5.2258 162 0.7695 0.3034 0.7695 0.8772
No log 5.2903 164 0.8500 0.3130 0.8500 0.9220
No log 5.3548 166 0.8860 0.3143 0.8860 0.9413
No log 5.4194 168 0.9066 0.3500 0.9066 0.9522
No log 5.4839 170 0.8947 0.3112 0.8947 0.9459
No log 5.5484 172 0.9066 0.2941 0.9066 0.9522
No log 5.6129 174 1.0192 0.3580 1.0192 1.0095
No log 5.6774 176 1.0235 0.3671 1.0235 1.0117
No log 5.7419 178 0.8991 0.3906 0.8991 0.9482
No log 5.8065 180 0.7897 0.2349 0.7897 0.8887
No log 5.8710 182 0.7640 0.2947 0.7640 0.8741
No log 5.9355 184 0.9083 0.3847 0.9083 0.9531
No log 6.0 186 1.0110 0.3662 1.0110 1.0055
No log 6.0645 188 0.8329 0.3521 0.8329 0.9127
No log 6.1290 190 0.6887 0.2591 0.6887 0.8299
No log 6.1935 192 0.6792 0.3336 0.6792 0.8241
No log 6.2581 194 0.6994 0.3574 0.6994 0.8363
No log 6.3226 196 0.7237 0.3363 0.7237 0.8507
No log 6.3871 198 0.7813 0.3316 0.7813 0.8839
No log 6.4516 200 0.8230 0.2774 0.8230 0.9072
No log 6.5161 202 0.7865 0.2987 0.7865 0.8868
No log 6.5806 204 0.8167 0.2798 0.8167 0.9037
No log 6.6452 206 0.8051 0.2931 0.8051 0.8973
No log 6.7097 208 0.8186 0.3388 0.8186 0.9048
No log 6.7742 210 0.7938 0.3243 0.7938 0.8910
No log 6.8387 212 0.8161 0.3388 0.8161 0.9034
No log 6.9032 214 0.8671 0.3376 0.8671 0.9312
No log 6.9677 216 0.9037 0.3389 0.9037 0.9506
No log 7.0323 218 0.8952 0.3173 0.8952 0.9461
No log 7.0968 220 0.8775 0.3330 0.8775 0.9367
No log 7.1613 222 0.8848 0.2998 0.8848 0.9406
No log 7.2258 224 0.8015 0.2831 0.8015 0.8953
No log 7.2903 226 0.7608 0.3175 0.7608 0.8723
No log 7.3548 228 0.7722 0.3659 0.7722 0.8788
No log 7.4194 230 0.7571 0.3097 0.7571 0.8701
No log 7.4839 232 0.7025 0.3530 0.7025 0.8381
No log 7.5484 234 0.7718 0.3590 0.7718 0.8785
No log 7.6129 236 0.9170 0.2832 0.9170 0.9576
No log 7.6774 238 0.8828 0.3586 0.8828 0.9396
No log 7.7419 240 0.8315 0.3650 0.8315 0.9119
No log 7.8065 242 0.8768 0.3295 0.8768 0.9364
No log 7.8710 244 0.9539 0.3195 0.9539 0.9767
No log 7.9355 246 0.9793 0.2873 0.9793 0.9896
No log 8.0 248 0.8838 0.3569 0.8838 0.9401
No log 8.0645 250 0.8450 0.3880 0.8450 0.9192
No log 8.1290 252 0.8251 0.3492 0.8251 0.9083
No log 8.1935 254 0.9171 0.3173 0.9171 0.9577
No log 8.2581 256 1.0398 0.2485 1.0398 1.0197
No log 8.3226 258 0.9962 0.2853 0.9962 0.9981
No log 8.3871 260 0.7978 0.3526 0.7978 0.8932
No log 8.4516 262 0.7380 0.2563 0.7380 0.8591
No log 8.5161 264 0.7565 0.3163 0.7565 0.8698
No log 8.5806 266 0.7415 0.3442 0.7415 0.8611
No log 8.6452 268 0.7502 0.3136 0.7502 0.8661
No log 8.7097 270 0.7372 0.3285 0.7372 0.8586
No log 8.7742 272 0.7355 0.2888 0.7355 0.8576
No log 8.8387 274 0.7610 0.2805 0.7610 0.8723
No log 8.9032 276 0.7888 0.3677 0.7888 0.8881
No log 8.9677 278 0.7345 0.3425 0.7345 0.8570
No log 9.0323 280 0.7415 0.3688 0.7415 0.8611
No log 9.0968 282 0.7557 0.3099 0.7557 0.8693
No log 9.1613 284 0.7079 0.4397 0.7079 0.8414
No log 9.2258 286 0.6928 0.2929 0.6928 0.8324
No log 9.2903 288 0.6976 0.2867 0.6976 0.8352
No log 9.3548 290 0.7155 0.3239 0.7155 0.8458
No log 9.4194 292 0.7173 0.3211 0.7173 0.8469
No log 9.4839 294 0.7012 0.2867 0.7012 0.8374
No log 9.5484 296 0.7081 0.2310 0.7081 0.8415
No log 9.6129 298 0.7310 0.2310 0.7310 0.8550
No log 9.6774 300 0.7493 0.2310 0.7493 0.8656
No log 9.7419 302 0.8060 0.2262 0.8060 0.8978
No log 9.8065 304 0.8866 0.2781 0.8866 0.9416
No log 9.8710 306 0.8137 0.1992 0.8137 0.9020
No log 9.9355 308 0.7904 0.2895 0.7904 0.8891
No log 10.0 310 0.7777 0.2661 0.7777 0.8818
No log 10.0645 312 0.7951 0.2953 0.7951 0.8917
No log 10.1290 314 0.8044 0.3287 0.8044 0.8969
No log 10.1935 316 0.7683 0.3688 0.7683 0.8765
No log 10.2581 318 0.7506 0.2913 0.7506 0.8664
No log 10.3226 320 0.7613 0.3545 0.7613 0.8725
No log 10.3871 322 0.7580 0.2621 0.7580 0.8706
No log 10.4516 324 0.7498 0.3253 0.7498 0.8659
No log 10.5161 326 0.7422 0.2535 0.7422 0.8615
No log 10.5806 328 0.7418 0.1720 0.7418 0.8613
No log 10.6452 330 0.7423 0.2092 0.7423 0.8615
No log 10.7097 332 0.7577 0.3267 0.7577 0.8705
No log 10.7742 334 0.7686 0.3238 0.7686 0.8767
No log 10.8387 336 0.7327 0.2043 0.7327 0.8560
No log 10.9032 338 0.7337 0.2078 0.7337 0.8566
No log 10.9677 340 0.7375 0.3224 0.7375 0.8588
No log 11.0323 342 0.7462 0.3155 0.7462 0.8638
No log 11.0968 344 0.7622 0.3088 0.7622 0.8731
No log 11.1613 346 0.7250 0.3434 0.7250 0.8514
No log 11.2258 348 0.7427 0.2889 0.7427 0.8618
No log 11.2903 350 0.7461 0.3335 0.7461 0.8637
No log 11.3548 352 0.8013 0.3231 0.8013 0.8951
No log 11.4194 354 0.8968 0.2779 0.8968 0.9470
No log 11.4839 356 0.8349 0.2590 0.8349 0.9137
No log 11.5484 358 0.7453 0.3813 0.7453 0.8633
No log 11.6129 360 0.7487 0.2652 0.7487 0.8653
No log 11.6774 362 0.7695 0.2204 0.7695 0.8772
No log 11.7419 364 0.7356 0.3702 0.7356 0.8577
No log 11.8065 366 0.7931 0.3121 0.7931 0.8905
No log 11.8710 368 0.9270 0.2651 0.9270 0.9628
No log 11.9355 370 0.9101 0.2928 0.9101 0.9540
No log 12.0 372 0.7925 0.2920 0.7925 0.8902
No log 12.0645 374 0.7747 0.3541 0.7747 0.8802
No log 12.1290 376 0.7859 0.3331 0.7859 0.8865
No log 12.1935 378 0.7453 0.3582 0.7453 0.8633
No log 12.2581 380 0.7398 0.2445 0.7398 0.8601
No log 12.3226 382 0.8391 0.3280 0.8391 0.9160
No log 12.3871 384 0.9107 0.3310 0.9107 0.9543
No log 12.4516 386 0.7940 0.2843 0.7940 0.8910
No log 12.5161 388 0.6618 0.3224 0.6618 0.8135
No log 12.5806 390 0.6468 0.4276 0.6468 0.8042
No log 12.6452 392 0.6458 0.4059 0.6458 0.8036
No log 12.7097 394 0.6987 0.3433 0.6987 0.8359
No log 12.7742 396 0.7549 0.3344 0.7549 0.8688
No log 12.8387 398 0.7126 0.3344 0.7126 0.8441
No log 12.9032 400 0.6319 0.4747 0.6319 0.7949
No log 12.9677 402 0.6259 0.4809 0.6259 0.7912
No log 13.0323 404 0.6322 0.4762 0.6322 0.7951
No log 13.0968 406 0.6425 0.4342 0.6425 0.8016
No log 13.1613 408 0.6753 0.4125 0.6753 0.8218
No log 13.2258 410 0.6937 0.3793 0.6937 0.8329
No log 13.2903 412 0.6805 0.4322 0.6805 0.8249
No log 13.3548 414 0.6780 0.4322 0.6780 0.8234
No log 13.4194 416 0.7029 0.4336 0.7029 0.8384
No log 13.4839 418 0.7125 0.3793 0.7125 0.8441
No log 13.5484 420 0.7673 0.3633 0.7673 0.8760
No log 13.6129 422 0.8067 0.3798 0.8067 0.8982
No log 13.6774 424 0.8631 0.3889 0.8631 0.9290
No log 13.7419 426 0.8950 0.4468 0.8950 0.9460
No log 13.8065 428 0.8086 0.3486 0.8086 0.8992
No log 13.8710 430 0.7689 0.3470 0.7689 0.8769
No log 13.9355 432 0.8466 0.4486 0.8466 0.9201
No log 14.0 434 0.9812 0.4125 0.9812 0.9905
No log 14.0645 436 0.9273 0.4426 0.9273 0.9630
No log 14.1290 438 0.7441 0.4025 0.7441 0.8626
No log 14.1935 440 0.6708 0.3791 0.6708 0.8190
No log 14.2581 442 0.6638 0.4329 0.6638 0.8147
No log 14.3226 444 0.6451 0.3425 0.6451 0.8032
No log 14.3871 446 0.6571 0.3840 0.6571 0.8106
No log 14.4516 448 0.7044 0.3287 0.7044 0.8393
No log 14.5161 450 0.7909 0.3520 0.7909 0.8893
No log 14.5806 452 0.7630 0.3095 0.7630 0.8735
No log 14.6452 454 0.6741 0.3910 0.6741 0.8211
No log 14.7097 456 0.6741 0.3321 0.6741 0.8210
No log 14.7742 458 0.6721 0.2652 0.6721 0.8198
No log 14.8387 460 0.6721 0.2929 0.6721 0.8198
No log 14.9032 462 0.7141 0.3966 0.7141 0.8450
No log 14.9677 464 0.7180 0.4234 0.7180 0.8473
No log 15.0323 466 0.6795 0.3762 0.6795 0.8243
No log 15.0968 468 0.6647 0.2744 0.6647 0.8153
No log 15.1613 470 0.6574 0.3702 0.6574 0.8108
No log 15.2258 472 0.6834 0.3966 0.6834 0.8267
No log 15.2903 474 0.7326 0.4212 0.7326 0.8559
No log 15.3548 476 0.6925 0.3966 0.6925 0.8321
No log 15.4194 478 0.6537 0.3835 0.6537 0.8085
No log 15.4839 480 0.6730 0.3704 0.6730 0.8204
No log 15.5484 482 0.6853 0.3259 0.6853 0.8278
No log 15.6129 484 0.6791 0.3814 0.6791 0.8240
No log 15.6774 486 0.7545 0.3740 0.7545 0.8686
No log 15.7419 488 0.7741 0.3510 0.7741 0.8798
No log 15.8065 490 0.7128 0.3492 0.7128 0.8443
No log 15.8710 492 0.6668 0.3769 0.6668 0.8166
No log 15.9355 494 0.6578 0.4595 0.6578 0.8110
No log 16.0 496 0.6502 0.4006 0.6502 0.8063
No log 16.0645 498 0.6634 0.3299 0.6634 0.8145
0.314 16.1290 500 0.6994 0.3590 0.6994 0.8363
0.314 16.1935 502 0.7304 0.3149 0.7304 0.8547
0.314 16.2581 504 0.7448 0.3379 0.7448 0.8630
0.314 16.3226 506 0.8088 0.3868 0.8088 0.8994
0.314 16.3871 508 0.8193 0.3868 0.8193 0.9052
0.314 16.4516 510 0.7393 0.4350 0.7393 0.8598
0.314 16.5161 512 0.7275 0.3867 0.7275 0.8529
0.314 16.5806 514 0.7590 0.3985 0.7590 0.8712
0.314 16.6452 516 0.7554 0.3867 0.7554 0.8692
0.314 16.7097 518 0.7301 0.3867 0.7301 0.8545
0.314 16.7742 520 0.6997 0.2530 0.6997 0.8365
0.314 16.8387 522 0.7140 0.2530 0.7140 0.8450
0.314 16.9032 524 0.7570 0.3355 0.7570 0.8701
0.314 16.9677 526 0.7691 0.3121 0.7691 0.8770
0.314 17.0323 528 0.7293 0.3382 0.7293 0.8540

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k9_task7_organization

Finetuned
(4019)
this model