ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k9_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8787
  • Qwk: 0.3622
  • Mse: 0.8787
  • Rmse: 0.9374

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0476 2 4.3477 -0.0335 4.3477 2.0851
No log 0.0952 4 2.9581 0.0111 2.9581 1.7199
No log 0.1429 6 2.1094 -0.0053 2.1094 1.4524
No log 0.1905 8 1.2592 0.0640 1.2592 1.1221
No log 0.2381 10 1.2560 0.0427 1.2560 1.1207
No log 0.2857 12 1.2022 0.0548 1.2022 1.0965
No log 0.3333 14 1.1509 0.0914 1.1509 1.0728
No log 0.3810 16 1.2972 0.0256 1.2972 1.1390
No log 0.4286 18 1.1945 0.0612 1.1945 1.0929
No log 0.4762 20 1.2011 0.1028 1.2011 1.0960
No log 0.5238 22 1.6233 0.1083 1.6233 1.2741
No log 0.5714 24 1.6716 0.0749 1.6716 1.2929
No log 0.6190 26 1.2884 0.0922 1.2884 1.1351
No log 0.6667 28 1.0803 0.2341 1.0803 1.0394
No log 0.7143 30 1.0669 0.2351 1.0669 1.0329
No log 0.7619 32 1.1707 0.1876 1.1707 1.0820
No log 0.8095 34 1.5061 0.1966 1.5061 1.2272
No log 0.8571 36 1.3713 0.1523 1.3713 1.1710
No log 0.9048 38 1.0819 0.1521 1.0819 1.0401
No log 0.9524 40 1.2122 0.2587 1.2122 1.1010
No log 1.0 42 1.4464 0.0946 1.4464 1.2027
No log 1.0476 44 1.5026 0.0855 1.5026 1.2258
No log 1.0952 46 1.2681 0.0661 1.2681 1.1261
No log 1.1429 48 1.0549 0.2474 1.0549 1.0271
No log 1.1905 50 1.0400 0.1841 1.0400 1.0198
No log 1.2381 52 1.1018 0.1918 1.1018 1.0497
No log 1.2857 54 1.1109 0.1894 1.1109 1.0540
No log 1.3333 56 1.1166 0.2344 1.1166 1.0567
No log 1.3810 58 1.1786 0.2813 1.1786 1.0856
No log 1.4286 60 1.2663 0.2750 1.2663 1.1253
No log 1.4762 62 1.2919 0.2750 1.2919 1.1366
No log 1.5238 64 1.2247 0.2690 1.2247 1.1067
No log 1.5714 66 1.1653 0.2316 1.1653 1.0795
No log 1.6190 68 1.1193 0.2931 1.1193 1.0580
No log 1.6667 70 1.1859 0.3027 1.1859 1.0890
No log 1.7143 72 1.1692 0.2651 1.1692 1.0813
No log 1.7619 74 1.1275 0.2556 1.1275 1.0618
No log 1.8095 76 1.0851 0.2670 1.0851 1.0417
No log 1.8571 78 1.0385 0.2951 1.0385 1.0191
No log 1.9048 80 1.0968 0.3394 1.0968 1.0473
No log 1.9524 82 1.1075 0.2864 1.1075 1.0524
No log 2.0 84 1.0140 0.3503 1.0140 1.0070
No log 2.0476 86 0.9095 0.3678 0.9095 0.9537
No log 2.0952 88 0.9150 0.3445 0.9150 0.9566
No log 2.1429 90 0.9573 0.3318 0.9573 0.9784
No log 2.1905 92 0.9473 0.3318 0.9473 0.9733
No log 2.2381 94 0.9666 0.3861 0.9666 0.9832
No log 2.2857 96 0.9750 0.3732 0.9750 0.9874
No log 2.3333 98 0.9602 0.3596 0.9602 0.9799
No log 2.3810 100 0.9346 0.3897 0.9346 0.9667
No log 2.4286 102 0.9708 0.4224 0.9708 0.9853
No log 2.4762 104 0.9500 0.3939 0.9500 0.9747
No log 2.5238 106 0.9464 0.3637 0.9464 0.9728
No log 2.5714 108 0.9141 0.3112 0.9141 0.9561
No log 2.6190 110 0.9219 0.3076 0.9219 0.9601
No log 2.6667 112 0.9136 0.4145 0.9136 0.9558
No log 2.7143 114 0.9416 0.4123 0.9416 0.9704
No log 2.7619 116 0.9776 0.4266 0.9776 0.9887
No log 2.8095 118 0.9876 0.4526 0.9876 0.9938
No log 2.8571 120 1.0194 0.4148 1.0194 1.0097
No log 2.9048 122 1.0495 0.4350 1.0495 1.0245
No log 2.9524 124 1.0355 0.4102 1.0355 1.0176
No log 3.0 126 0.9934 0.3765 0.9934 0.9967
No log 3.0476 128 1.0190 0.3453 1.0190 1.0095
No log 3.0952 130 0.9729 0.3330 0.9729 0.9864
No log 3.1429 132 0.9430 0.3782 0.9430 0.9711
No log 3.1905 134 0.9217 0.4180 0.9217 0.9601
No log 3.2381 136 0.9386 0.3465 0.9386 0.9688
No log 3.2857 138 0.9590 0.3725 0.9590 0.9793
No log 3.3333 140 1.0037 0.3763 1.0037 1.0018
No log 3.3810 142 1.0803 0.2797 1.0803 1.0394
No log 3.4286 144 1.0017 0.3363 1.0017 1.0009
No log 3.4762 146 0.9374 0.2718 0.9374 0.9682
No log 3.5238 148 0.9637 0.2103 0.9637 0.9817
No log 3.5714 150 0.9330 0.2351 0.9330 0.9659
No log 3.6190 152 0.9587 0.3658 0.9587 0.9791
No log 3.6667 154 1.0179 0.4071 1.0179 1.0089
No log 3.7143 156 1.0385 0.4071 1.0385 1.0191
No log 3.7619 158 0.9983 0.3842 0.9983 0.9991
No log 3.8095 160 0.9327 0.3914 0.9327 0.9658
No log 3.8571 162 0.9257 0.4690 0.9257 0.9621
No log 3.9048 164 0.9268 0.4045 0.9268 0.9627
No log 3.9524 166 0.9830 0.4168 0.9830 0.9915
No log 4.0 168 1.0850 0.3845 1.0850 1.0417
No log 4.0476 170 1.2600 0.3283 1.2600 1.1225
No log 4.0952 172 1.2706 0.3361 1.2706 1.1272
No log 4.1429 174 1.0844 0.4135 1.0844 1.0413
No log 4.1905 176 0.9927 0.3974 0.9927 0.9963
No log 4.2381 178 1.0159 0.3924 1.0159 1.0079
No log 4.2857 180 1.0318 0.3807 1.0318 1.0158
No log 4.3333 182 1.0596 0.3961 1.0596 1.0294
No log 4.3810 184 1.0559 0.4565 1.0559 1.0276
No log 4.4286 186 1.0278 0.4328 1.0278 1.0138
No log 4.4762 188 0.9943 0.4344 0.9943 0.9971
No log 4.5238 190 0.9654 0.3744 0.9654 0.9826
No log 4.5714 192 0.9612 0.3151 0.9612 0.9804
No log 4.6190 194 0.9659 0.3445 0.9659 0.9828
No log 4.6667 196 0.9983 0.3921 0.9983 0.9992
No log 4.7143 198 1.0194 0.3921 1.0194 1.0097
No log 4.7619 200 0.9818 0.3959 0.9818 0.9909
No log 4.8095 202 0.9851 0.3074 0.9851 0.9925
No log 4.8571 204 0.9865 0.4407 0.9865 0.9932
No log 4.9048 206 0.9967 0.4466 0.9967 0.9983
No log 4.9524 208 1.0512 0.3773 1.0512 1.0253
No log 5.0 210 1.1076 0.3921 1.1076 1.0524
No log 5.0476 212 1.0211 0.4310 1.0211 1.0105
No log 5.0952 214 1.0215 0.3590 1.0215 1.0107
No log 5.1429 216 1.1108 0.2559 1.1108 1.0539
No log 5.1905 218 1.0742 0.2748 1.0742 1.0365
No log 5.2381 220 0.9555 0.2988 0.9555 0.9775
No log 5.2857 222 0.9349 0.3236 0.9349 0.9669
No log 5.3333 224 0.9535 0.3891 0.9535 0.9765
No log 5.3810 226 0.9306 0.3622 0.9306 0.9647
No log 5.4286 228 0.9830 0.4597 0.9830 0.9915
No log 5.4762 230 1.1919 0.2667 1.1919 1.0917
No log 5.5238 232 1.2717 0.2767 1.2717 1.1277
No log 5.5714 234 1.1323 0.3009 1.1323 1.0641
No log 5.6190 236 0.9976 0.3264 0.9976 0.9988
No log 5.6667 238 0.9599 0.3652 0.9599 0.9798
No log 5.7143 240 0.9735 0.3637 0.9735 0.9866
No log 5.7619 242 1.0565 0.3503 1.0565 1.0279
No log 5.8095 244 1.0466 0.3806 1.0466 1.0230
No log 5.8571 246 0.9643 0.4491 0.9643 0.9820
No log 5.9048 248 0.9314 0.4175 0.9314 0.9651
No log 5.9524 250 0.9317 0.4030 0.9317 0.9653
No log 6.0 252 0.9435 0.4123 0.9435 0.9713
No log 6.0476 254 0.9531 0.4123 0.9531 0.9763
No log 6.0952 256 0.9573 0.3323 0.9573 0.9784
No log 6.1429 258 1.0836 0.2601 1.0836 1.0409
No log 6.1905 260 1.1157 0.2601 1.1157 1.0563
No log 6.2381 262 1.0051 0.3210 1.0051 1.0026
No log 6.2857 264 1.0180 0.4220 1.0180 1.0090
No log 6.3333 266 1.0484 0.3973 1.0484 1.0239
No log 6.3810 268 1.0023 0.3647 1.0023 1.0012
No log 6.4286 270 1.0089 0.3666 1.0089 1.0045
No log 6.4762 272 0.9851 0.3896 0.9851 0.9925
No log 6.5238 274 1.0234 0.3844 1.0234 1.0116
No log 6.5714 276 1.0758 0.3243 1.0758 1.0372
No log 6.6190 278 1.0653 0.3695 1.0653 1.0322
No log 6.6667 280 0.9566 0.3424 0.9566 0.9781
No log 6.7143 282 0.9435 0.3424 0.9435 0.9713
No log 6.7619 284 0.9740 0.3760 0.9740 0.9869
No log 6.8095 286 1.0011 0.3351 1.0011 1.0006
No log 6.8571 288 1.0648 0.1900 1.0648 1.0319
No log 6.9048 290 1.0856 0.1474 1.0856 1.0419
No log 6.9524 292 1.0520 0.2449 1.0520 1.0257
No log 7.0 294 1.0232 0.3207 1.0232 1.0115
No log 7.0476 296 1.0208 0.3207 1.0208 1.0104
No log 7.0952 298 1.0888 0.2678 1.0888 1.0435
No log 7.1429 300 1.0869 0.3374 1.0869 1.0425
No log 7.1905 302 1.0380 0.3643 1.0380 1.0188
No log 7.2381 304 0.9816 0.3563 0.9816 0.9908
No log 7.2857 306 0.9720 0.3485 0.9720 0.9859
No log 7.3333 308 0.9767 0.3915 0.9767 0.9883
No log 7.3810 310 0.9809 0.3052 0.9809 0.9904
No log 7.4286 312 1.0082 0.2942 1.0082 1.0041
No log 7.4762 314 1.0205 0.3044 1.0205 1.0102
No log 7.5238 316 1.0240 0.3383 1.0240 1.0119
No log 7.5714 318 1.0067 0.3543 1.0067 1.0033
No log 7.6190 320 1.0036 0.3914 1.0036 1.0018
No log 7.6667 322 1.0085 0.3682 1.0085 1.0042
No log 7.7143 324 1.0073 0.3543 1.0073 1.0036
No log 7.7619 326 1.0349 0.3278 1.0349 1.0173
No log 7.8095 328 1.0516 0.2796 1.0516 1.0255
No log 7.8571 330 1.0563 0.2796 1.0563 1.0278
No log 7.9048 332 0.9772 0.2623 0.9772 0.9885
No log 7.9524 334 0.9302 0.2967 0.9302 0.9645
No log 8.0 336 0.9228 0.3117 0.9228 0.9606
No log 8.0476 338 0.9366 0.2942 0.9366 0.9678
No log 8.0952 340 0.9569 0.2499 0.9569 0.9782
No log 8.1429 342 0.9532 0.2647 0.9532 0.9763
No log 8.1905 344 0.9754 0.2623 0.9754 0.9876
No log 8.2381 346 0.9663 0.2941 0.9663 0.9830
No log 8.2857 348 0.9013 0.3658 0.9013 0.9494
No log 8.3333 350 0.8753 0.4345 0.8753 0.9356
No log 8.3810 352 0.8761 0.4345 0.8761 0.9360
No log 8.4286 354 0.9401 0.3921 0.9401 0.9696
No log 8.4762 356 1.1038 0.3614 1.1038 1.0506
No log 8.5238 358 1.1676 0.3452 1.1676 1.0806
No log 8.5714 360 1.0865 0.3405 1.0865 1.0423
No log 8.6190 362 0.9659 0.2941 0.9659 0.9828
No log 8.6667 364 0.9081 0.3494 0.9081 0.9530
No log 8.7143 366 0.8976 0.3637 0.8976 0.9474
No log 8.7619 368 0.8937 0.3682 0.8937 0.9453
No log 8.8095 370 0.9439 0.4476 0.9439 0.9716
No log 8.8571 372 1.0016 0.4815 1.0016 1.0008
No log 8.9048 374 0.9714 0.4709 0.9714 0.9856
No log 8.9524 376 0.9462 0.4606 0.9462 0.9727
No log 9.0 378 0.8894 0.4423 0.8894 0.9431
No log 9.0476 380 0.8990 0.4955 0.8990 0.9481
No log 9.0952 382 0.8710 0.4676 0.8710 0.9333
No log 9.1429 384 0.8621 0.4269 0.8621 0.9285
No log 9.1905 386 0.8604 0.4133 0.8604 0.9276
No log 9.2381 388 0.8596 0.3878 0.8596 0.9272
No log 9.2857 390 0.8752 0.3583 0.8752 0.9355
No log 9.3333 392 0.8662 0.3721 0.8662 0.9307
No log 9.3810 394 0.8460 0.4540 0.8460 0.9198
No log 9.4286 396 0.8439 0.4540 0.8439 0.9186
No log 9.4762 398 0.8364 0.4151 0.8364 0.9146
No log 9.5238 400 0.8677 0.3293 0.8677 0.9315
No log 9.5714 402 0.9277 0.3044 0.9277 0.9632
No log 9.6190 404 0.9469 0.3044 0.9469 0.9731
No log 9.6667 406 0.8989 0.3485 0.8989 0.9481
No log 9.7143 408 0.8831 0.4378 0.8831 0.9398
No log 9.7619 410 0.8843 0.3666 0.8843 0.9404
No log 9.8095 412 0.9087 0.4594 0.9087 0.9533
No log 9.8571 414 0.9016 0.3966 0.9016 0.9495
No log 9.9048 416 0.8968 0.3949 0.8968 0.9470
No log 9.9524 418 0.8924 0.4407 0.8924 0.9447
No log 10.0 420 0.8884 0.4629 0.8884 0.9426
No log 10.0476 422 0.8772 0.3717 0.8772 0.9366
No log 10.0952 424 0.8851 0.3476 0.8851 0.9408
No log 10.1429 426 0.9006 0.3517 0.9006 0.9490
No log 10.1905 428 0.9134 0.2834 0.9134 0.9557
No log 10.2381 430 0.9004 0.3517 0.9004 0.9489
No log 10.2857 432 0.9112 0.4981 0.9112 0.9545
No log 10.3333 434 0.9648 0.4350 0.9648 0.9822
No log 10.3810 436 0.9669 0.4250 0.9669 0.9833
No log 10.4286 438 0.9202 0.4854 0.9202 0.9593
No log 10.4762 440 0.9075 0.4158 0.9075 0.9526
No log 10.5238 442 0.9031 0.4371 0.9031 0.9503
No log 10.5714 444 0.9040 0.4613 0.9040 0.9508
No log 10.6190 446 0.9143 0.4198 0.9143 0.9562
No log 10.6667 448 0.9341 0.4192 0.9341 0.9665
No log 10.7143 450 0.9148 0.4327 0.9148 0.9564
No log 10.7619 452 0.8787 0.4353 0.8787 0.9374
No log 10.8095 454 0.8735 0.4030 0.8735 0.9346
No log 10.8571 456 0.8748 0.4030 0.8748 0.9353
No log 10.9048 458 0.8864 0.4398 0.8864 0.9415
No log 10.9524 460 0.9059 0.4266 0.9059 0.9518
No log 11.0 462 0.9309 0.4581 0.9309 0.9648
No log 11.0476 464 0.8948 0.4595 0.8948 0.9460
No log 11.0952 466 0.8865 0.3673 0.8865 0.9415
No log 11.1429 468 0.8947 0.3280 0.8947 0.9459
No log 11.1905 470 0.9152 0.3528 0.9152 0.9567
No log 11.2381 472 0.9406 0.3214 0.9406 0.9699
No log 11.2857 474 0.9475 0.3089 0.9475 0.9734
No log 11.3333 476 0.9432 0.3194 0.9432 0.9712
No log 11.3810 478 0.9457 0.3271 0.9457 0.9725
No log 11.4286 480 0.9587 0.3874 0.9587 0.9791
No log 11.4762 482 0.9731 0.3715 0.9731 0.9865
No log 11.5238 484 1.0719 0.3929 1.0719 1.0353
No log 11.5714 486 1.1212 0.375 1.1212 1.0588
No log 11.6190 488 1.1212 0.3617 1.1212 1.0589
No log 11.6667 490 1.1092 0.3532 1.1092 1.0532
No log 11.7143 492 1.0344 0.3862 1.0344 1.0170
No log 11.7619 494 0.9447 0.3859 0.9447 0.9720
No log 11.8095 496 0.9239 0.3896 0.9239 0.9612
No log 11.8571 498 0.9302 0.3631 0.9302 0.9645
0.2661 11.9048 500 0.9176 0.3915 0.9176 0.9579
0.2661 11.9524 502 0.9150 0.3996 0.9150 0.9566
0.2661 12.0 504 0.9288 0.4391 0.9288 0.9637
0.2661 12.0476 506 0.9217 0.4515 0.9217 0.9601
0.2661 12.0952 508 0.9152 0.4676 0.9152 0.9567
0.2661 12.1429 510 0.9040 0.4313 0.9040 0.9508
0.2661 12.1905 512 0.8940 0.4527 0.8940 0.9455
0.2661 12.2381 514 0.8951 0.4878 0.8951 0.9461
0.2661 12.2857 516 0.8806 0.4269 0.8806 0.9384
0.2661 12.3333 518 0.8752 0.4051 0.8752 0.9355
0.2661 12.3810 520 0.8749 0.4051 0.8749 0.9354
0.2661 12.4286 522 0.9176 0.4164 0.9176 0.9579
0.2661 12.4762 524 0.9539 0.4112 0.9539 0.9767
0.2661 12.5238 526 0.8993 0.4763 0.8993 0.9483
0.2661 12.5714 528 0.8542 0.4014 0.8542 0.9243
0.2661 12.6190 530 0.8581 0.4391 0.8581 0.9263
0.2661 12.6667 532 0.8702 0.4499 0.8702 0.9328
0.2661 12.7143 534 0.8693 0.3941 0.8693 0.9324
0.2661 12.7619 536 0.8634 0.3258 0.8634 0.9292
0.2661 12.8095 538 0.8705 0.3335 0.8705 0.9330
0.2661 12.8571 540 0.8896 0.3636 0.8896 0.9432
0.2661 12.9048 542 0.8905 0.3915 0.8905 0.9437
0.2661 12.9524 544 0.8787 0.3622 0.8787 0.9374

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k9_task5_organization

Finetuned
(4019)
this model