ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k2_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9477
  • Qwk: 0.5031
  • Mse: 0.9477
  • Rmse: 0.9735

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.2 2 4.6258 0.0042 4.6258 2.1508
No log 0.4 4 2.7940 -0.0640 2.7940 1.6715
No log 0.6 6 1.9046 0.0198 1.9046 1.3801
No log 0.8 8 1.6346 0.0372 1.6346 1.2785
No log 1.0 10 1.3686 0.0872 1.3686 1.1699
No log 1.2 12 1.2701 0.0061 1.2701 1.1270
No log 1.4 14 1.2662 -0.0071 1.2662 1.1252
No log 1.6 16 1.3709 0.0169 1.3709 1.1709
No log 1.8 18 1.4336 -0.0149 1.4336 1.1973
No log 2.0 20 1.3823 0.0019 1.3823 1.1757
No log 2.2 22 1.4177 0.0019 1.4177 1.1907
No log 2.4 24 1.4604 -0.0359 1.4604 1.2085
No log 2.6 26 1.3579 0.0992 1.3579 1.1653
No log 2.8 28 1.2265 0.1959 1.2265 1.1075
No log 3.0 30 1.3550 0.1866 1.3550 1.1640
No log 3.2 32 1.2271 0.2069 1.2271 1.1078
No log 3.4 34 1.0439 0.3038 1.0439 1.0217
No log 3.6 36 1.0852 0.2537 1.0852 1.0417
No log 3.8 38 1.2748 0.1364 1.2748 1.1291
No log 4.0 40 1.7207 0.1705 1.7207 1.3117
No log 4.2 42 1.9614 0.1355 1.9614 1.4005
No log 4.4 44 2.0290 0.1516 2.0290 1.4244
No log 4.6 46 1.8442 0.1006 1.8442 1.3580
No log 4.8 48 1.6006 0.1120 1.6006 1.2651
No log 5.0 50 1.2716 0.2605 1.2716 1.1277
No log 5.2 52 1.1127 0.3025 1.1127 1.0548
No log 5.4 54 0.9571 0.3381 0.9571 0.9783
No log 5.6 56 0.8813 0.3493 0.8813 0.9388
No log 5.8 58 0.8959 0.3021 0.8959 0.9465
No log 6.0 60 0.9404 0.3857 0.9404 0.9697
No log 6.2 62 0.8893 0.2676 0.8893 0.9430
No log 6.4 64 1.0420 0.4100 1.0420 1.0208
No log 6.6 66 1.7140 0.2472 1.7140 1.3092
No log 6.8 68 1.8652 0.2358 1.8652 1.3657
No log 7.0 70 1.7182 0.2472 1.7182 1.3108
No log 7.2 72 1.3248 0.2880 1.3248 1.1510
No log 7.4 74 1.0329 0.3747 1.0329 1.0163
No log 7.6 76 0.9467 0.4690 0.9467 0.9730
No log 7.8 78 1.0603 0.4205 1.0603 1.0297
No log 8.0 80 1.3520 0.2692 1.3520 1.1628
No log 8.2 82 1.3087 0.2568 1.3087 1.1440
No log 8.4 84 1.0012 0.4048 1.0012 1.0006
No log 8.6 86 0.8649 0.4086 0.8649 0.9300
No log 8.8 88 0.8778 0.4560 0.8778 0.9369
No log 9.0 90 0.8747 0.4760 0.8747 0.9353
No log 9.2 92 1.1099 0.3192 1.1099 1.0535
No log 9.4 94 1.6767 0.2299 1.6767 1.2949
No log 9.6 96 1.7436 0.2659 1.7436 1.3205
No log 9.8 98 1.3308 0.2808 1.3308 1.1536
No log 10.0 100 0.9608 0.4849 0.9608 0.9802
No log 10.2 102 0.9350 0.3558 0.9350 0.9670
No log 10.4 104 0.9316 0.3660 0.9316 0.9652
No log 10.6 106 0.8686 0.3513 0.8686 0.9320
No log 10.8 108 0.9048 0.5219 0.9048 0.9512
No log 11.0 110 1.0996 0.3952 1.0996 1.0486
No log 11.2 112 1.1355 0.3952 1.1355 1.0656
No log 11.4 114 0.9331 0.5311 0.9331 0.9660
No log 11.6 116 0.8640 0.5327 0.8640 0.9295
No log 11.8 118 0.8374 0.4933 0.8374 0.9151
No log 12.0 120 0.8388 0.4885 0.8388 0.9159
No log 12.2 122 0.9412 0.5094 0.9412 0.9702
No log 12.4 124 1.0976 0.4186 1.0976 1.0477
No log 12.6 126 1.3581 0.3636 1.3581 1.1654
No log 12.8 128 1.2978 0.4076 1.2978 1.1392
No log 13.0 130 1.0331 0.4672 1.0331 1.0164
No log 13.2 132 0.9695 0.4509 0.9695 0.9846
No log 13.4 134 0.9126 0.4641 0.9126 0.9553
No log 13.6 136 0.8660 0.4666 0.8660 0.9306
No log 13.8 138 0.9122 0.5203 0.9122 0.9551
No log 14.0 140 1.0196 0.4634 1.0196 1.0098
No log 14.2 142 1.1226 0.4410 1.1226 1.0595
No log 14.4 144 1.3104 0.3443 1.3104 1.1447
No log 14.6 146 1.2637 0.2926 1.2637 1.1242
No log 14.8 148 1.0272 0.4990 1.0272 1.0135
No log 15.0 150 0.8296 0.4834 0.8296 0.9108
No log 15.2 152 0.8287 0.4045 0.8287 0.9103
No log 15.4 154 0.8191 0.4045 0.8191 0.9051
No log 15.6 156 0.8346 0.5686 0.8346 0.9135
No log 15.8 158 1.0071 0.4990 1.0071 1.0035
No log 16.0 160 1.0303 0.4328 1.0303 1.0150
No log 16.2 162 0.9331 0.4807 0.9331 0.9660
No log 16.4 164 0.9100 0.4807 0.9100 0.9539
No log 16.6 166 0.8454 0.5295 0.8454 0.9194
No log 16.8 168 0.8005 0.5238 0.8005 0.8947
No log 17.0 170 0.7732 0.4460 0.7732 0.8793
No log 17.2 172 0.7781 0.4494 0.7781 0.8821
No log 17.4 174 0.9145 0.5394 0.9145 0.9563
No log 17.6 176 1.2115 0.4485 1.2115 1.1007
No log 17.8 178 1.2881 0.3824 1.2881 1.1349
No log 18.0 180 1.1810 0.4618 1.1810 1.0867
No log 18.2 182 0.8827 0.4961 0.8827 0.9395
No log 18.4 184 0.8180 0.4885 0.8180 0.9044
No log 18.6 186 0.8682 0.4774 0.8682 0.9318
No log 18.8 188 0.9183 0.4943 0.9183 0.9583
No log 19.0 190 0.9025 0.5037 0.9025 0.9500
No log 19.2 192 0.8667 0.5037 0.8667 0.9310
No log 19.4 194 0.8643 0.5115 0.8643 0.9297
No log 19.6 196 0.8544 0.4815 0.8544 0.9243
No log 19.8 198 0.8622 0.4045 0.8622 0.9285
No log 20.0 200 0.8494 0.4945 0.8494 0.9216
No log 20.2 202 0.8940 0.5238 0.8940 0.9455
No log 20.4 204 0.9761 0.3660 0.9761 0.9880
No log 20.6 206 0.9530 0.4353 0.9530 0.9762
No log 20.8 208 0.8938 0.4542 0.8938 0.9454
No log 21.0 210 0.9065 0.4275 0.9065 0.9521
No log 21.2 212 0.9301 0.4347 0.9301 0.9644
No log 21.4 214 0.9609 0.4595 0.9609 0.9803
No log 21.6 216 1.0839 0.4358 1.0839 1.0411
No log 21.8 218 1.0715 0.4596 1.0715 1.0351
No log 22.0 220 1.0187 0.5045 1.0187 1.0093
No log 22.2 222 0.9453 0.5555 0.9453 0.9722
No log 22.4 224 0.8696 0.5821 0.8696 0.9325
No log 22.6 226 0.8486 0.5912 0.8486 0.9212
No log 22.8 228 0.8606 0.5912 0.8606 0.9277
No log 23.0 230 0.8132 0.5746 0.8132 0.9018
No log 23.2 232 0.8103 0.5721 0.8103 0.9002
No log 23.4 234 0.8275 0.6094 0.8275 0.9097
No log 23.6 236 0.7996 0.5632 0.7996 0.8942
No log 23.8 238 0.8026 0.4644 0.8026 0.8959
No log 24.0 240 0.8077 0.4996 0.8077 0.8987
No log 24.2 242 0.8258 0.4615 0.8258 0.9087
No log 24.4 244 0.9089 0.5267 0.9089 0.9534
No log 24.6 246 0.9440 0.4898 0.9440 0.9716
No log 24.8 248 0.9142 0.5267 0.9142 0.9561
No log 25.0 250 0.8903 0.3908 0.8903 0.9435
No log 25.2 252 0.9190 0.3639 0.9190 0.9587
No log 25.4 254 0.8897 0.4278 0.8897 0.9432
No log 25.6 256 0.8475 0.4455 0.8475 0.9206
No log 25.8 258 0.8299 0.5455 0.8299 0.9110
No log 26.0 260 0.8991 0.5081 0.8991 0.9482
No log 26.2 262 0.9224 0.4931 0.9224 0.9604
No log 26.4 264 0.8360 0.5482 0.8360 0.9143
No log 26.6 266 0.8140 0.5120 0.8140 0.9022
No log 26.8 268 0.8707 0.3879 0.8707 0.9331
No log 27.0 270 0.8980 0.2967 0.8980 0.9476
No log 27.2 272 0.8854 0.3483 0.8854 0.9410
No log 27.4 274 0.8741 0.4476 0.8741 0.9349
No log 27.6 276 0.9212 0.4898 0.9212 0.9598
No log 27.8 278 0.9732 0.4505 0.9732 0.9865
No log 28.0 280 0.9922 0.4471 0.9922 0.9961
No log 28.2 282 0.9018 0.4807 0.9018 0.9496
No log 28.4 284 0.8018 0.4991 0.8018 0.8954
No log 28.6 286 0.7977 0.5505 0.7977 0.8931
No log 28.8 288 0.8417 0.4792 0.8417 0.9174
No log 29.0 290 0.8783 0.4096 0.8783 0.9372
No log 29.2 292 0.9357 0.4480 0.9357 0.9673
No log 29.4 294 0.9716 0.4717 0.9716 0.9857
No log 29.6 296 1.0374 0.4358 1.0374 1.0185
No log 29.8 298 1.1311 0.4414 1.1311 1.0635
No log 30.0 300 1.1381 0.4587 1.1381 1.0668
No log 30.2 302 1.0563 0.4358 1.0563 1.0278
No log 30.4 304 0.8845 0.5167 0.8845 0.9405
No log 30.6 306 0.8057 0.5102 0.8057 0.8976
No log 30.8 308 0.8035 0.5009 0.8035 0.8964
No log 31.0 310 0.8557 0.5409 0.8557 0.9250
No log 31.2 312 1.0449 0.4777 1.0449 1.0222
No log 31.4 314 1.1651 0.4502 1.1651 1.0794
No log 31.6 316 1.0633 0.4587 1.0633 1.0312
No log 31.8 318 0.9107 0.4685 0.9107 0.9543
No log 32.0 320 0.8210 0.4792 0.8210 0.9061
No log 32.2 322 0.8467 0.3852 0.8467 0.9201
No log 32.4 324 0.8564 0.3845 0.8564 0.9254
No log 32.6 326 0.8477 0.3920 0.8477 0.9207
No log 32.8 328 0.8715 0.5023 0.8715 0.9335
No log 33.0 330 0.8996 0.4723 0.8996 0.9485
No log 33.2 332 0.9035 0.4855 0.9035 0.9505
No log 33.4 334 0.8708 0.5127 0.8708 0.9332
No log 33.6 336 0.8418 0.4676 0.8418 0.9175
No log 33.8 338 0.8305 0.4316 0.8305 0.9113
No log 34.0 340 0.8253 0.5094 0.8253 0.9085
No log 34.2 342 0.8166 0.5186 0.8166 0.9036
No log 34.4 344 0.8343 0.5376 0.8343 0.9134
No log 34.6 346 0.9160 0.4689 0.9160 0.9571
No log 34.8 348 0.9806 0.4205 0.9806 0.9903
No log 35.0 350 0.9813 0.4205 0.9813 0.9906
No log 35.2 352 0.9632 0.4206 0.9632 0.9814
No log 35.4 354 0.8914 0.4898 0.8914 0.9441
No log 35.6 356 0.8481 0.4745 0.8481 0.9209
No log 35.8 358 0.8214 0.4975 0.8214 0.9063
No log 36.0 360 0.8120 0.4749 0.8120 0.9011
No log 36.2 362 0.8064 0.4514 0.8064 0.8980
No log 36.4 364 0.8084 0.4803 0.8084 0.8991
No log 36.6 366 0.8101 0.5376 0.8101 0.9000
No log 36.8 368 0.8376 0.4975 0.8376 0.9152
No log 37.0 370 0.9298 0.4650 0.9298 0.9642
No log 37.2 372 0.9542 0.4650 0.9542 0.9768
No log 37.4 374 0.9194 0.4650 0.9194 0.9588
No log 37.6 376 0.9119 0.5018 0.9119 0.9549
No log 37.8 378 0.9055 0.5102 0.9055 0.9516
No log 38.0 380 0.8883 0.5177 0.8883 0.9425
No log 38.2 382 0.8796 0.5050 0.8796 0.9379
No log 38.4 384 0.8821 0.4657 0.8821 0.9392
No log 38.6 386 0.8806 0.4696 0.8806 0.9384
No log 38.8 388 0.8769 0.4696 0.8769 0.9365
No log 39.0 390 0.8633 0.4991 0.8633 0.9292
No log 39.2 392 0.8645 0.5345 0.8645 0.9298
No log 39.4 394 0.8576 0.5345 0.8576 0.9261
No log 39.6 396 0.8570 0.5554 0.8570 0.9258
No log 39.8 398 0.8877 0.5264 0.8877 0.9422
No log 40.0 400 0.8795 0.4898 0.8795 0.9378
No log 40.2 402 0.8771 0.5203 0.8771 0.9365
No log 40.4 404 0.9001 0.5122 0.9001 0.9487
No log 40.6 406 0.8807 0.5203 0.8807 0.9385
No log 40.8 408 0.8484 0.5509 0.8484 0.9211
No log 41.0 410 0.7995 0.5603 0.7995 0.8942
No log 41.2 412 0.7437 0.5988 0.7437 0.8624
No log 41.4 414 0.7437 0.5510 0.7437 0.8624
No log 41.6 416 0.7805 0.5375 0.7805 0.8835
No log 41.8 418 0.8026 0.5345 0.8026 0.8959
No log 42.0 420 0.7935 0.5404 0.7935 0.8908
No log 42.2 422 0.7933 0.5470 0.7933 0.8907
No log 42.4 424 0.7940 0.5500 0.7940 0.8911
No log 42.6 426 0.7986 0.5409 0.7986 0.8936
No log 42.8 428 0.7986 0.5291 0.7986 0.8937
No log 43.0 430 0.7987 0.5592 0.7987 0.8937
No log 43.2 432 0.8144 0.5500 0.8144 0.9025
No log 43.4 434 0.8299 0.5200 0.8299 0.9110
No log 43.6 436 0.8366 0.5190 0.8366 0.9147
No log 43.8 438 0.8243 0.5781 0.8243 0.9079
No log 44.0 440 0.8126 0.5781 0.8126 0.9014
No log 44.2 442 0.8008 0.5686 0.8008 0.8949
No log 44.4 444 0.8027 0.5592 0.8027 0.8960
No log 44.6 446 0.8010 0.5802 0.8010 0.8950
No log 44.8 448 0.8012 0.5802 0.8012 0.8951
No log 45.0 450 0.7997 0.5710 0.7997 0.8943
No log 45.2 452 0.7865 0.5684 0.7865 0.8868
No log 45.4 454 0.7848 0.5877 0.7848 0.8859
No log 45.6 456 0.8051 0.5877 0.8051 0.8973
No log 45.8 458 0.8452 0.4652 0.8452 0.9193
No log 46.0 460 0.8802 0.4706 0.8802 0.9382
No log 46.2 462 0.8968 0.4893 0.8968 0.9470
No log 46.4 464 0.8773 0.5007 0.8773 0.9366
No log 46.6 466 0.8265 0.5376 0.8265 0.9091
No log 46.8 468 0.7931 0.5852 0.7931 0.8906
No log 47.0 470 0.7759 0.5580 0.7759 0.8808
No log 47.2 472 0.7657 0.5580 0.7657 0.8750
No log 47.4 474 0.7495 0.5932 0.7495 0.8657
No log 47.6 476 0.7728 0.5734 0.7728 0.8791
No log 47.8 478 0.8018 0.5643 0.8018 0.8954
No log 48.0 480 0.8093 0.5436 0.8093 0.8996
No log 48.2 482 0.8009 0.5318 0.8009 0.8949
No log 48.4 484 0.7832 0.5686 0.7832 0.8850
No log 48.6 486 0.7747 0.5686 0.7747 0.8802
No log 48.8 488 0.7806 0.5592 0.7806 0.8835
No log 49.0 490 0.8157 0.5485 0.8157 0.9032
No log 49.2 492 0.8809 0.5111 0.8809 0.9385
No log 49.4 494 0.9128 0.5330 0.9128 0.9554
No log 49.6 496 0.8837 0.5111 0.8837 0.9400
No log 49.8 498 0.8361 0.5203 0.8361 0.9144
0.2733 50.0 500 0.8128 0.5394 0.8128 0.9016
0.2733 50.2 502 0.8110 0.5554 0.8110 0.9005
0.2733 50.4 504 0.8011 0.5554 0.8011 0.8950
0.2733 50.6 506 0.8105 0.5578 0.8105 0.9003
0.2733 50.8 508 0.8166 0.5394 0.8166 0.9037
0.2733 51.0 510 0.8210 0.5081 0.8210 0.9061
0.2733 51.2 512 0.8432 0.5203 0.8432 0.9183
0.2733 51.4 514 0.8299 0.5394 0.8299 0.9110
0.2733 51.6 516 0.8228 0.5578 0.8228 0.9071
0.2733 51.8 518 0.8458 0.5303 0.8458 0.9197
0.2733 52.0 520 0.8490 0.5303 0.8490 0.9214
0.2733 52.2 522 0.8395 0.5394 0.8395 0.9162
0.2733 52.4 524 0.8518 0.5394 0.8518 0.9229
0.2733 52.6 526 0.8479 0.5578 0.8479 0.9208
0.2733 52.8 528 0.8640 0.5488 0.8640 0.9295
0.2733 53.0 530 0.8904 0.4890 0.8904 0.9436
0.2733 53.2 532 0.9338 0.5031 0.9338 0.9663
0.2733 53.4 534 1.0035 0.5247 1.0035 1.0018
0.2733 53.6 536 1.0028 0.5247 1.0028 1.0014
0.2733 53.8 538 0.9477 0.5031 0.9477 0.9735

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k2_task2_organization

Finetuned
(4019)
this model