ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k3_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9242
  • Qwk: 0.6462
  • Mse: 0.9242
  • Rmse: 0.9613

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1667 2 6.8104 0.0176 6.8104 2.6097
No log 0.3333 4 4.4213 0.0741 4.4213 2.1027
No log 0.5 6 3.4042 0.0 3.4042 1.8450
No log 0.6667 8 2.3504 0.1429 2.3504 1.5331
No log 0.8333 10 1.9124 0.1062 1.9124 1.3829
No log 1.0 12 1.6230 0.1905 1.6230 1.2740
No log 1.1667 14 1.5751 0.1682 1.5751 1.2550
No log 1.3333 16 1.5758 0.1835 1.5758 1.2553
No log 1.5 18 1.7985 0.2833 1.7985 1.3411
No log 1.6667 20 1.9370 0.2581 1.9370 1.3918
No log 1.8333 22 2.0341 0.2137 2.0341 1.4262
No log 2.0 24 1.9211 0.3008 1.9211 1.3860
No log 2.1667 26 1.6428 0.3651 1.6428 1.2817
No log 2.3333 28 1.3809 0.4167 1.3809 1.1751
No log 2.5 30 1.2852 0.3761 1.2852 1.1337
No log 2.6667 32 1.3114 0.4590 1.3114 1.1452
No log 2.8333 34 1.4247 0.4062 1.4247 1.1936
No log 3.0 36 1.5395 0.4580 1.5395 1.2408
No log 3.1667 38 1.8845 0.2899 1.8845 1.3728
No log 3.3333 40 1.6028 0.3438 1.6028 1.2660
No log 3.5 42 1.3941 0.3740 1.3941 1.1807
No log 3.6667 44 1.2312 0.496 1.2312 1.1096
No log 3.8333 46 1.1838 0.4754 1.1838 1.0880
No log 4.0 48 1.1842 0.4553 1.1842 1.0882
No log 4.1667 50 1.1734 0.512 1.1734 1.0832
No log 4.3333 52 1.2306 0.4724 1.2306 1.1093
No log 4.5 54 1.3776 0.3710 1.3776 1.1737
No log 4.6667 56 1.6495 0.3382 1.6495 1.2843
No log 4.8333 58 1.9388 0.3776 1.9388 1.3924
No log 5.0 60 1.7965 0.3830 1.7965 1.3404
No log 5.1667 62 1.8081 0.3830 1.8081 1.3447
No log 5.3333 64 1.6131 0.3824 1.6131 1.2701
No log 5.5 66 1.4887 0.3817 1.4887 1.2201
No log 5.6667 68 1.6898 0.3636 1.6898 1.2999
No log 5.8333 70 1.3501 0.4844 1.3501 1.1619
No log 6.0 72 1.2102 0.4480 1.2102 1.1001
No log 6.1667 74 1.1772 0.4918 1.1772 1.0850
No log 6.3333 76 1.1674 0.4706 1.1674 1.0805
No log 6.5 78 1.1770 0.5161 1.1770 1.0849
No log 6.6667 80 1.2386 0.4724 1.2386 1.1129
No log 6.8333 82 1.4208 0.3937 1.4208 1.1920
No log 7.0 84 1.4325 0.3622 1.4325 1.1969
No log 7.1667 86 1.2652 0.5077 1.2652 1.1248
No log 7.3333 88 1.1930 0.5606 1.1930 1.0922
No log 7.5 90 1.1980 0.5116 1.1980 1.0945
No log 7.6667 92 1.1534 0.4961 1.1534 1.0740
No log 7.8333 94 1.1154 0.5625 1.1154 1.0561
No log 8.0 96 1.3753 0.4127 1.3753 1.1727
No log 8.1667 98 1.2946 0.4580 1.2946 1.1378
No log 8.3333 100 1.1133 0.5693 1.1133 1.0551
No log 8.5 102 1.0247 0.5926 1.0247 1.0123
No log 8.6667 104 1.0190 0.5985 1.0190 1.0095
No log 8.8333 106 1.1829 0.5571 1.1829 1.0876
No log 9.0 108 1.3329 0.4932 1.3329 1.1545
No log 9.1667 110 1.4278 0.4667 1.4278 1.1949
No log 9.3333 112 1.3958 0.4898 1.3958 1.1814
No log 9.5 114 1.1120 0.5481 1.1120 1.0545
No log 9.6667 116 1.1018 0.5839 1.1018 1.0497
No log 9.8333 118 1.2246 0.4776 1.2246 1.1066
No log 10.0 120 1.1985 0.4741 1.1985 1.0948
No log 10.1667 122 1.1132 0.6154 1.1132 1.0551
No log 10.3333 124 1.1430 0.5839 1.1430 1.0691
No log 10.5 126 1.1586 0.5755 1.1586 1.0764
No log 10.6667 128 1.1490 0.5985 1.1490 1.0719
No log 10.8333 130 1.1910 0.5455 1.1910 1.0913
No log 11.0 132 1.1459 0.5231 1.1459 1.0705
No log 11.1667 134 1.0392 0.5522 1.0392 1.0194
No log 11.3333 136 0.9658 0.5985 0.9658 0.9828
No log 11.5 138 0.9408 0.6667 0.9408 0.9699
No log 11.6667 140 0.9375 0.6667 0.9375 0.9683
No log 11.8333 142 0.9516 0.6519 0.9516 0.9755
No log 12.0 144 0.9546 0.5970 0.9546 0.9771
No log 12.1667 146 1.0050 0.5821 1.0050 1.0025
No log 12.3333 148 1.0335 0.5821 1.0335 1.0166
No log 12.5 150 1.0458 0.5821 1.0458 1.0226
No log 12.6667 152 1.0516 0.5649 1.0516 1.0255
No log 12.8333 154 1.0206 0.5649 1.0206 1.0102
No log 13.0 156 0.9646 0.6154 0.9646 0.9822
No log 13.1667 158 0.9866 0.6260 0.9866 0.9933
No log 13.3333 160 1.0088 0.6047 1.0088 1.0044
No log 13.5 162 1.0157 0.6107 1.0157 1.0078
No log 13.6667 164 1.0577 0.5891 1.0577 1.0284
No log 13.8333 166 1.0422 0.6 1.0422 1.0209
No log 14.0 168 1.0269 0.6212 1.0269 1.0134
No log 14.1667 170 1.0284 0.6212 1.0284 1.0141
No log 14.3333 172 1.0525 0.5846 1.0525 1.0259
No log 14.5 174 1.0759 0.6 1.0759 1.0373
No log 14.6667 176 1.0911 0.5846 1.0911 1.0445
No log 14.8333 178 1.0836 0.5954 1.0836 1.0409
No log 15.0 180 1.0998 0.5649 1.0998 1.0487
No log 15.1667 182 1.1707 0.5263 1.1707 1.0820
No log 15.3333 184 1.2068 0.4962 1.2068 1.0985
No log 15.5 186 1.2712 0.5152 1.2712 1.1275
No log 15.6667 188 1.3771 0.5 1.3771 1.1735
No log 15.8333 190 1.2875 0.5373 1.2875 1.1347
No log 16.0 192 1.2137 0.5113 1.2137 1.1017
No log 16.1667 194 1.1400 0.5714 1.1400 1.0677
No log 16.3333 196 1.0877 0.5606 1.0877 1.0429
No log 16.5 198 1.1064 0.5496 1.1064 1.0518
No log 16.6667 200 1.1062 0.5496 1.1062 1.0518
No log 16.8333 202 1.0579 0.5758 1.0579 1.0285
No log 17.0 204 1.0041 0.6061 1.0041 1.0021
No log 17.1667 206 1.0094 0.5802 1.0094 1.0047
No log 17.3333 208 1.0701 0.5758 1.0701 1.0345
No log 17.5 210 1.0659 0.5714 1.0659 1.0324
No log 17.6667 212 1.0959 0.5926 1.0959 1.0468
No log 17.8333 214 1.0589 0.6176 1.0589 1.0290
No log 18.0 216 1.0565 0.5926 1.0565 1.0279
No log 18.1667 218 1.0780 0.5839 1.0780 1.0383
No log 18.3333 220 1.0831 0.6277 1.0831 1.0407
No log 18.5 222 1.0542 0.6176 1.0542 1.0267
No log 18.6667 224 1.0187 0.6015 1.0187 1.0093
No log 18.8333 226 1.0103 0.6 1.0103 1.0051
No log 19.0 228 1.0466 0.5538 1.0466 1.0231
No log 19.1667 230 1.0930 0.5692 1.0930 1.0455
No log 19.3333 232 1.1449 0.4925 1.1449 1.0700
No log 19.5 234 1.1061 0.5693 1.1061 1.0517
No log 19.6667 236 1.0534 0.6423 1.0534 1.0263
No log 19.8333 238 1.0662 0.5882 1.0662 1.0326
No log 20.0 240 1.1270 0.6087 1.1270 1.0616
No log 20.1667 242 1.1161 0.6043 1.1161 1.0564
No log 20.3333 244 1.1431 0.5714 1.1431 1.0691
No log 20.5 246 1.0763 0.5714 1.0763 1.0374
No log 20.6667 248 1.0288 0.5865 1.0288 1.0143
No log 20.8333 250 1.0388 0.5942 1.0388 1.0192
No log 21.0 252 1.0589 0.5693 1.0589 1.0290
No log 21.1667 254 1.0279 0.6294 1.0279 1.0139
No log 21.3333 256 0.9847 0.6429 0.9847 0.9923
No log 21.5 258 0.9680 0.6232 0.9680 0.9839
No log 21.6667 260 0.9920 0.6165 0.9920 0.9960
No log 21.8333 262 1.0395 0.5802 1.0395 1.0195
No log 22.0 264 1.1138 0.5197 1.1138 1.0554
No log 22.1667 266 1.1074 0.5385 1.1074 1.0523
No log 22.3333 268 1.0336 0.5672 1.0336 1.0167
No log 22.5 270 0.9835 0.6087 0.9835 0.9917
No log 22.6667 272 0.9695 0.5985 0.9695 0.9846
No log 22.8333 274 0.9756 0.6331 0.9756 0.9877
No log 23.0 276 0.9732 0.6015 0.9732 0.9865
No log 23.1667 278 0.9941 0.5649 0.9941 0.9970
No log 23.3333 280 1.0603 0.5909 1.0603 1.0297
No log 23.5 282 1.1749 0.5588 1.1749 1.0839
No log 23.6667 284 1.2342 0.5263 1.2342 1.1110
No log 23.8333 286 1.2000 0.5303 1.2000 1.0955
No log 24.0 288 1.1449 0.5564 1.1449 1.0700
No log 24.1667 290 1.0702 0.5564 1.0702 1.0345
No log 24.3333 292 1.0026 0.5758 1.0026 1.0013
No log 24.5 294 0.9532 0.5758 0.9532 0.9763
No log 24.6667 296 0.9489 0.5758 0.9489 0.9741
No log 24.8333 298 0.9638 0.5714 0.9638 0.9817
No log 25.0 300 1.0123 0.6212 1.0123 1.0061
No log 25.1667 302 1.0657 0.6370 1.0657 1.0323
No log 25.3333 304 1.0122 0.6212 1.0122 1.0061
No log 25.5 306 0.9740 0.6119 0.9740 0.9869
No log 25.6667 308 0.9523 0.6212 0.9523 0.9758
No log 25.8333 310 0.9441 0.6565 0.9441 0.9716
No log 26.0 312 0.9545 0.6260 0.9545 0.9770
No log 26.1667 314 0.9501 0.5802 0.9501 0.9747
No log 26.3333 316 0.9499 0.6119 0.9499 0.9746
No log 26.5 318 0.9327 0.5865 0.9327 0.9658
No log 26.6667 320 0.9082 0.6618 0.9082 0.9530
No log 26.8333 322 0.9159 0.6763 0.9159 0.9570
No log 27.0 324 0.9220 0.6714 0.9220 0.9602
No log 27.1667 326 0.9034 0.6619 0.9034 0.9505
No log 27.3333 328 0.8913 0.6324 0.8913 0.9441
No log 27.5 330 0.9175 0.6165 0.9175 0.9579
No log 27.6667 332 0.9259 0.6165 0.9259 0.9622
No log 27.8333 334 0.9103 0.6412 0.9103 0.9541
No log 28.0 336 0.9175 0.6061 0.9175 0.9579
No log 28.1667 338 0.9414 0.5926 0.9414 0.9702
No log 28.3333 340 0.9490 0.5802 0.9490 0.9742
No log 28.5 342 0.9495 0.5846 0.9495 0.9744
No log 28.6667 344 0.9691 0.6176 0.9691 0.9844
No log 28.8333 346 0.9905 0.6176 0.9905 0.9952
No log 29.0 348 0.9936 0.6074 0.9936 0.9968
No log 29.1667 350 0.9718 0.6074 0.9718 0.9858
No log 29.3333 352 0.9625 0.6061 0.9625 0.9811
No log 29.5 354 0.9969 0.5538 0.9969 0.9984
No log 29.6667 356 1.0550 0.5649 1.0550 1.0271
No log 29.8333 358 1.0900 0.5344 1.0900 1.0441
No log 30.0 360 1.0877 0.5714 1.0877 1.0429
No log 30.1667 362 1.1125 0.5373 1.1125 1.0547
No log 30.3333 364 1.0807 0.5942 1.0807 1.0395
No log 30.5 366 1.0009 0.5970 1.0009 1.0004
No log 30.6667 368 0.9489 0.6412 0.9489 0.9741
No log 30.8333 370 0.9615 0.6154 0.9615 0.9806
No log 31.0 372 0.9739 0.5984 0.9739 0.9869
No log 31.1667 374 0.9961 0.5781 0.9961 0.9980
No log 31.3333 376 1.0670 0.5865 1.0670 1.0330
No log 31.5 378 1.1232 0.5116 1.1232 1.0598
No log 31.6667 380 1.1395 0.5077 1.1395 1.0675
No log 31.8333 382 1.0750 0.5692 1.0750 1.0368
No log 32.0 384 0.9920 0.5669 0.9920 0.9960
No log 32.1667 386 0.9633 0.5938 0.9633 0.9815
No log 32.3333 388 0.9595 0.6154 0.9595 0.9796
No log 32.5 390 0.9606 0.5625 0.9606 0.9801
No log 32.6667 392 0.9473 0.6047 0.9473 0.9733
No log 32.8333 394 0.9456 0.5891 0.9456 0.9724
No log 33.0 396 0.9338 0.6222 0.9338 0.9663
No log 33.1667 398 0.9435 0.6074 0.9435 0.9713
No log 33.3333 400 0.9793 0.5693 0.9793 0.9896
No log 33.5 402 1.0179 0.6143 1.0179 1.0089
No log 33.6667 404 0.9693 0.6043 0.9693 0.9845
No log 33.8333 406 0.9197 0.6471 0.9197 0.9590
No log 34.0 408 0.9133 0.6222 0.9133 0.9557
No log 34.1667 410 0.9196 0.6377 0.9196 0.9590
No log 34.3333 412 0.9305 0.6522 0.9305 0.9646
No log 34.5 414 0.9298 0.6522 0.9298 0.9643
No log 34.6667 416 0.9321 0.6471 0.9321 0.9655
No log 34.8333 418 0.9509 0.6364 0.9509 0.9752
No log 35.0 420 0.9638 0.6364 0.9638 0.9817
No log 35.1667 422 0.9667 0.6154 0.9667 0.9832
No log 35.3333 424 0.9762 0.6466 0.9762 0.9881
No log 35.5 426 0.9803 0.6418 0.9803 0.9901
No log 35.6667 428 1.0035 0.6418 1.0035 1.0017
No log 35.8333 430 1.0528 0.5970 1.0528 1.0261
No log 36.0 432 1.0965 0.5839 1.0965 1.0471
No log 36.1667 434 1.0834 0.5630 1.0834 1.0409
No log 36.3333 436 1.0341 0.6015 1.0341 1.0169
No log 36.5 438 0.9754 0.6 0.9754 0.9876
No log 36.6667 440 0.9428 0.6260 0.9428 0.9710
No log 36.8333 442 0.9154 0.6667 0.9154 0.9568
No log 37.0 444 0.9122 0.6716 0.9122 0.9551
No log 37.1667 446 0.9125 0.6308 0.9125 0.9552
No log 37.3333 448 0.9330 0.5984 0.9330 0.9659
No log 37.5 450 0.9612 0.6357 0.9612 0.9804
No log 37.6667 452 1.0172 0.6212 1.0172 1.0086
No log 37.8333 454 1.0353 0.5692 1.0353 1.0175
No log 38.0 456 0.9971 0.5581 0.9971 0.9985
No log 38.1667 458 0.9507 0.5802 0.9507 0.9751
No log 38.3333 460 0.9236 0.6015 0.9236 0.9610
No log 38.5 462 0.8980 0.6519 0.8980 0.9476
No log 38.6667 464 0.8618 0.6667 0.8618 0.9283
No log 38.8333 466 0.8395 0.6667 0.8395 0.9163
No log 39.0 468 0.8365 0.6617 0.8365 0.9146
No log 39.1667 470 0.8393 0.7153 0.8393 0.9162
No log 39.3333 472 0.8393 0.7059 0.8393 0.9161
No log 39.5 474 0.8536 0.6917 0.8536 0.9239
No log 39.6667 476 0.8629 0.6815 0.8629 0.9290
No log 39.8333 478 0.9069 0.6619 0.9069 0.9523
No log 40.0 480 0.9941 0.6486 0.9941 0.9970
No log 40.1667 482 1.0472 0.6443 1.0472 1.0233
No log 40.3333 484 0.9948 0.6573 0.9948 0.9974
No log 40.5 486 0.9312 0.6522 0.9312 0.9650
No log 40.6667 488 0.8774 0.6316 0.8774 0.9367
No log 40.8333 490 0.8636 0.6165 0.8636 0.9293
No log 41.0 492 0.8574 0.6412 0.8574 0.9260
No log 41.1667 494 0.8524 0.6617 0.8524 0.9233
No log 41.3333 496 0.8509 0.6617 0.8509 0.9224
No log 41.5 498 0.8423 0.6963 0.8423 0.9178
0.2957 41.6667 500 0.8382 0.6963 0.8382 0.9155
0.2957 41.8333 502 0.8401 0.6716 0.8401 0.9165
0.2957 42.0 504 0.8465 0.6767 0.8465 0.9201
0.2957 42.1667 506 0.8624 0.6767 0.8624 0.9286
0.2957 42.3333 508 0.8846 0.6667 0.8846 0.9405
0.2957 42.5 510 0.9242 0.6462 0.9242 0.9613

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k3_task1_organization

Finetuned
(4019)
this model