ArabicNewSplits5_FineTuningAraBERT_run2_AugV5_k8_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6048
  • Qwk: 0.2893
  • Mse: 0.6048
  • Rmse: 0.7777

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0417 2 3.3908 -0.0053 3.3908 1.8414
No log 0.0833 4 1.9450 -0.0424 1.9450 1.3946
No log 0.125 6 1.3533 0.0255 1.3533 1.1633
No log 0.1667 8 1.0953 0.0588 1.0953 1.0466
No log 0.2083 10 0.8841 0.1385 0.8841 0.9403
No log 0.25 12 0.8083 0.0531 0.8083 0.8990
No log 0.2917 14 0.9662 0.0078 0.9662 0.9829
No log 0.3333 16 1.3155 0.0 1.3155 1.1470
No log 0.375 18 1.7059 -0.0327 1.7059 1.3061
No log 0.4167 20 1.9172 -0.0070 1.9172 1.3846
No log 0.4583 22 1.6078 0.0 1.6078 1.2680
No log 0.5 24 1.2485 0.0 1.2485 1.1174
No log 0.5417 26 1.0247 0.0 1.0247 1.0123
No log 0.5833 28 0.9303 0.0 0.9303 0.9645
No log 0.625 30 0.8010 -0.0741 0.8010 0.8950
No log 0.6667 32 0.6921 -0.1556 0.6921 0.8319
No log 0.7083 34 0.6844 -0.0233 0.6844 0.8273
No log 0.75 36 0.7185 -0.0556 0.7185 0.8477
No log 0.7917 38 0.7843 -0.1692 0.7843 0.8856
No log 0.8333 40 0.9162 0.0078 0.9162 0.9572
No log 0.875 42 1.0531 0.0038 1.0531 1.0262
No log 0.9167 44 1.5203 0.0 1.5203 1.2330
No log 0.9583 46 1.6901 0.0 1.6901 1.3000
No log 1.0 48 1.6139 0.0 1.6139 1.2704
No log 1.0417 50 1.4053 0.0 1.4053 1.1854
No log 1.0833 52 1.1007 0.0 1.1007 1.0491
No log 1.125 54 0.8881 0.0159 0.8881 0.9424
No log 1.1667 56 0.7177 0.0952 0.7177 0.8472
No log 1.2083 58 0.6496 0.0569 0.6496 0.8060
No log 1.25 60 0.6571 0.1773 0.6571 0.8106
No log 1.2917 62 0.8475 0.0085 0.8475 0.9206
No log 1.3333 64 1.1951 -0.0233 1.1951 1.0932
No log 1.375 66 1.3348 -0.0268 1.3348 1.1553
No log 1.4167 68 1.2656 -0.0233 1.2656 1.1250
No log 1.4583 70 0.8049 0.1238 0.8049 0.8972
No log 1.5 72 0.6136 0.0569 0.6136 0.7833
No log 1.5417 74 0.6172 0.0569 0.6172 0.7856
No log 1.5833 76 0.6260 0.0569 0.6260 0.7912
No log 1.625 78 0.6543 -0.0853 0.6543 0.8089
No log 1.6667 80 0.9680 0.0 0.9680 0.9839
No log 1.7083 82 1.5751 -0.0233 1.5751 1.2550
No log 1.75 84 1.8811 0.0 1.8811 1.3715
No log 1.7917 86 1.7331 -0.0268 1.7331 1.3165
No log 1.8333 88 1.3815 -0.0196 1.3815 1.1754
No log 1.875 90 0.9536 0.0222 0.9536 0.9765
No log 1.9167 92 0.6655 0.1008 0.6655 0.8158
No log 1.9583 94 0.6899 0.0 0.6899 0.8306
No log 2.0 96 0.6724 0.0 0.6724 0.8200
No log 2.0417 98 0.6355 0.0476 0.6355 0.7972
No log 2.0833 100 0.6975 0.1724 0.6975 0.8351
No log 2.125 102 0.7499 0.0899 0.7499 0.8660
No log 2.1667 104 0.6748 0.0920 0.6748 0.8215
No log 2.2083 106 0.7490 0.1489 0.7490 0.8654
No log 2.25 108 0.8596 0.1256 0.8596 0.9272
No log 2.2917 110 0.7770 0.1503 0.7770 0.8815
No log 2.3333 112 0.7115 0.1158 0.7115 0.8435
No log 2.375 114 0.9169 0.1256 0.9169 0.9575
No log 2.4167 116 0.9919 0.1441 0.9919 0.9959
No log 2.4583 118 0.6813 0.2315 0.6813 0.8254
No log 2.5 120 0.9499 0.1931 0.9499 0.9746
No log 2.5417 122 1.1607 0.0815 1.1607 1.0774
No log 2.5833 124 1.1608 0.1396 1.1608 1.0774
No log 2.625 126 0.9939 0.2000 0.9939 0.9970
No log 2.6667 128 0.5921 0.2251 0.5921 0.7695
No log 2.7083 130 0.8181 0.1925 0.8181 0.9045
No log 2.75 132 0.8014 0.1925 0.8014 0.8952
No log 2.7917 134 0.6104 0.2967 0.6104 0.7813
No log 2.8333 136 0.7031 0.2621 0.7031 0.8385
No log 2.875 138 0.9826 0.1939 0.9826 0.9913
No log 2.9167 140 0.9721 0.1939 0.9721 0.9860
No log 2.9583 142 0.6557 0.2653 0.6557 0.8098
No log 3.0 144 0.6831 0.2432 0.6831 0.8265
No log 3.0417 146 1.0547 0.1020 1.0547 1.0270
No log 3.0833 148 1.1101 0.0916 1.1101 1.0536
No log 3.125 150 0.7931 0.1402 0.7931 0.8906
No log 3.1667 152 0.6200 0.2787 0.6200 0.7874
No log 3.2083 154 1.0228 0.1655 1.0228 1.0113
No log 3.25 156 1.0181 0.1429 1.0181 1.0090
No log 3.2917 158 0.6491 0.2670 0.6491 0.8057
No log 3.3333 160 0.5630 0.2683 0.5630 0.7503
No log 3.375 162 0.6020 0.2370 0.6020 0.7759
No log 3.4167 164 0.5608 0.1895 0.5608 0.7489
No log 3.4583 166 0.8881 0.1673 0.8881 0.9424
No log 3.5 168 1.1072 0.1127 1.1072 1.0522
No log 3.5417 170 0.9068 0.2000 0.9068 0.9523
No log 3.5833 172 0.6663 0.2941 0.6663 0.8163
No log 3.625 174 0.5973 0.2593 0.5973 0.7729
No log 3.6667 176 0.6737 0.3905 0.6737 0.8208
No log 3.7083 178 0.9188 0.2510 0.9188 0.9585
No log 3.75 180 0.9770 0.1938 0.9770 0.9884
No log 3.7917 182 0.8342 0.2618 0.8342 0.9134
No log 3.8333 184 0.7054 0.2830 0.7054 0.8399
No log 3.875 186 0.5566 0.2457 0.5566 0.7461
No log 3.9167 188 0.6216 0.3469 0.6216 0.7884
No log 3.9583 190 0.6061 0.3469 0.6061 0.7785
No log 4.0 192 0.5389 0.2457 0.5389 0.7341
No log 4.0417 194 0.6676 0.2621 0.6676 0.8171
No log 4.0833 196 1.0313 0.1692 1.0313 1.0155
No log 4.125 198 1.1388 0.2000 1.1388 1.0671
No log 4.1667 200 0.8755 0.2333 0.8755 0.9357
No log 4.2083 202 0.5504 0.3073 0.5504 0.7419
No log 4.25 204 0.5684 0.3725 0.5684 0.7539
No log 4.2917 206 0.5580 0.4098 0.5580 0.7470
No log 4.3333 208 0.7554 0.2877 0.7554 0.8691
No log 4.375 210 1.0010 0.2000 1.0010 1.0005
No log 4.4167 212 1.0334 0.2000 1.0334 1.0165
No log 4.4583 214 0.7984 0.2676 0.7984 0.8936
No log 4.5 216 0.5791 0.2273 0.5791 0.7610
No log 4.5417 218 0.5856 0.2871 0.5856 0.7652
No log 4.5833 220 0.5878 0.3231 0.5878 0.7667
No log 4.625 222 0.5550 0.4098 0.5550 0.7450
No log 4.6667 224 0.6302 0.2169 0.6302 0.7938
No log 4.7083 226 0.8204 0.2554 0.8204 0.9057
No log 4.75 228 0.9147 0.2191 0.9147 0.9564
No log 4.7917 230 0.8085 0.2900 0.8085 0.8991
No log 4.8333 232 0.6461 0.3103 0.6461 0.8038
No log 4.875 234 0.6425 0.3103 0.6425 0.8016
No log 4.9167 236 0.7089 0.4286 0.7089 0.8420
No log 4.9583 238 0.8817 0.2195 0.8817 0.9390
No log 5.0 240 0.9661 0.2258 0.9661 0.9829
No log 5.0417 242 0.8205 0.2542 0.8205 0.9058
No log 5.0833 244 0.6885 0.4545 0.6885 0.8298
No log 5.125 246 0.5882 0.3131 0.5882 0.7670
No log 5.1667 248 0.5858 0.3131 0.5858 0.7654
No log 5.2083 250 0.6260 0.4341 0.6260 0.7912
No log 5.25 252 0.7892 0.3162 0.7892 0.8884
No log 5.2917 254 0.9051 0.2863 0.9051 0.9514
No log 5.3333 256 0.8304 0.2900 0.8304 0.9113
No log 5.375 258 0.6874 0.3010 0.6874 0.8291
No log 5.4167 260 0.5200 0.3149 0.5200 0.7211
No log 5.4583 262 0.5034 0.3953 0.5034 0.7095
No log 5.5 264 0.5019 0.3846 0.5019 0.7085
No log 5.5417 266 0.5694 0.2179 0.5694 0.7546
No log 5.5833 268 0.6285 0.2258 0.6285 0.7928
No log 5.625 270 0.5741 0.2179 0.5741 0.7577
No log 5.6667 272 0.4959 0.3898 0.4959 0.7042
No log 5.7083 274 0.5047 0.4225 0.5047 0.7104
No log 5.75 276 0.4919 0.3898 0.4919 0.7014
No log 5.7917 278 0.5372 0.3297 0.5372 0.7329
No log 5.8333 280 0.6679 0.2637 0.6679 0.8172
No log 5.875 282 0.7536 0.2986 0.7536 0.8681
No log 5.9167 284 0.7428 0.2637 0.7428 0.8619
No log 5.9583 286 0.6514 0.2251 0.6514 0.8071
No log 6.0 288 0.5371 0.3297 0.5371 0.7329
No log 6.0417 290 0.5210 0.2370 0.5210 0.7218
No log 6.0833 292 0.5516 0.2746 0.5516 0.7427
No log 6.125 294 0.6656 0.3299 0.6656 0.8158
No log 6.1667 296 0.7573 0.2637 0.7573 0.8702
No log 6.2083 298 0.7615 0.2637 0.7615 0.8727
No log 6.25 300 0.7655 0.2637 0.7655 0.8749
No log 6.2917 302 0.6850 0.3267 0.6850 0.8276
No log 6.3333 304 0.6090 0.3641 0.6090 0.7804
No log 6.375 306 0.5290 0.3402 0.5290 0.7273
No log 6.4167 308 0.5197 0.4105 0.5197 0.7209
No log 6.4583 310 0.5241 0.3402 0.5241 0.7240
No log 6.5 312 0.6096 0.4 0.6096 0.7808
No log 6.5417 314 0.7513 0.3271 0.7513 0.8668
No log 6.5833 316 0.7364 0.2941 0.7364 0.8581
No log 6.625 318 0.6358 0.2990 0.6358 0.7974
No log 6.6667 320 0.5407 0.2865 0.5407 0.7353
No log 6.7083 322 0.4999 0.3478 0.4999 0.7071
No log 6.75 324 0.5015 0.3149 0.5015 0.7082
No log 6.7917 326 0.5323 0.3333 0.5323 0.7296
No log 6.8333 328 0.5676 0.2174 0.5676 0.7534
No log 6.875 330 0.6419 0.2990 0.6419 0.8012
No log 6.9167 332 0.6714 0.2670 0.6714 0.8194
No log 6.9583 334 0.6194 0.2990 0.6194 0.7870
No log 7.0 336 0.5768 0.3641 0.5768 0.7595
No log 7.0417 338 0.5277 0.3369 0.5277 0.7265
No log 7.0833 340 0.5328 0.3508 0.5328 0.7299
No log 7.125 342 0.5348 0.3161 0.5348 0.7313
No log 7.1667 344 0.5897 0.3535 0.5897 0.7679
No log 7.2083 346 0.7369 0.2941 0.7369 0.8585
No log 7.25 348 0.8533 0.3028 0.8533 0.9237
No log 7.2917 350 0.8293 0.2593 0.8293 0.9106
No log 7.3333 352 0.6998 0.2965 0.6998 0.8365
No log 7.375 354 0.5863 0.3231 0.5863 0.7657
No log 7.4167 356 0.5330 0.3478 0.5330 0.7300
No log 7.4583 358 0.5286 0.3333 0.5286 0.7271
No log 7.5 360 0.5326 0.3563 0.5326 0.7298
No log 7.5417 362 0.5545 0.3661 0.5545 0.7447
No log 7.5833 364 0.6469 0.2653 0.6469 0.8043
No log 7.625 366 0.7373 0.2653 0.7373 0.8587
No log 7.6667 368 0.7447 0.2637 0.7447 0.8630
No log 7.7083 370 0.6868 0.2653 0.6868 0.8288
No log 7.75 372 0.6445 0.2965 0.6445 0.8028
No log 7.7917 374 0.6066 0.36 0.6066 0.7788
No log 7.8333 376 0.5844 0.3161 0.5844 0.7645
No log 7.875 378 0.5683 0.2965 0.5683 0.7539
No log 7.9167 380 0.5589 0.3171 0.5589 0.7476
No log 7.9583 382 0.5635 0.3333 0.5635 0.7507
No log 8.0 384 0.5893 0.2917 0.5893 0.7677
No log 8.0417 386 0.6338 0.2965 0.6338 0.7961
No log 8.0833 388 0.6848 0.2653 0.6848 0.8275
No log 8.125 390 0.6845 0.2653 0.6845 0.8274
No log 8.1667 392 0.6660 0.2653 0.6660 0.8161
No log 8.2083 394 0.6229 0.2577 0.6229 0.7893
No log 8.25 396 0.5881 0.2917 0.5881 0.7669
No log 8.2917 398 0.5738 0.2917 0.5738 0.7575
No log 8.3333 400 0.5592 0.2917 0.5592 0.7478
No log 8.375 402 0.5447 0.3508 0.5447 0.7380
No log 8.4167 404 0.5367 0.3073 0.5367 0.7326
No log 8.4583 406 0.5359 0.3073 0.5359 0.7320
No log 8.5 408 0.5471 0.2542 0.5471 0.7397
No log 8.5417 410 0.5703 0.2917 0.5703 0.7552
No log 8.5833 412 0.6164 0.2893 0.6164 0.7851
No log 8.625 414 0.6665 0.2653 0.6665 0.8164
No log 8.6667 416 0.6847 0.2653 0.6847 0.8274
No log 8.7083 418 0.6791 0.2653 0.6791 0.8241
No log 8.75 420 0.6632 0.2653 0.6632 0.8144
No log 8.7917 422 0.6323 0.3267 0.6323 0.7952
No log 8.8333 424 0.6033 0.2893 0.6033 0.7767
No log 8.875 426 0.5935 0.2893 0.5935 0.7704
No log 8.9167 428 0.5932 0.2893 0.5932 0.7702
No log 8.9583 430 0.5908 0.2893 0.5908 0.7687
No log 9.0 432 0.5946 0.2893 0.5946 0.7711
No log 9.0417 434 0.5971 0.2893 0.5971 0.7727
No log 9.0833 436 0.6106 0.2893 0.6106 0.7814
No log 9.125 438 0.6179 0.2893 0.6179 0.7861
No log 9.1667 440 0.6245 0.2893 0.6245 0.7903
No log 9.2083 442 0.6259 0.2893 0.6259 0.7911
No log 9.25 444 0.6222 0.2893 0.6222 0.7888
No log 9.2917 446 0.6131 0.2893 0.6131 0.7830
No log 9.3333 448 0.6009 0.2893 0.6009 0.7752
No log 9.375 450 0.5893 0.2917 0.5893 0.7676
No log 9.4167 452 0.5783 0.2917 0.5783 0.7605
No log 9.4583 454 0.5700 0.2917 0.5700 0.7550
No log 9.5 456 0.5637 0.2917 0.5637 0.7508
No log 9.5417 458 0.5631 0.2917 0.5631 0.7504
No log 9.5833 460 0.5681 0.2917 0.5681 0.7537
No log 9.625 462 0.5760 0.2917 0.5760 0.7589
No log 9.6667 464 0.5796 0.2917 0.5796 0.7613
No log 9.7083 466 0.5883 0.2893 0.5883 0.7670
No log 9.75 468 0.5982 0.2893 0.5982 0.7734
No log 9.7917 470 0.6030 0.2893 0.6030 0.7766
No log 9.8333 472 0.6046 0.2893 0.6046 0.7775
No log 9.875 474 0.6054 0.2893 0.6054 0.7781
No log 9.9167 476 0.6054 0.2893 0.6054 0.7780
No log 9.9583 478 0.6051 0.2893 0.6051 0.7779
No log 10.0 480 0.6048 0.2893 0.6048 0.7777

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits5_FineTuningAraBERT_run2_AugV5_k8_task3_organization

Finetuned
(4023)
this model