ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k9_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5788
  • Qwk: 0.3725
  • Mse: 0.5788
  • Rmse: 0.7608

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.05 2 3.0480 0.0243 3.0480 1.7459
No log 0.1 4 2.0274 0.0 2.0274 1.4239
No log 0.15 6 1.3139 0.0255 1.3139 1.1462
No log 0.2 8 1.2496 -0.0327 1.2496 1.1178
No log 0.25 10 0.7973 0.0370 0.7973 0.8929
No log 0.3 12 0.7432 0.1765 0.7432 0.8621
No log 0.35 14 1.0186 0.0388 1.0186 1.0093
No log 0.4 16 1.1159 0.0388 1.1159 1.0564
No log 0.45 18 0.8029 0.1579 0.8029 0.8960
No log 0.5 20 0.6351 0.0815 0.6351 0.7969
No log 0.55 22 0.5943 -0.0081 0.5943 0.7709
No log 0.6 24 0.5801 0.0 0.5801 0.7616
No log 0.65 26 0.5615 0.0388 0.5615 0.7493
No log 0.7 28 1.2011 0.0817 1.2011 1.0959
No log 0.75 30 0.9884 0.1111 0.9884 0.9942
No log 0.8 32 0.5902 0.0256 0.5902 0.7683
No log 0.85 34 0.6365 0.0720 0.6365 0.7978
No log 0.9 36 0.6860 0.2000 0.6860 0.8283
No log 0.95 38 0.6061 0.0 0.6061 0.7785
No log 1.0 40 0.5664 0.0 0.5664 0.7526
No log 1.05 42 0.5758 0.0 0.5758 0.7588
No log 1.1 44 0.5739 0.0 0.5739 0.7576
No log 1.15 46 0.5298 0.0 0.5298 0.7278
No log 1.2 48 0.5290 0.0 0.5290 0.7273
No log 1.25 50 0.5289 0.0 0.5289 0.7272
No log 1.3 52 0.5855 0.0400 0.5855 0.7652
No log 1.35 54 0.6567 0.1913 0.6567 0.8104
No log 1.4 56 0.8138 0.1453 0.8138 0.9021
No log 1.45 58 0.6289 0.1590 0.6289 0.7930
No log 1.5 60 0.5590 0.1448 0.5590 0.7477
No log 1.55 62 0.6107 0.2000 0.6107 0.7815
No log 1.6 64 0.5629 0.1688 0.5629 0.7503
No log 1.65 66 0.5921 0.2370 0.5921 0.7695
No log 1.7 68 0.6366 0.3073 0.6366 0.7979
No log 1.75 70 0.6554 0.1813 0.6554 0.8096
No log 1.8 72 0.8733 0.1628 0.8733 0.9345
No log 1.85 74 0.5684 0.3086 0.5684 0.7539
No log 1.9 76 0.7414 0.2919 0.7414 0.8611
No log 1.95 78 0.6789 0.3402 0.6789 0.8240
No log 2.0 80 0.5692 0.2663 0.5692 0.7545
No log 2.05 82 0.6102 0.3548 0.6102 0.7811
No log 2.1 84 0.5705 0.2298 0.5705 0.7553
No log 2.15 86 0.6879 0.2184 0.6879 0.8294
No log 2.2 88 0.9164 0.1861 0.9164 0.9573
No log 2.25 90 0.7008 0.2093 0.7008 0.8371
No log 2.3 92 0.6497 0.1823 0.6497 0.8060
No log 2.35 94 0.6836 0.2251 0.6836 0.8268
No log 2.4 96 0.5905 0.1304 0.5905 0.7684
No log 2.45 98 0.7857 0.1841 0.7857 0.8864
No log 2.5 100 0.6160 0.2088 0.6160 0.7848
No log 2.55 102 0.7152 0.3103 0.7152 0.8457
No log 2.6 104 0.6149 0.1556 0.6149 0.7841
No log 2.65 106 1.1708 0.1367 1.1708 1.0820
No log 2.7 108 1.4435 0.1656 1.4435 1.2014
No log 2.75 110 0.8315 0.1776 0.8315 0.9118
No log 2.8 112 0.6941 0.3469 0.6941 0.8331
No log 2.85 114 0.6401 0.3016 0.6401 0.8001
No log 2.9 116 0.6196 0.2746 0.6196 0.7871
No log 2.95 118 0.8945 0.1535 0.8945 0.9458
No log 3.0 120 1.2372 0.1943 1.2372 1.1123
No log 3.05 122 0.7503 0.2239 0.7503 0.8662
No log 3.1 124 0.6009 0.2513 0.6009 0.7752
No log 3.15 126 0.5915 0.3182 0.5915 0.7691
No log 3.2 128 0.6258 0.2432 0.6258 0.7911
No log 3.25 130 0.6099 0.1617 0.6099 0.7810
No log 3.3 132 0.9063 0.1366 0.9063 0.9520
No log 3.35 134 1.0869 0.1008 1.0869 1.0425
No log 3.4 136 0.7834 0.1020 0.7834 0.8851
No log 3.45 138 0.8323 0.1220 0.8323 0.9123
No log 3.5 140 1.2093 0.1045 1.2093 1.0997
No log 3.55 142 0.9645 0.0847 0.9645 0.9821
No log 3.6 144 0.6953 0.1086 0.6953 0.8339
No log 3.65 146 0.6707 0.1807 0.6707 0.8189
No log 3.7 148 0.7618 0.1340 0.7618 0.8728
No log 3.75 150 0.6711 0.2432 0.6711 0.8192
No log 3.8 152 0.6559 0.2527 0.6559 0.8099
No log 3.85 154 0.7880 0.2153 0.7880 0.8877
No log 3.9 156 0.6388 0.2917 0.6388 0.7993
No log 3.95 158 0.6537 0.2893 0.6537 0.8085
No log 4.0 160 0.7289 0.2453 0.7289 0.8538
No log 4.05 162 0.8108 0.2222 0.8108 0.9005
No log 4.1 164 0.7773 0.2432 0.7773 0.8816
No log 4.15 166 0.7917 0.2146 0.7917 0.8898
No log 4.2 168 0.7133 0.2744 0.7133 0.8446
No log 4.25 170 0.6893 0.1923 0.6893 0.8303
No log 4.3 172 0.6517 0.2079 0.6517 0.8073
No log 4.35 174 0.6129 0.4019 0.6129 0.7829
No log 4.4 176 0.7200 0.2536 0.7200 0.8485
No log 4.45 178 0.6449 0.2000 0.6449 0.8031
No log 4.5 180 0.5633 0.3371 0.5633 0.7505
No log 4.55 182 0.5970 0.2967 0.5970 0.7727
No log 4.6 184 0.5323 0.4157 0.5323 0.7296
No log 4.65 186 0.8318 0.2281 0.8318 0.9121
No log 4.7 188 1.1001 0.2226 1.1001 1.0488
No log 4.75 190 0.8454 0.2203 0.8454 0.9194
No log 4.8 192 0.5858 0.3641 0.5858 0.7654
No log 4.85 194 0.5878 0.3831 0.5878 0.7667
No log 4.9 196 0.7279 0.2000 0.7279 0.8532
No log 4.95 198 1.1122 0.2109 1.1122 1.0546
No log 5.0 200 1.0419 0.1880 1.0419 1.0207
No log 5.05 202 0.7099 0.3301 0.7099 0.8426
No log 5.1 204 0.6556 0.3905 0.6556 0.8097
No log 5.15 206 0.6510 0.4074 0.6510 0.8068
No log 5.2 208 0.6810 0.2830 0.6810 0.8252
No log 5.25 210 0.9593 0.1165 0.9593 0.9795
No log 5.3 212 0.9210 0.1148 0.9210 0.9597
No log 5.35 214 0.6410 0.4010 0.6410 0.8006
No log 5.4 216 0.5971 0.4502 0.5971 0.7727
No log 5.45 218 0.6880 0.3116 0.6880 0.8295
No log 5.5 220 0.8885 0.1545 0.8885 0.9426
No log 5.55 222 1.0841 0.2174 1.0841 1.0412
No log 5.6 224 0.8579 0.1933 0.8579 0.9262
No log 5.65 226 0.5633 0.4343 0.5633 0.7505
No log 5.7 228 0.5882 0.3548 0.5882 0.7669
No log 5.75 230 0.5591 0.4167 0.5591 0.7477
No log 5.8 232 0.6480 0.3077 0.6480 0.8050
No log 5.85 234 1.0624 0.2171 1.0624 1.0307
No log 5.9 236 1.1417 0.2165 1.1417 1.0685
No log 5.95 238 0.8783 0.2459 0.8783 0.9372
No log 6.0 240 0.6387 0.3786 0.6387 0.7992
No log 6.05 242 0.6097 0.4118 0.6097 0.7808
No log 6.1 244 0.6686 0.2475 0.6686 0.8177
No log 6.15 246 0.7290 0.1402 0.7290 0.8538
No log 6.2 248 0.8428 0.2212 0.8428 0.9180
No log 6.25 250 0.7428 0.1416 0.7428 0.8619
No log 6.3 252 0.5837 0.3978 0.5837 0.7640
No log 6.35 254 0.5653 0.3913 0.5653 0.7519
No log 6.4 256 0.5866 0.4462 0.5866 0.7659
No log 6.45 258 0.6605 0.2161 0.6605 0.8127
No log 6.5 260 0.7655 0.1402 0.7655 0.8749
No log 6.55 262 0.7815 0.1416 0.7815 0.8840
No log 6.6 264 0.7131 0.2079 0.7131 0.8444
No log 6.65 266 0.6597 0.28 0.6597 0.8122
No log 6.7 268 0.6029 0.375 0.6029 0.7765
No log 6.75 270 0.5736 0.4098 0.5736 0.7574
No log 6.8 272 0.5832 0.4225 0.5832 0.7637
No log 6.85 274 0.5967 0.3061 0.5967 0.7725
No log 6.9 276 0.5855 0.4462 0.5855 0.7652
No log 6.95 278 0.5811 0.3641 0.5811 0.7623
No log 7.0 280 0.5975 0.3706 0.5975 0.7730
No log 7.05 282 0.6201 0.3301 0.6201 0.7875
No log 7.1 284 0.5930 0.3706 0.5930 0.7701
No log 7.15 286 0.5694 0.4409 0.5694 0.7546
No log 7.2 288 0.5669 0.4286 0.5669 0.7529
No log 7.25 290 0.5616 0.4409 0.5616 0.7494
No log 7.3 292 0.5931 0.3769 0.5931 0.7701
No log 7.35 294 0.6111 0.3469 0.6111 0.7817
No log 7.4 296 0.5988 0.3161 0.5988 0.7738
No log 7.45 298 0.5863 0.3469 0.5863 0.7657
No log 7.5 300 0.5705 0.3769 0.5705 0.7553
No log 7.55 302 0.5765 0.3469 0.5765 0.7593
No log 7.6 304 0.5727 0.3769 0.5727 0.7567
No log 7.65 306 0.6075 0.3786 0.6075 0.7794
No log 7.7 308 0.6456 0.3462 0.6456 0.8035
No log 7.75 310 0.7480 0.1416 0.7480 0.8648
No log 7.8 312 0.7922 0.1786 0.7922 0.8901
No log 7.85 314 0.7298 0.2072 0.7298 0.8543
No log 7.9 316 0.6409 0.3744 0.6409 0.8006
No log 7.95 318 0.5720 0.4286 0.5720 0.7563
No log 8.0 320 0.5637 0.3862 0.5637 0.7508
No log 8.05 322 0.5593 0.3862 0.5593 0.7479
No log 8.1 324 0.5621 0.4343 0.5621 0.7497
No log 8.15 326 0.6325 0.3744 0.6325 0.7953
No log 8.2 328 0.7035 0.2793 0.7035 0.8387
No log 8.25 330 0.6895 0.3067 0.6895 0.8303
No log 8.3 332 0.6157 0.3684 0.6157 0.7846
No log 8.35 334 0.5644 0.4112 0.5644 0.7513
No log 8.4 336 0.5474 0.4227 0.5474 0.7398
No log 8.45 338 0.5490 0.4167 0.5490 0.7409
No log 8.5 340 0.5457 0.4227 0.5457 0.7387
No log 8.55 342 0.5543 0.4167 0.5543 0.7445
No log 8.6 344 0.5736 0.3769 0.5736 0.7574
No log 8.65 346 0.5983 0.3725 0.5983 0.7735
No log 8.7 348 0.5949 0.3725 0.5949 0.7713
No log 8.75 350 0.5866 0.3725 0.5866 0.7659
No log 8.8 352 0.5854 0.3725 0.5854 0.7651
No log 8.85 354 0.5923 0.3725 0.5923 0.7696
No log 8.9 356 0.5913 0.3725 0.5913 0.7689
No log 8.95 358 0.6097 0.3103 0.6097 0.7808
No log 9.0 360 0.6325 0.2780 0.6325 0.7953
No log 9.05 362 0.6622 0.3143 0.6622 0.8138
No log 9.1 364 0.6601 0.3143 0.6601 0.8125
No log 9.15 366 0.6480 0.3143 0.6480 0.8050
No log 9.2 368 0.6269 0.3427 0.6269 0.7918
No log 9.25 370 0.6036 0.3398 0.6036 0.7769
No log 9.3 372 0.5774 0.3725 0.5774 0.7599
No log 9.35 374 0.5685 0.4059 0.5685 0.7540
No log 9.4 376 0.5634 0.4059 0.5634 0.7506
No log 9.45 378 0.5624 0.4059 0.5624 0.7499
No log 9.5 380 0.5639 0.4059 0.5639 0.7509
No log 9.55 382 0.5635 0.4059 0.5635 0.7507
No log 9.6 384 0.5671 0.4010 0.5671 0.7530
No log 9.65 386 0.5716 0.4010 0.5716 0.7560
No log 9.7 388 0.5700 0.4010 0.5700 0.7550
No log 9.75 390 0.5717 0.4010 0.5717 0.7561
No log 9.8 392 0.5739 0.4010 0.5739 0.7576
No log 9.85 394 0.5770 0.3725 0.5770 0.7596
No log 9.9 396 0.5785 0.3725 0.5785 0.7606
No log 9.95 398 0.5788 0.3725 0.5788 0.7608
No log 10.0 400 0.5788 0.3725 0.5788 0.7608

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k9_task3_organization

Finetuned
(4023)
this model