ArabicNewSplits5_FineTuningAraBERT_run2_AugV5_k9_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8244
  • Qwk: 0.6091
  • Mse: 0.8244
  • Rmse: 0.9080

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0345 2 5.1862 -0.0378 5.1862 2.2773
No log 0.0690 4 3.0577 0.0812 3.0577 1.7486
No log 0.1034 6 2.0758 0.0051 2.0758 1.4408
No log 0.1379 8 2.4292 -0.1028 2.4292 1.5586
No log 0.1724 10 2.1197 -0.0538 2.1197 1.4559
No log 0.2069 12 1.4985 0.0905 1.4985 1.2241
No log 0.2414 14 1.3001 0.2057 1.3001 1.1402
No log 0.2759 16 1.1769 0.2605 1.1769 1.0848
No log 0.3103 18 1.1800 0.3648 1.1800 1.0863
No log 0.3448 20 1.1733 0.3280 1.1733 1.0832
No log 0.3793 22 1.2307 0.3653 1.2307 1.1094
No log 0.4138 24 1.6613 0.2496 1.6613 1.2889
No log 0.4483 26 2.1120 0.2388 2.1120 1.4533
No log 0.4828 28 1.9788 0.2675 1.9788 1.4067
No log 0.5172 30 1.6182 0.3065 1.6182 1.2721
No log 0.5517 32 1.3496 0.2931 1.3496 1.1617
No log 0.5862 34 1.3186 0.3334 1.3186 1.1483
No log 0.6207 36 1.6070 0.2877 1.6070 1.2677
No log 0.6552 38 2.0197 0.2404 2.0197 1.4212
No log 0.6897 40 1.8706 0.2875 1.8706 1.3677
No log 0.7241 42 1.2809 0.3309 1.2809 1.1318
No log 0.7586 44 0.9582 0.4354 0.9582 0.9789
No log 0.7931 46 1.0200 0.3792 1.0200 1.0100
No log 0.8276 48 1.0736 0.3206 1.0736 1.0361
No log 0.8621 50 0.9901 0.3910 0.9901 0.9951
No log 0.8966 52 0.9590 0.4662 0.9590 0.9793
No log 0.9310 54 1.1032 0.4964 1.1032 1.0503
No log 0.9655 56 1.3656 0.3294 1.3656 1.1686
No log 1.0 58 1.8693 0.2750 1.8693 1.3672
No log 1.0345 60 2.0892 0.2645 2.0892 1.4454
No log 1.0690 62 2.0997 0.2595 2.0997 1.4490
No log 1.1034 64 1.9481 0.2444 1.9481 1.3957
No log 1.1379 66 1.5582 0.2982 1.5582 1.2483
No log 1.1724 68 1.2422 0.4276 1.2422 1.1145
No log 1.2069 70 1.0069 0.4864 1.0069 1.0034
No log 1.2414 72 0.9569 0.5010 0.9569 0.9782
No log 1.2759 74 0.9885 0.5349 0.9885 0.9942
No log 1.3103 76 0.9916 0.5301 0.9916 0.9958
No log 1.3448 78 1.0251 0.5215 1.0251 1.0125
No log 1.3793 80 1.0781 0.5541 1.0781 1.0383
No log 1.4138 82 1.1173 0.5613 1.1173 1.0570
No log 1.4483 84 1.2440 0.5503 1.2440 1.1153
No log 1.4828 86 1.4597 0.4959 1.4597 1.2082
No log 1.5172 88 1.3854 0.4884 1.3854 1.1770
No log 1.5517 90 1.0968 0.4999 1.0968 1.0473
No log 1.5862 92 1.0354 0.5004 1.0354 1.0175
No log 1.6207 94 0.9106 0.6163 0.9106 0.9542
No log 1.6552 96 0.8705 0.5996 0.8705 0.9330
No log 1.6897 98 0.8819 0.5991 0.8819 0.9391
No log 1.7241 100 0.8762 0.5658 0.8762 0.9361
No log 1.7586 102 0.8939 0.5758 0.8939 0.9455
No log 1.7931 104 0.9885 0.5706 0.9885 0.9942
No log 1.8276 106 1.0530 0.5321 1.0530 1.0262
No log 1.8621 108 1.1204 0.4646 1.1204 1.0585
No log 1.8966 110 1.1526 0.4234 1.1526 1.0736
No log 1.9310 112 1.0426 0.5480 1.0426 1.0211
No log 1.9655 114 0.9864 0.5704 0.9864 0.9932
No log 2.0 116 0.9486 0.5881 0.9486 0.9740
No log 2.0345 118 1.0775 0.5425 1.0775 1.0380
No log 2.0690 120 1.1877 0.4808 1.1877 1.0898
No log 2.1034 122 1.0193 0.5717 1.0193 1.0096
No log 2.1379 124 0.8318 0.5682 0.8318 0.9120
No log 2.1724 126 0.7992 0.6287 0.7992 0.8940
No log 2.2069 128 0.7969 0.6017 0.7969 0.8927
No log 2.2414 130 0.8220 0.6365 0.8220 0.9066
No log 2.2759 132 0.8184 0.6720 0.8184 0.9047
No log 2.3103 134 0.8653 0.6555 0.8653 0.9302
No log 2.3448 136 0.9262 0.5937 0.9262 0.9624
No log 2.3793 138 0.9743 0.5853 0.9743 0.9871
No log 2.4138 140 0.9900 0.5962 0.9900 0.9950
No log 2.4483 142 0.9976 0.6118 0.9976 0.9988
No log 2.4828 144 1.0002 0.6432 1.0002 1.0001
No log 2.5172 146 1.0366 0.6233 1.0366 1.0181
No log 2.5517 148 1.0499 0.5901 1.0499 1.0247
No log 2.5862 150 1.0993 0.5633 1.0993 1.0485
No log 2.6207 152 1.1328 0.5153 1.1328 1.0643
No log 2.6552 154 1.0077 0.5631 1.0077 1.0038
No log 2.6897 156 0.8849 0.5859 0.8849 0.9407
No log 2.7241 158 0.8705 0.6344 0.8705 0.9330
No log 2.7586 160 0.8758 0.6288 0.8758 0.9358
No log 2.7931 162 0.8691 0.6327 0.8691 0.9323
No log 2.8276 164 0.8413 0.6212 0.8413 0.9172
No log 2.8621 166 0.8577 0.5587 0.8577 0.9261
No log 2.8966 168 0.9693 0.5391 0.9693 0.9845
No log 2.9310 170 0.9276 0.5478 0.9276 0.9631
No log 2.9655 172 0.8227 0.5835 0.8227 0.9070
No log 3.0 174 0.7952 0.5973 0.7952 0.8918
No log 3.0345 176 0.8059 0.6346 0.8059 0.8977
No log 3.0690 178 0.8350 0.6656 0.8350 0.9138
No log 3.1034 180 0.8630 0.6502 0.8630 0.9290
No log 3.1379 182 0.8751 0.6254 0.8751 0.9355
No log 3.1724 184 0.8492 0.6206 0.8492 0.9215
No log 3.2069 186 0.8503 0.6138 0.8503 0.9221
No log 3.2414 188 0.8914 0.6342 0.8914 0.9441
No log 3.2759 190 0.9021 0.6320 0.9021 0.9498
No log 3.3103 192 0.8681 0.6172 0.8681 0.9317
No log 3.3448 194 0.8997 0.6183 0.8997 0.9485
No log 3.3793 196 0.9586 0.5990 0.9586 0.9791
No log 3.4138 198 0.9985 0.5997 0.9985 0.9992
No log 3.4483 200 0.9730 0.5824 0.9730 0.9864
No log 3.4828 202 0.8893 0.5772 0.8893 0.9430
No log 3.5172 204 0.8137 0.6307 0.8137 0.9021
No log 3.5517 206 0.7808 0.6866 0.7808 0.8836
No log 3.5862 208 0.7784 0.6681 0.7784 0.8823
No log 3.6207 210 0.8037 0.6582 0.8037 0.8965
No log 3.6552 212 0.8322 0.6367 0.8322 0.9122
No log 3.6897 214 0.8242 0.6469 0.8242 0.9079
No log 3.7241 216 0.7905 0.6456 0.7905 0.8891
No log 3.7586 218 0.7697 0.6693 0.7697 0.8773
No log 3.7931 220 0.7798 0.6668 0.7798 0.8830
No log 3.8276 222 0.8125 0.6337 0.8125 0.9014
No log 3.8621 224 0.8186 0.6308 0.8186 0.9048
No log 3.8966 226 0.8101 0.6239 0.8101 0.9001
No log 3.9310 228 0.8003 0.6230 0.8003 0.8946
No log 3.9655 230 0.8002 0.6542 0.8002 0.8946
No log 4.0 232 0.8078 0.6261 0.8078 0.8988
No log 4.0345 234 0.7876 0.6253 0.7876 0.8875
No log 4.0690 236 0.7859 0.6279 0.7859 0.8865
No log 4.1034 238 0.8155 0.6437 0.8155 0.9031
No log 4.1379 240 0.9027 0.6154 0.9027 0.9501
No log 4.1724 242 0.9800 0.5809 0.9800 0.9900
No log 4.2069 244 1.0224 0.5860 1.0224 1.0111
No log 4.2414 246 1.0002 0.5898 1.0002 1.0001
No log 4.2759 248 0.9372 0.5952 0.9372 0.9681
No log 4.3103 250 0.8470 0.6253 0.8470 0.9203
No log 4.3448 252 0.8053 0.6094 0.8053 0.8974
No log 4.3793 254 0.8052 0.5762 0.8052 0.8974
No log 4.4138 256 0.8181 0.5869 0.8181 0.9045
No log 4.4483 258 0.8486 0.5992 0.8486 0.9212
No log 4.4828 260 0.8767 0.5374 0.8767 0.9363
No log 4.5172 262 0.8819 0.5362 0.8819 0.9391
No log 4.5517 264 0.8765 0.5325 0.8765 0.9362
No log 4.5862 266 0.8593 0.5820 0.8593 0.9270
No log 4.6207 268 0.8529 0.5717 0.8529 0.9235
No log 4.6552 270 0.8255 0.5880 0.8255 0.9086
No log 4.6897 272 0.8054 0.6123 0.8054 0.8974
No log 4.7241 274 0.8270 0.6630 0.8270 0.9094
No log 4.7586 276 0.8816 0.6357 0.8816 0.9389
No log 4.7931 278 0.9325 0.6212 0.9325 0.9656
No log 4.8276 280 0.9271 0.5934 0.9271 0.9629
No log 4.8621 282 0.8843 0.5554 0.8843 0.9404
No log 4.8966 284 0.8733 0.5796 0.8733 0.9345
No log 4.9310 286 0.8639 0.6110 0.8639 0.9295
No log 4.9655 288 0.8368 0.6267 0.8368 0.9148
No log 5.0 290 0.8151 0.6442 0.8151 0.9029
No log 5.0345 292 0.8249 0.6245 0.8249 0.9082
No log 5.0690 294 0.8696 0.6321 0.8696 0.9325
No log 5.1034 296 0.9125 0.6272 0.9125 0.9552
No log 5.1379 298 0.9156 0.6071 0.9156 0.9569
No log 5.1724 300 0.8853 0.6149 0.8853 0.9409
No log 5.2069 302 0.8270 0.6239 0.8270 0.9094
No log 5.2414 304 0.7786 0.6433 0.7786 0.8824
No log 5.2759 306 0.7710 0.6435 0.7710 0.8781
No log 5.3103 308 0.7713 0.6586 0.7713 0.8782
No log 5.3448 310 0.7813 0.6382 0.7813 0.8839
No log 5.3793 312 0.7920 0.6209 0.7920 0.8899
No log 5.4138 314 0.7769 0.6441 0.7769 0.8814
No log 5.4483 316 0.7613 0.6443 0.7613 0.8725
No log 5.4828 318 0.7796 0.6292 0.7796 0.8830
No log 5.5172 320 0.7924 0.6477 0.7924 0.8902
No log 5.5517 322 0.8116 0.6160 0.8116 0.9009
No log 5.5862 324 0.8286 0.6434 0.8286 0.9103
No log 5.6207 326 0.8319 0.6461 0.8319 0.9121
No log 5.6552 328 0.8366 0.6555 0.8366 0.9147
No log 5.6897 330 0.8129 0.6239 0.8129 0.9016
No log 5.7241 332 0.7796 0.6535 0.7796 0.8829
No log 5.7586 334 0.7701 0.6686 0.7701 0.8776
No log 5.7931 336 0.7783 0.6428 0.7783 0.8822
No log 5.8276 338 0.7886 0.6498 0.7886 0.8881
No log 5.8621 340 0.8053 0.6443 0.8053 0.8974
No log 5.8966 342 0.8500 0.6195 0.8500 0.9219
No log 5.9310 344 0.8878 0.6098 0.8878 0.9422
No log 5.9655 346 0.9447 0.6062 0.9447 0.9719
No log 6.0 348 0.9971 0.5868 0.9971 0.9985
No log 6.0345 350 0.9891 0.5868 0.9891 0.9946
No log 6.0690 352 0.9200 0.6044 0.9200 0.9592
No log 6.1034 354 0.8334 0.6546 0.8334 0.9129
No log 6.1379 356 0.7824 0.6502 0.7824 0.8845
No log 6.1724 358 0.7696 0.6406 0.7696 0.8773
No log 6.2069 360 0.7662 0.6475 0.7662 0.8753
No log 6.2414 362 0.7744 0.6515 0.7744 0.8800
No log 6.2759 364 0.8062 0.6573 0.8062 0.8979
No log 6.3103 366 0.8571 0.6044 0.8571 0.9258
No log 6.3448 368 0.8916 0.6054 0.8916 0.9442
No log 6.3793 370 0.8528 0.6229 0.8528 0.9234
No log 6.4138 372 0.8160 0.6206 0.8160 0.9033
No log 6.4483 374 0.7783 0.6363 0.7783 0.8822
No log 6.4828 376 0.7504 0.6261 0.7504 0.8662
No log 6.5172 378 0.7478 0.6107 0.7478 0.8647
No log 6.5517 380 0.7483 0.6096 0.7483 0.8650
No log 6.5862 382 0.7495 0.5955 0.7495 0.8657
No log 6.6207 384 0.7625 0.5958 0.7625 0.8732
No log 6.6552 386 0.7878 0.5838 0.7878 0.8876
No log 6.6897 388 0.8065 0.5595 0.8065 0.8981
No log 6.7241 390 0.8212 0.5693 0.8212 0.9062
No log 6.7586 392 0.8179 0.5671 0.8179 0.9044
No log 6.7931 394 0.8003 0.5709 0.8003 0.8946
No log 6.8276 396 0.7844 0.5890 0.7844 0.8857
No log 6.8621 398 0.7878 0.6082 0.7878 0.8876
No log 6.8966 400 0.7900 0.6243 0.7900 0.8888
No log 6.9310 402 0.7953 0.6137 0.7953 0.8918
No log 6.9655 404 0.8013 0.6137 0.8013 0.8952
No log 7.0 406 0.8065 0.6054 0.8065 0.8981
No log 7.0345 408 0.8116 0.6012 0.8116 0.9009
No log 7.0690 410 0.8182 0.5762 0.8182 0.9046
No log 7.1034 412 0.8219 0.5748 0.8219 0.9066
No log 7.1379 414 0.8165 0.5770 0.8165 0.9036
No log 7.1724 416 0.8047 0.5787 0.8047 0.8970
No log 7.2069 418 0.8103 0.6099 0.8103 0.9002
No log 7.2414 420 0.8379 0.6212 0.8379 0.9154
No log 7.2759 422 0.8432 0.6318 0.8432 0.9183
No log 7.3103 424 0.8385 0.6221 0.8385 0.9157
No log 7.3448 426 0.8218 0.6220 0.8218 0.9065
No log 7.3793 428 0.7953 0.6406 0.7953 0.8918
No log 7.4138 430 0.7705 0.6372 0.7705 0.8778
No log 7.4483 432 0.7566 0.6420 0.7566 0.8698
No log 7.4828 434 0.7525 0.6527 0.7525 0.8675
No log 7.5172 436 0.7558 0.6594 0.7558 0.8694
No log 7.5517 438 0.7631 0.6386 0.7631 0.8735
No log 7.5862 440 0.7777 0.6197 0.7777 0.8819
No log 7.6207 442 0.7835 0.6169 0.7835 0.8852
No log 7.6552 444 0.7854 0.6347 0.7854 0.8862
No log 7.6897 446 0.7961 0.6302 0.7961 0.8923
No log 7.7241 448 0.8192 0.6263 0.8192 0.9051
No log 7.7586 450 0.8525 0.6337 0.8525 0.9233
No log 7.7931 452 0.8724 0.6541 0.8724 0.9340
No log 7.8276 454 0.8730 0.6541 0.8730 0.9343
No log 7.8621 456 0.8736 0.6541 0.8736 0.9347
No log 7.8966 458 0.8654 0.6337 0.8654 0.9303
No log 7.9310 460 0.8506 0.6085 0.8506 0.9223
No log 7.9655 462 0.8392 0.5956 0.8392 0.9161
No log 8.0 464 0.8375 0.6125 0.8375 0.9152
No log 8.0345 466 0.8358 0.6120 0.8358 0.9142
No log 8.0690 468 0.8359 0.6078 0.8359 0.9143
No log 8.1034 470 0.8433 0.5947 0.8433 0.9183
No log 8.1379 472 0.8567 0.5937 0.8567 0.9256
No log 8.1724 474 0.8696 0.5941 0.8696 0.9325
No log 8.2069 476 0.8697 0.5810 0.8697 0.9326
No log 8.2414 478 0.8664 0.5941 0.8664 0.9308
No log 8.2759 480 0.8609 0.6011 0.8609 0.9278
No log 8.3103 482 0.8435 0.6108 0.8435 0.9184
No log 8.3448 484 0.8279 0.6108 0.8279 0.9099
No log 8.3793 486 0.8036 0.6136 0.8036 0.8964
No log 8.4138 488 0.7892 0.6229 0.7892 0.8884
No log 8.4483 490 0.7830 0.6168 0.7830 0.8849
No log 8.4828 492 0.7812 0.6314 0.7812 0.8838
No log 8.5172 494 0.7811 0.6361 0.7811 0.8838
No log 8.5517 496 0.7836 0.6421 0.7836 0.8852
No log 8.5862 498 0.7912 0.6383 0.7912 0.8895
0.4074 8.6207 500 0.8087 0.6313 0.8087 0.8992
0.4074 8.6552 502 0.8252 0.6366 0.8252 0.9084
0.4074 8.6897 504 0.8374 0.6351 0.8374 0.9151
0.4074 8.7241 506 0.8525 0.6351 0.8525 0.9233
0.4074 8.7586 508 0.8610 0.6414 0.8610 0.9279
0.4074 8.7931 510 0.8536 0.6329 0.8536 0.9239
0.4074 8.8276 512 0.8385 0.6502 0.8385 0.9157
0.4074 8.8621 514 0.8217 0.6255 0.8217 0.9065
0.4074 8.8966 516 0.8027 0.6301 0.8027 0.8960
0.4074 8.9310 518 0.7878 0.6234 0.7878 0.8876
0.4074 8.9655 520 0.7777 0.6287 0.7777 0.8819
0.4074 9.0 522 0.7716 0.6386 0.7716 0.8784
0.4074 9.0345 524 0.7697 0.6386 0.7697 0.8774
0.4074 9.0690 526 0.7737 0.6437 0.7737 0.8796
0.4074 9.1034 528 0.7784 0.6437 0.7784 0.8823
0.4074 9.1379 530 0.7828 0.6488 0.7828 0.8848
0.4074 9.1724 532 0.7899 0.6391 0.7899 0.8888
0.4074 9.2069 534 0.7989 0.6323 0.7989 0.8938
0.4074 9.2414 536 0.8094 0.6389 0.8094 0.8997
0.4074 9.2759 538 0.8183 0.6145 0.8183 0.9046
0.4074 9.3103 540 0.8234 0.6131 0.8234 0.9074
0.4074 9.3448 542 0.8286 0.6167 0.8286 0.9103
0.4074 9.3793 544 0.8364 0.6173 0.8364 0.9146
0.4074 9.4138 546 0.8412 0.6078 0.8412 0.9172
0.4074 9.4483 548 0.8434 0.6078 0.8434 0.9184
0.4074 9.4828 550 0.8434 0.5952 0.8434 0.9184
0.4074 9.5172 552 0.8417 0.6046 0.8417 0.9175
0.4074 9.5517 554 0.8420 0.6078 0.8420 0.9176
0.4074 9.5862 556 0.8424 0.6078 0.8424 0.9178
0.4074 9.6207 558 0.8445 0.6146 0.8445 0.9190
0.4074 9.6552 560 0.8430 0.6182 0.8430 0.9182
0.4074 9.6897 562 0.8387 0.6182 0.8387 0.9158
0.4074 9.7241 564 0.8342 0.6182 0.8342 0.9134
0.4074 9.7586 566 0.8311 0.6114 0.8311 0.9116
0.4074 9.7931 568 0.8278 0.6091 0.8278 0.9099
0.4074 9.8276 570 0.8259 0.6187 0.8259 0.9088
0.4074 9.8621 572 0.8249 0.6187 0.8249 0.9082
0.4074 9.8966 574 0.8240 0.6187 0.8240 0.9077
0.4074 9.9310 576 0.8241 0.6187 0.8241 0.9078
0.4074 9.9655 578 0.8244 0.6091 0.8244 0.9080
0.4074 10.0 580 0.8244 0.6091 0.8244 0.9080

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits5_FineTuningAraBERT_run2_AugV5_k9_task1_organization

Finetuned
(4023)
this model