ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k8_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7652
  • Qwk: 0.6876
  • Mse: 0.7652
  • Rmse: 0.8747

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0426 2 5.1290 -0.0267 5.1290 2.2647
No log 0.0851 4 2.8137 0.0945 2.8137 1.6774
No log 0.1277 6 1.5807 0.1111 1.5807 1.2573
No log 0.1702 8 1.3821 0.1152 1.3821 1.1756
No log 0.2128 10 1.1704 0.1848 1.1704 1.0818
No log 0.2553 12 1.2191 0.1567 1.2191 1.1041
No log 0.2979 14 1.3354 0.1920 1.3354 1.1556
No log 0.3404 16 1.4753 -0.0008 1.4753 1.2146
No log 0.3830 18 1.7196 -0.0031 1.7196 1.3113
No log 0.4255 20 1.5587 -0.0105 1.5587 1.2485
No log 0.4681 22 1.3481 0.1449 1.3481 1.1611
No log 0.5106 24 1.2442 0.2514 1.2442 1.1155
No log 0.5532 26 1.1872 0.2869 1.1872 1.0896
No log 0.5957 28 1.1682 0.1218 1.1682 1.0808
No log 0.6383 30 1.2844 0.1224 1.2844 1.1333
No log 0.6809 32 1.3499 0.1882 1.3499 1.1619
No log 0.7234 34 1.2693 0.1173 1.2693 1.1266
No log 0.7660 36 1.1612 0.2902 1.1612 1.0776
No log 0.8085 38 1.2254 0.2116 1.2254 1.1070
No log 0.8511 40 1.4816 0.0946 1.4816 1.2172
No log 0.8936 42 1.4669 0.0829 1.4669 1.2112
No log 0.9362 44 1.2644 0.1495 1.2644 1.1245
No log 0.9787 46 1.1477 0.2587 1.1477 1.0713
No log 1.0213 48 1.1174 0.3015 1.1174 1.0571
No log 1.0638 50 1.0785 0.2943 1.0785 1.0385
No log 1.1064 52 1.0571 0.2964 1.0571 1.0282
No log 1.1489 54 1.0895 0.3061 1.0895 1.0438
No log 1.1915 56 1.3182 0.2886 1.3182 1.1481
No log 1.2340 58 1.6592 0.2844 1.6592 1.2881
No log 1.2766 60 2.1351 0.2244 2.1351 1.4612
No log 1.3191 62 2.3925 0.1896 2.3925 1.5468
No log 1.3617 64 2.2834 0.2204 2.2834 1.5111
No log 1.4043 66 1.8869 0.2609 1.8869 1.3736
No log 1.4468 68 1.3980 0.4233 1.3980 1.1824
No log 1.4894 70 1.4009 0.4233 1.4009 1.1836
No log 1.5319 72 1.3039 0.4215 1.3039 1.1419
No log 1.5745 74 1.2527 0.4266 1.2527 1.1192
No log 1.6170 76 1.5784 0.3460 1.5784 1.2564
No log 1.6596 78 1.5617 0.3504 1.5617 1.2497
No log 1.7021 80 1.1156 0.5294 1.1156 1.0562
No log 1.7447 82 0.8723 0.5786 0.8723 0.9339
No log 1.7872 84 0.8187 0.5855 0.8187 0.9048
No log 1.8298 86 0.8369 0.5757 0.8369 0.9148
No log 1.8723 88 0.8866 0.5601 0.8866 0.9416
No log 1.9149 90 0.8652 0.5907 0.8652 0.9302
No log 1.9574 92 0.8369 0.5811 0.8369 0.9148
No log 2.0 94 0.8786 0.5795 0.8786 0.9373
No log 2.0426 96 0.9282 0.5698 0.9282 0.9634
No log 2.0851 98 0.9100 0.6018 0.9100 0.9540
No log 2.1277 100 0.8691 0.6023 0.8691 0.9322
No log 2.1702 102 0.8709 0.6545 0.8709 0.9332
No log 2.2128 104 0.8412 0.6890 0.8412 0.9172
No log 2.2553 106 0.8190 0.6928 0.8190 0.9050
No log 2.2979 108 0.8188 0.6794 0.8188 0.9049
No log 2.3404 110 0.7958 0.6639 0.7958 0.8921
No log 2.3830 112 0.8449 0.6672 0.8449 0.9192
No log 2.4255 114 1.0413 0.5631 1.0413 1.0205
No log 2.4681 116 0.9599 0.5770 0.9599 0.9797
No log 2.5106 118 0.7623 0.6566 0.7623 0.8731
No log 2.5532 120 0.7566 0.6630 0.7566 0.8698
No log 2.5957 122 0.7341 0.7169 0.7341 0.8568
No log 2.6383 124 0.6804 0.6847 0.6804 0.8249
No log 2.6809 126 0.7300 0.6573 0.7300 0.8544
No log 2.7234 128 0.7042 0.6663 0.7042 0.8392
No log 2.7660 130 0.8731 0.6359 0.8731 0.9344
No log 2.8085 132 1.2626 0.5936 1.2626 1.1236
No log 2.8511 134 1.2711 0.5770 1.2711 1.1274
No log 2.8936 136 1.0044 0.6297 1.0044 1.0022
No log 2.9362 138 0.7615 0.6999 0.7615 0.8727
No log 2.9787 140 0.7369 0.6835 0.7369 0.8584
No log 3.0213 142 0.7272 0.6880 0.7272 0.8528
No log 3.0638 144 0.7272 0.7102 0.7272 0.8528
No log 3.1064 146 0.8458 0.6650 0.8458 0.9196
No log 3.1489 148 0.8474 0.6467 0.8474 0.9205
No log 3.1915 150 0.7684 0.6537 0.7684 0.8766
No log 3.2340 152 0.7436 0.6644 0.7436 0.8623
No log 3.2766 154 0.8325 0.5950 0.8325 0.9124
No log 3.3191 156 0.8558 0.5959 0.8558 0.9251
No log 3.3617 158 0.8375 0.6401 0.8375 0.9151
No log 3.4043 160 0.8775 0.6710 0.8775 0.9367
No log 3.4468 162 0.8909 0.6637 0.8909 0.9439
No log 3.4894 164 0.8364 0.6528 0.8364 0.9145
No log 3.5319 166 0.8259 0.6185 0.8259 0.9088
No log 3.5745 168 0.8270 0.6291 0.8270 0.9094
No log 3.6170 170 0.7649 0.6350 0.7649 0.8746
No log 3.6596 172 0.7578 0.6671 0.7578 0.8705
No log 3.7021 174 0.8702 0.6604 0.8702 0.9328
No log 3.7447 176 0.9083 0.6602 0.9083 0.9531
No log 3.7872 178 0.8220 0.6389 0.8220 0.9066
No log 3.8298 180 0.7578 0.6751 0.7578 0.8705
No log 3.8723 182 0.7215 0.6489 0.7215 0.8494
No log 3.9149 184 0.7449 0.6746 0.7449 0.8631
No log 3.9574 186 0.7440 0.6789 0.7440 0.8626
No log 4.0 188 0.7548 0.7006 0.7548 0.8688
No log 4.0426 190 0.9422 0.6654 0.9422 0.9707
No log 4.0851 192 1.0901 0.6524 1.0901 1.0441
No log 4.1277 194 0.9891 0.6757 0.9891 0.9945
No log 4.1702 196 0.7977 0.7027 0.7977 0.8931
No log 4.2128 198 0.7527 0.6926 0.7527 0.8676
No log 4.2553 200 0.7503 0.6995 0.7503 0.8662
No log 4.2979 202 0.7067 0.6941 0.7067 0.8406
No log 4.3404 204 0.7509 0.6925 0.7509 0.8665
No log 4.3830 206 0.7993 0.6883 0.7993 0.8940
No log 4.4255 208 0.7423 0.6965 0.7423 0.8616
No log 4.4681 210 0.7027 0.7081 0.7027 0.8383
No log 4.5106 212 0.7514 0.6802 0.7514 0.8668
No log 4.5532 214 0.7546 0.6952 0.7546 0.8687
No log 4.5957 216 0.7191 0.7238 0.7191 0.8480
No log 4.6383 218 0.7263 0.7271 0.7263 0.8522
No log 4.6809 220 0.8071 0.6669 0.8071 0.8984
No log 4.7234 222 0.8106 0.6605 0.8106 0.9003
No log 4.7660 224 0.7427 0.7268 0.7427 0.8618
No log 4.8085 226 0.6899 0.7063 0.6899 0.8306
No log 4.8511 228 0.6744 0.7263 0.6744 0.8212
No log 4.8936 230 0.6718 0.7211 0.6718 0.8196
No log 4.9362 232 0.6975 0.7119 0.6975 0.8352
No log 4.9787 234 0.7528 0.6846 0.7528 0.8676
No log 5.0213 236 0.7929 0.6829 0.7929 0.8905
No log 5.0638 238 0.7811 0.6952 0.7811 0.8838
No log 5.1064 240 0.7699 0.6894 0.7699 0.8775
No log 5.1489 242 0.7524 0.6983 0.7524 0.8674
No log 5.1915 244 0.7260 0.6912 0.7260 0.8521
No log 5.2340 246 0.7097 0.6838 0.7097 0.8424
No log 5.2766 248 0.6931 0.7273 0.6931 0.8326
No log 5.3191 250 0.6929 0.7160 0.6929 0.8324
No log 5.3617 252 0.6979 0.7328 0.6979 0.8354
No log 5.4043 254 0.7251 0.6890 0.7251 0.8515
No log 5.4468 256 0.7521 0.6929 0.7521 0.8673
No log 5.4894 258 0.7058 0.7228 0.7058 0.8401
No log 5.5319 260 0.6788 0.6993 0.6788 0.8239
No log 5.5745 262 0.7075 0.6752 0.7075 0.8411
No log 5.6170 264 0.6755 0.6940 0.6755 0.8219
No log 5.6596 266 0.6493 0.7149 0.6493 0.8058
No log 5.7021 268 0.6457 0.7350 0.6457 0.8036
No log 5.7447 270 0.6542 0.7350 0.6542 0.8088
No log 5.7872 272 0.6752 0.7284 0.6752 0.8217
No log 5.8298 274 0.6909 0.6948 0.6909 0.8312
No log 5.8723 276 0.6975 0.6782 0.6975 0.8352
No log 5.9149 278 0.6735 0.7226 0.6735 0.8207
No log 5.9574 280 0.6651 0.7629 0.6651 0.8155
No log 6.0 282 0.6722 0.7592 0.6722 0.8199
No log 6.0426 284 0.6851 0.7535 0.6851 0.8277
No log 6.0851 286 0.7120 0.7100 0.7120 0.8438
No log 6.1277 288 0.7702 0.6844 0.7702 0.8776
No log 6.1702 290 0.7902 0.6705 0.7902 0.8889
No log 6.2128 292 0.7731 0.6643 0.7731 0.8793
No log 6.2553 294 0.7175 0.6831 0.7175 0.8470
No log 6.2979 296 0.7023 0.6635 0.7023 0.8380
No log 6.3404 298 0.6994 0.6663 0.6994 0.8363
No log 6.3830 300 0.6901 0.6786 0.6901 0.8307
No log 6.4255 302 0.6927 0.6700 0.6927 0.8323
No log 6.4681 304 0.7161 0.6562 0.7161 0.8462
No log 6.5106 306 0.7620 0.6537 0.7620 0.8729
No log 6.5532 308 0.7627 0.6629 0.7627 0.8733
No log 6.5957 310 0.7274 0.6762 0.7274 0.8529
No log 6.6383 312 0.7122 0.6760 0.7122 0.8439
No log 6.6809 314 0.7177 0.6458 0.7177 0.8472
No log 6.7234 316 0.7320 0.6267 0.7320 0.8556
No log 6.7660 318 0.7344 0.6314 0.7344 0.8570
No log 6.8085 320 0.7247 0.6623 0.7247 0.8513
No log 6.8511 322 0.7473 0.6823 0.7473 0.8645
No log 6.8936 324 0.7836 0.6747 0.7836 0.8852
No log 6.9362 326 0.7922 0.6835 0.7922 0.8901
No log 6.9787 328 0.7655 0.6643 0.7655 0.8749
No log 7.0213 330 0.7523 0.6926 0.7523 0.8673
No log 7.0638 332 0.7366 0.6877 0.7366 0.8583
No log 7.1064 334 0.7391 0.6877 0.7391 0.8597
No log 7.1489 336 0.7623 0.6609 0.7623 0.8731
No log 7.1915 338 0.8316 0.6706 0.8316 0.9119
No log 7.2340 340 0.8728 0.6782 0.8728 0.9342
No log 7.2766 342 0.8618 0.6683 0.8618 0.9284
No log 7.3191 344 0.8267 0.6680 0.8267 0.9092
No log 7.3617 346 0.8134 0.6722 0.8134 0.9019
No log 7.4043 348 0.7989 0.6702 0.7989 0.8938
No log 7.4468 350 0.7857 0.6769 0.7857 0.8864
No log 7.4894 352 0.7886 0.6702 0.7886 0.8880
No log 7.5319 354 0.7910 0.6702 0.7910 0.8894
No log 7.5745 356 0.7895 0.6548 0.7895 0.8886
No log 7.6170 358 0.7810 0.6590 0.7810 0.8837
No log 7.6596 360 0.8008 0.6540 0.8008 0.8949
No log 7.7021 362 0.8402 0.6590 0.8402 0.9166
No log 7.7447 364 0.9029 0.6496 0.9029 0.9502
No log 7.7872 366 0.9635 0.6544 0.9635 0.9816
No log 7.8298 368 0.9671 0.6544 0.9671 0.9834
No log 7.8723 370 0.9176 0.6537 0.9176 0.9579
No log 7.9149 372 0.8473 0.6567 0.8473 0.9205
No log 7.9574 374 0.7799 0.6525 0.7799 0.8831
No log 8.0 376 0.7404 0.6881 0.7404 0.8605
No log 8.0426 378 0.7326 0.7138 0.7326 0.8559
No log 8.0851 380 0.7461 0.6859 0.7461 0.8638
No log 8.1277 382 0.7488 0.6859 0.7488 0.8653
No log 8.1702 384 0.7434 0.6859 0.7434 0.8622
No log 8.2128 386 0.7458 0.6859 0.7458 0.8636
No log 8.2553 388 0.7513 0.6894 0.7513 0.8668
No log 8.2979 390 0.7564 0.6852 0.7564 0.8697
No log 8.3404 392 0.7751 0.6794 0.7751 0.8804
No log 8.3830 394 0.7836 0.6857 0.7836 0.8852
No log 8.4255 396 0.7804 0.6870 0.7804 0.8834
No log 8.4681 398 0.7567 0.6794 0.7567 0.8699
No log 8.5106 400 0.7296 0.6852 0.7296 0.8542
No log 8.5532 402 0.7141 0.6894 0.7141 0.8451
No log 8.5957 404 0.7117 0.6888 0.7117 0.8436
No log 8.6383 406 0.7171 0.6888 0.7171 0.8468
No log 8.6809 408 0.7249 0.6857 0.7249 0.8514
No log 8.7234 410 0.7261 0.6930 0.7261 0.8521
No log 8.7660 412 0.7228 0.6868 0.7228 0.8502
No log 8.8085 414 0.7122 0.6806 0.7122 0.8439
No log 8.8511 416 0.7074 0.6805 0.7074 0.8411
No log 8.8936 418 0.7043 0.6805 0.7043 0.8392
No log 8.9362 420 0.7012 0.6906 0.7012 0.8373
No log 8.9787 422 0.6994 0.6938 0.6994 0.8363
No log 9.0213 424 0.7079 0.6805 0.7079 0.8414
No log 9.0638 426 0.7156 0.6832 0.7156 0.8459
No log 9.1064 428 0.7167 0.6820 0.7167 0.8466
No log 9.1489 430 0.7153 0.6918 0.7153 0.8458
No log 9.1915 432 0.7175 0.6918 0.7175 0.8471
No log 9.2340 434 0.7160 0.6918 0.7160 0.8462
No log 9.2766 436 0.7132 0.6944 0.7132 0.8445
No log 9.3191 438 0.7091 0.7051 0.7091 0.8421
No log 9.3617 440 0.7097 0.6926 0.7097 0.8424
No log 9.4043 442 0.7143 0.6900 0.7143 0.8452
No log 9.4468 444 0.7206 0.6918 0.7206 0.8489
No log 9.4894 446 0.7239 0.6931 0.7239 0.8508
No log 9.5319 448 0.7284 0.6924 0.7284 0.8535
No log 9.5745 450 0.7350 0.6828 0.7350 0.8573
No log 9.6170 452 0.7432 0.6811 0.7432 0.8621
No log 9.6596 454 0.7486 0.6811 0.7486 0.8652
No log 9.7021 456 0.7554 0.6894 0.7554 0.8692
No log 9.7447 458 0.7610 0.6876 0.7610 0.8724
No log 9.7872 460 0.7649 0.6876 0.7649 0.8746
No log 9.8298 462 0.7647 0.6876 0.7647 0.8745
No log 9.8723 464 0.7637 0.6876 0.7637 0.8739
No log 9.9149 466 0.7643 0.6876 0.7643 0.8743
No log 9.9574 468 0.7646 0.6876 0.7646 0.8744
No log 10.0 470 0.7652 0.6876 0.7652 0.8747

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k8_task1_organization

Finetuned
(4023)
this model