ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k8_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6275
  • Qwk: 0.4934
  • Mse: 0.6275
  • Rmse: 0.7921

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0690 2 2.5413 -0.0788 2.5413 1.5941
No log 0.1379 4 1.3006 0.0692 1.3006 1.1405
No log 0.2069 6 1.1754 -0.1304 1.1754 1.0841
No log 0.2759 8 1.0247 -0.0918 1.0247 1.0123
No log 0.3448 10 1.0538 -0.0007 1.0538 1.0265
No log 0.4138 12 1.0131 -0.1389 1.0131 1.0065
No log 0.4828 14 0.9353 -0.0054 0.9353 0.9671
No log 0.5517 16 0.9644 0.0053 0.9644 0.9820
No log 0.6207 18 1.0759 0.0957 1.0759 1.0372
No log 0.6897 20 1.1501 0.0975 1.1501 1.0724
No log 0.7586 22 1.0725 0.0970 1.0725 1.0356
No log 0.8276 24 0.8867 0.1372 0.8867 0.9416
No log 0.8966 26 0.8001 0.1232 0.8001 0.8945
No log 0.9655 28 0.7965 0.0966 0.7965 0.8924
No log 1.0345 30 0.7803 0.0966 0.7803 0.8833
No log 1.1034 32 0.7457 0.1718 0.7457 0.8635
No log 1.1724 34 0.8340 0.1358 0.8340 0.9132
No log 1.2414 36 0.9803 0.0970 0.9803 0.9901
No log 1.3103 38 0.9511 0.1639 0.9511 0.9752
No log 1.3793 40 0.7930 0.1786 0.7930 0.8905
No log 1.4483 42 0.7776 0.0481 0.7776 0.8818
No log 1.5172 44 0.7767 0.0937 0.7767 0.8813
No log 1.5862 46 0.7914 0.0937 0.7914 0.8896
No log 1.6552 48 0.7987 0.0937 0.7987 0.8937
No log 1.7241 50 0.8408 0.0944 0.8408 0.9169
No log 1.7931 52 0.7635 0.0937 0.7635 0.8738
No log 1.8621 54 0.7503 0.0937 0.7503 0.8662
No log 1.9310 56 0.7614 0.0937 0.7614 0.8726
No log 2.0 58 0.7287 0.1591 0.7287 0.8536
No log 2.0690 60 0.7439 0.0393 0.7439 0.8625
No log 2.1379 62 0.7656 0.0893 0.7656 0.8750
No log 2.2069 64 0.7430 0.0444 0.7430 0.8620
No log 2.2759 66 0.7531 -0.0051 0.7531 0.8678
No log 2.3448 68 0.7795 0.1550 0.7795 0.8829
No log 2.4138 70 0.7995 0.1863 0.7995 0.8941
No log 2.4828 72 0.7965 0.1407 0.7965 0.8925
No log 2.5517 74 0.7686 0.0376 0.7686 0.8767
No log 2.6207 76 0.7707 -0.0426 0.7707 0.8779
No log 2.6897 78 0.8117 0.1352 0.8117 0.9009
No log 2.7586 80 0.8000 0.0501 0.8000 0.8944
No log 2.8276 82 0.7567 0.0330 0.7567 0.8699
No log 2.8966 84 0.7568 0.3197 0.7568 0.8700
No log 2.9655 86 0.7506 0.2121 0.7506 0.8664
No log 3.0345 88 0.7545 0.1094 0.7545 0.8686
No log 3.1034 90 0.7647 0.0643 0.7647 0.8745
No log 3.1724 92 0.7551 0.0327 0.7551 0.8689
No log 3.2414 94 0.7481 0.1050 0.7481 0.8649
No log 3.3103 96 0.7469 0.2890 0.7469 0.8643
No log 3.3793 98 0.7855 0.1492 0.7855 0.8863
No log 3.4483 100 0.9540 0.1476 0.9540 0.9767
No log 3.5172 102 1.0538 0.2044 1.0538 1.0265
No log 3.5862 104 0.9654 0.1476 0.9654 0.9825
No log 3.6552 106 0.8472 0.1682 0.8472 0.9205
No log 3.7241 108 0.8145 0.2718 0.8145 0.9025
No log 3.7931 110 0.8615 0.2605 0.8615 0.9282
No log 3.8621 112 0.8661 0.2932 0.8661 0.9307
No log 3.9310 114 0.7920 0.1953 0.7920 0.8899
No log 4.0 116 0.8130 0.2621 0.8130 0.9017
No log 4.0690 118 0.8785 0.2096 0.8785 0.9373
No log 4.1379 120 0.8202 0.1140 0.8202 0.9056
No log 4.2069 122 0.8129 0.2038 0.8129 0.9016
No log 4.2759 124 0.8400 0.2325 0.8400 0.9165
No log 4.3448 126 0.8746 0.1754 0.8746 0.9352
No log 4.4138 128 0.8364 0.1259 0.8364 0.9145
No log 4.4828 130 0.8221 0.2163 0.8221 0.9067
No log 4.5517 132 0.8159 0.2458 0.8159 0.9033
No log 4.6207 134 0.7998 0.2458 0.7998 0.8943
No log 4.6897 136 0.7482 0.2802 0.7482 0.8650
No log 4.7586 138 0.7217 0.4459 0.7217 0.8495
No log 4.8276 140 0.7483 0.3155 0.7483 0.8650
No log 4.8966 142 0.7419 0.2447 0.7419 0.8613
No log 4.9655 144 0.8373 0.1941 0.8373 0.9150
No log 5.0345 146 0.9230 0.2758 0.9230 0.9607
No log 5.1034 148 0.8071 0.2130 0.8071 0.8984
No log 5.1724 150 0.7390 0.2302 0.7390 0.8597
No log 5.2414 152 0.8772 0.4295 0.8772 0.9366
No log 5.3103 154 0.9428 0.3827 0.9428 0.9710
No log 5.3793 156 0.8349 0.4144 0.8349 0.9137
No log 5.4483 158 0.7343 0.3665 0.7343 0.8569
No log 5.5172 160 0.7193 0.2965 0.7193 0.8481
No log 5.5862 162 0.7562 0.1432 0.7562 0.8696
No log 5.6552 164 0.7578 0.1967 0.7578 0.8705
No log 5.7241 166 0.7475 0.3504 0.7475 0.8646
No log 5.7931 168 0.8184 0.3314 0.8184 0.9047
No log 5.8621 170 0.9001 0.3844 0.9001 0.9487
No log 5.9310 172 0.9111 0.3505 0.9111 0.9545
No log 6.0 174 0.9161 0.3567 0.9161 0.9571
No log 6.0690 176 0.8737 0.3221 0.8737 0.9347
No log 6.1379 178 0.8788 0.3221 0.8788 0.9375
No log 6.2069 180 0.8810 0.3235 0.8810 0.9386
No log 6.2759 182 0.8647 0.3076 0.8647 0.9299
No log 6.3448 184 0.7950 0.2634 0.7950 0.8916
No log 6.4138 186 0.8203 0.1775 0.8203 0.9057
No log 6.4828 188 0.8236 0.2434 0.8236 0.9075
No log 6.5517 190 0.7953 0.2053 0.7953 0.8918
No log 6.6207 192 0.7920 0.1754 0.7920 0.8900
No log 6.6897 194 0.8102 0.2098 0.8102 0.9001
No log 6.7586 196 0.7856 0.2373 0.7856 0.8863
No log 6.8276 198 0.7636 0.3022 0.7636 0.8738
No log 6.8966 200 0.7652 0.2834 0.7652 0.8747
No log 6.9655 202 0.7427 0.2689 0.7427 0.8618
No log 7.0345 204 0.7007 0.3603 0.7007 0.8371
No log 7.1034 206 0.6916 0.2958 0.6916 0.8317
No log 7.1724 208 0.7207 0.2130 0.7207 0.8489
No log 7.2414 210 0.7473 0.2439 0.7473 0.8645
No log 7.3103 212 0.6506 0.3390 0.6506 0.8066
No log 7.3793 214 0.6428 0.4816 0.6428 0.8017
No log 7.4483 216 0.6736 0.5056 0.6736 0.8207
No log 7.5172 218 0.6605 0.4194 0.6605 0.8127
No log 7.5862 220 0.6667 0.2958 0.6667 0.8165
No log 7.6552 222 0.6794 0.2456 0.6794 0.8243
No log 7.7241 224 0.6793 0.2746 0.6793 0.8242
No log 7.7931 226 0.6794 0.2746 0.6794 0.8243
No log 7.8621 228 0.6661 0.3031 0.6661 0.8162
No log 7.9310 230 0.6604 0.2777 0.6604 0.8126
No log 8.0 232 0.6735 0.3417 0.6735 0.8207
No log 8.0690 234 0.7154 0.3656 0.7154 0.8458
No log 8.1379 236 0.7458 0.3212 0.7458 0.8636
No log 8.2069 238 0.7452 0.2874 0.7452 0.8632
No log 8.2759 240 0.7508 0.2634 0.7508 0.8665
No log 8.3448 242 0.8164 0.3101 0.8164 0.9036
No log 8.4138 244 0.9974 0.3438 0.9974 0.9987
No log 8.4828 246 1.1314 0.3576 1.1314 1.0637
No log 8.5517 248 0.9945 0.3654 0.9945 0.9972
No log 8.6207 250 0.8124 0.3796 0.8124 0.9013
No log 8.6897 252 0.8751 0.4281 0.8751 0.9354
No log 8.7586 254 0.8566 0.4327 0.8566 0.9255
No log 8.8276 256 0.7765 0.4071 0.7765 0.8812
No log 8.8966 258 0.8233 0.3944 0.8233 0.9074
No log 8.9655 260 1.0627 0.3716 1.0627 1.0309
No log 9.0345 262 1.1353 0.2910 1.1353 1.0655
No log 9.1034 264 1.0214 0.3716 1.0214 1.0107
No log 9.1724 266 0.7817 0.3781 0.7817 0.8841
No log 9.2414 268 0.6360 0.4300 0.6360 0.7975
No log 9.3103 270 0.6144 0.4217 0.6144 0.7838
No log 9.3793 272 0.6210 0.3964 0.6210 0.7880
No log 9.4483 274 0.6112 0.4217 0.6112 0.7818
No log 9.5172 276 0.5902 0.4380 0.5902 0.7683
No log 9.5862 278 0.5979 0.4114 0.5979 0.7733
No log 9.6552 280 0.6421 0.4114 0.6421 0.8013
No log 9.7241 282 0.6746 0.3293 0.6746 0.8213
No log 9.7931 284 0.7042 0.2366 0.7042 0.8391
No log 9.8621 286 0.7277 0.1935 0.7277 0.8530
No log 9.9310 288 0.8669 0.4144 0.8669 0.9311
No log 10.0 290 1.0067 0.3161 1.0067 1.0033
No log 10.0690 292 1.0081 0.2850 1.0081 1.0041
No log 10.1379 294 0.8463 0.4243 0.8463 0.9199
No log 10.2069 296 0.7379 0.4212 0.7379 0.8590
No log 10.2759 298 0.6784 0.3887 0.6784 0.8236
No log 10.3448 300 0.6716 0.3467 0.6716 0.8195
No log 10.4138 302 0.6749 0.4287 0.6749 0.8216
No log 10.4828 304 0.6521 0.4575 0.6521 0.8075
No log 10.5517 306 0.6416 0.4059 0.6416 0.8010
No log 10.6207 308 0.6493 0.4984 0.6493 0.8058
No log 10.6897 310 0.6483 0.4984 0.6483 0.8052
No log 10.7586 312 0.6353 0.4059 0.6353 0.7971
No log 10.8276 314 0.6720 0.3051 0.6720 0.8197
No log 10.8966 316 0.6793 0.3051 0.6793 0.8242
No log 10.9655 318 0.6643 0.4059 0.6643 0.8150
No log 11.0345 320 0.7894 0.4423 0.7894 0.8885
No log 11.1034 322 0.9371 0.3579 0.9371 0.9681
No log 11.1724 324 0.9551 0.3579 0.9551 0.9773
No log 11.2414 326 0.8518 0.3955 0.8518 0.9230
No log 11.3103 328 0.7563 0.2652 0.7563 0.8696
No log 11.3793 330 0.7457 0.2038 0.7457 0.8636
No log 11.4483 332 0.7350 0.2366 0.7350 0.8573
No log 11.5172 334 0.7301 0.3171 0.7301 0.8545
No log 11.5862 336 0.7562 0.3613 0.7562 0.8696
No log 11.6552 338 0.8069 0.3711 0.8069 0.8983
No log 11.7241 340 0.8201 0.3822 0.8201 0.9056
No log 11.7931 342 0.7492 0.3279 0.7492 0.8655
No log 11.8621 344 0.7172 0.4051 0.7172 0.8469
No log 11.9310 346 0.7282 0.3450 0.7282 0.8534
No log 12.0 348 0.7182 0.3738 0.7182 0.8475
No log 12.0690 350 0.7303 0.3530 0.7303 0.8546
No log 12.1379 352 0.8282 0.3569 0.8282 0.9100
No log 12.2069 354 0.9177 0.4212 0.9177 0.9580
No log 12.2759 356 0.8556 0.4081 0.8556 0.9250
No log 12.3448 358 0.7562 0.3931 0.7562 0.8696
No log 12.4138 360 0.6793 0.3667 0.6793 0.8242
No log 12.4828 362 0.6601 0.3622 0.6601 0.8125
No log 12.5517 364 0.6477 0.3622 0.6477 0.8048
No log 12.6207 366 0.6425 0.4222 0.6425 0.8015
No log 12.6897 368 0.6640 0.4455 0.6640 0.8148
No log 12.7586 370 0.6881 0.4227 0.6881 0.8295
No log 12.8276 372 0.6842 0.4001 0.6842 0.8272
No log 12.8966 374 0.6510 0.4248 0.6510 0.8068
No log 12.9655 376 0.6464 0.4432 0.6464 0.8040
No log 13.0345 378 0.6646 0.4248 0.6646 0.8152
No log 13.1034 380 0.6357 0.5021 0.6357 0.7973
No log 13.1724 382 0.6275 0.4287 0.6275 0.7922
No log 13.2414 384 0.6193 0.4287 0.6193 0.7869
No log 13.3103 386 0.5964 0.4287 0.5964 0.7723
No log 13.3793 388 0.5846 0.4575 0.5846 0.7646
No log 13.4483 390 0.5981 0.4051 0.5981 0.7733
No log 13.5172 392 0.5926 0.4051 0.5926 0.7698
No log 13.5862 394 0.5827 0.4402 0.5827 0.7634
No log 13.6552 396 0.6144 0.4249 0.6144 0.7838
No log 13.7241 398 0.6842 0.3637 0.6842 0.8272
No log 13.7931 400 0.6575 0.4237 0.6575 0.8108
No log 13.8621 402 0.5999 0.4019 0.5999 0.7745
No log 13.9310 404 0.6370 0.4336 0.6370 0.7981
No log 14.0 406 0.6650 0.4336 0.6650 0.8155
No log 14.0690 408 0.6314 0.4419 0.6314 0.7946
No log 14.1379 410 0.6298 0.3363 0.6298 0.7936
No log 14.2069 412 0.6468 0.3729 0.6468 0.8042
No log 14.2759 414 0.6498 0.3835 0.6498 0.8061
No log 14.3448 416 0.6635 0.4300 0.6635 0.8146
No log 14.4138 418 0.7201 0.4302 0.7201 0.8486
No log 14.4828 420 0.7174 0.3971 0.7174 0.8470
No log 14.5517 422 0.6662 0.4106 0.6662 0.8162
No log 14.6207 424 0.6354 0.4322 0.6354 0.7971
No log 14.6897 426 0.6583 0.4656 0.6583 0.8114
No log 14.7586 428 0.6677 0.4099 0.6677 0.8171
No log 14.8276 430 0.6465 0.4638 0.6465 0.8041
No log 14.8966 432 0.6327 0.4535 0.6327 0.7954
No log 14.9655 434 0.6333 0.4448 0.6333 0.7958
No log 15.0345 436 0.6170 0.4838 0.6170 0.7855
No log 15.1034 438 0.6036 0.4910 0.6036 0.7769
No log 15.1724 440 0.6068 0.4697 0.6068 0.7790
No log 15.2414 442 0.6048 0.4820 0.6048 0.7777
No log 15.3103 444 0.6106 0.4820 0.6106 0.7814
No log 15.3793 446 0.6339 0.4726 0.6339 0.7962
No log 15.4483 448 0.6745 0.4207 0.6745 0.8213
No log 15.5172 450 0.6718 0.4314 0.6718 0.8196
No log 15.5862 452 0.7047 0.4186 0.7047 0.8395
No log 15.6552 454 0.7315 0.4423 0.7315 0.8553
No log 15.7241 456 0.7639 0.4353 0.7639 0.8740
No log 15.7931 458 0.7109 0.4353 0.7109 0.8431
No log 15.8621 460 0.6327 0.4114 0.6327 0.7954
No log 15.9310 462 0.6352 0.4105 0.6352 0.7970
No log 16.0 464 0.7837 0.3538 0.7837 0.8853
No log 16.0690 466 0.9726 0.3373 0.9726 0.9862
No log 16.1379 468 0.9417 0.3565 0.9417 0.9704
No log 16.2069 470 0.7510 0.3655 0.7510 0.8666
No log 16.2759 472 0.6208 0.3628 0.6208 0.7879
No log 16.3448 474 0.6557 0.4801 0.6557 0.8097
No log 16.4138 476 0.7318 0.4665 0.7318 0.8555
No log 16.4828 478 0.7053 0.4409 0.7053 0.8398
No log 16.5517 480 0.6644 0.4100 0.6644 0.8151
No log 16.6207 482 0.6234 0.3445 0.6234 0.7895
No log 16.6897 484 0.6146 0.3445 0.6146 0.7839
No log 16.7586 486 0.6343 0.4925 0.6343 0.7965
No log 16.8276 488 0.6176 0.4726 0.6176 0.7859
No log 16.8966 490 0.6178 0.5214 0.6178 0.7860
No log 16.9655 492 0.6380 0.5345 0.6380 0.7987
No log 17.0345 494 0.6242 0.5214 0.6242 0.7900
No log 17.1034 496 0.6058 0.5379 0.6058 0.7783
No log 17.1724 498 0.6080 0.5379 0.6080 0.7797
0.3183 17.2414 500 0.6107 0.5379 0.6107 0.7815
0.3183 17.3103 502 0.6054 0.5159 0.6054 0.7781
0.3183 17.3793 504 0.5963 0.4463 0.5963 0.7722
0.3183 17.4483 506 0.6016 0.4463 0.6016 0.7756
0.3183 17.5172 508 0.6126 0.4934 0.6126 0.7827
0.3183 17.5862 510 0.6275 0.4934 0.6275 0.7921

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k8_task7_organization

Finetuned
(4019)
this model