ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k10_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6088
  • Qwk: 0.4106
  • Mse: 0.6088
  • Rmse: 0.7803

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0385 2 2.6285 -0.0593 2.6285 1.6213
No log 0.0769 4 1.2769 0.0985 1.2769 1.1300
No log 0.1154 6 1.1434 -0.1038 1.1434 1.0693
No log 0.1538 8 1.1216 0.0929 1.1216 1.0591
No log 0.1923 10 0.9861 0.1178 0.9861 0.9930
No log 0.2308 12 1.1632 0.0074 1.1632 1.0785
No log 0.2692 14 1.2513 -0.1001 1.2513 1.1186
No log 0.3077 16 1.1174 0.0272 1.1174 1.0571
No log 0.3462 18 0.9601 0.0410 0.9601 0.9798
No log 0.3846 20 0.8892 -0.0426 0.8892 0.9430
No log 0.4231 22 0.8081 -0.0426 0.8081 0.8989
No log 0.4615 24 0.7471 0.0798 0.7471 0.8644
No log 0.5 26 0.7455 0.1807 0.7455 0.8634
No log 0.5385 28 0.7281 0.1236 0.7281 0.8533
No log 0.5769 30 0.7208 0.0327 0.7208 0.8490
No log 0.6154 32 0.7815 0.1050 0.7815 0.8840
No log 0.6538 34 0.8263 0.0778 0.8263 0.9090
No log 0.6923 36 0.9624 0.1639 0.9624 0.9810
No log 0.7308 38 1.0200 0.2285 1.0200 1.0099
No log 0.7692 40 1.0236 0.2642 1.0236 1.0117
No log 0.8077 42 0.8520 0.1786 0.8520 0.9231
No log 0.8462 44 0.7503 0.1094 0.7503 0.8662
No log 0.8846 46 0.8634 0.2409 0.8634 0.9292
No log 0.9231 48 0.9481 0.1911 0.9481 0.9737
No log 0.9615 50 0.8319 0.1007 0.8319 0.9121
No log 1.0 52 0.9099 0.0145 0.9099 0.9539
No log 1.0385 54 1.3161 -0.0483 1.3161 1.1472
No log 1.0769 56 1.3272 -0.0005 1.3272 1.1520
No log 1.1154 58 1.1568 0.0078 1.1568 1.0756
No log 1.1538 60 0.9448 0.0952 0.9448 0.9720
No log 1.1923 62 0.8248 0.0798 0.8248 0.9082
No log 1.2308 64 0.9004 0.0627 0.9004 0.9489
No log 1.2692 66 0.9985 0.1212 0.9985 0.9992
No log 1.3077 68 1.1436 -0.0670 1.1436 1.0694
No log 1.3462 70 1.1499 -0.0961 1.1499 1.0724
No log 1.3846 72 0.9530 0.0609 0.9530 0.9762
No log 1.4231 74 0.8422 0.0949 0.8422 0.9177
No log 1.4615 76 0.9608 0.1973 0.9608 0.9802
No log 1.5 78 1.0339 0.1565 1.0339 1.0168
No log 1.5385 80 0.9342 0.2308 0.9342 0.9665
No log 1.5769 82 0.7919 0.2051 0.7919 0.8899
No log 1.6154 84 0.7194 0.1646 0.7194 0.8482
No log 1.6538 86 0.6825 0.2206 0.6825 0.8262
No log 1.6923 88 0.6781 0.2677 0.6781 0.8235
No log 1.7308 90 0.7162 0.1282 0.7162 0.8463
No log 1.7692 92 0.8597 0.0949 0.8597 0.9272
No log 1.8077 94 1.2809 0.0233 1.2809 1.1318
No log 1.8462 96 1.2184 0.0209 1.2184 1.1038
No log 1.8846 98 0.8523 0.2526 0.8523 0.9232
No log 1.9231 100 0.7123 0.1372 0.7123 0.8440
No log 1.9615 102 0.7219 0.1744 0.7219 0.8497
No log 2.0 104 0.7187 0.2382 0.7187 0.8478
No log 2.0385 106 0.7739 0.2226 0.7739 0.8797
No log 2.0769 108 0.8926 0.2126 0.8926 0.9448
No log 2.1154 110 1.0774 0.1721 1.0774 1.0380
No log 2.1538 112 1.1043 0.1434 1.1043 1.0509
No log 2.1923 114 1.1620 0.1182 1.1620 1.0779
No log 2.2308 116 1.1067 0.2184 1.1067 1.0520
No log 2.2692 118 0.8651 0.2528 0.8651 0.9301
No log 2.3077 120 0.7483 0.1232 0.7483 0.8651
No log 2.3462 122 0.8024 0.0822 0.8024 0.8958
No log 2.3846 124 0.9477 0.0569 0.9477 0.9735
No log 2.4231 126 1.0077 0.0296 1.0077 1.0038
No log 2.4615 128 0.8589 0.0847 0.8589 0.9268
No log 2.5 130 0.8121 0.0822 0.8121 0.9012
No log 2.5385 132 0.8270 0.1215 0.8270 0.9094
No log 2.5769 134 0.8316 0.0870 0.8316 0.9119
No log 2.6154 136 0.8093 0.1255 0.8093 0.8996
No log 2.6538 138 0.8120 0.0886 0.8120 0.9011
No log 2.6923 140 0.8125 0.2214 0.8125 0.9014
No log 2.7308 142 0.7820 0.2150 0.7820 0.8843
No log 2.7692 144 0.8104 0.2214 0.8104 0.9002
No log 2.8077 146 0.8596 0.2448 0.8596 0.9272
No log 2.8462 148 0.9295 0.1798 0.9295 0.9641
No log 2.8846 150 0.9615 0.1787 0.9615 0.9806
No log 2.9231 152 0.9023 0.1595 0.9023 0.9499
No log 2.9615 154 0.8457 0.0537 0.8457 0.9196
No log 3.0 156 0.7958 0.1440 0.7958 0.8921
No log 3.0385 158 0.7872 0.1393 0.7872 0.8872
No log 3.0769 160 0.8838 0.2182 0.8838 0.9401
No log 3.1154 162 1.1053 0.0917 1.1053 1.0513
No log 3.1538 164 1.2618 0.0409 1.2618 1.1233
No log 3.1923 166 1.0794 0.1433 1.0794 1.0389
No log 3.2308 168 0.8266 0.1054 0.8266 0.9092
No log 3.2692 170 0.7734 0.3888 0.7734 0.8794
No log 3.3077 172 0.7082 0.3452 0.7082 0.8415
No log 3.3462 174 0.7074 0.2709 0.7074 0.8410
No log 3.3846 176 0.7489 0.2642 0.7489 0.8654
No log 3.4231 178 0.7411 0.3999 0.7411 0.8609
No log 3.4615 180 0.6644 0.3118 0.6644 0.8151
No log 3.5 182 0.6585 0.4007 0.6585 0.8115
No log 3.5385 184 0.6616 0.3458 0.6616 0.8134
No log 3.5769 186 0.6348 0.2987 0.6348 0.7967
No log 3.6154 188 0.8903 0.2285 0.8903 0.9436
No log 3.6538 190 1.1526 0.2827 1.1526 1.0736
No log 3.6923 192 1.0792 0.2395 1.0792 1.0389
No log 3.7308 194 0.8638 0.2602 0.8638 0.9294
No log 3.7692 196 0.6866 0.3160 0.6866 0.8286
No log 3.8077 198 0.7164 0.3700 0.7164 0.8464
No log 3.8462 200 0.7511 0.4315 0.7511 0.8667
No log 3.8846 202 0.6811 0.4007 0.6811 0.8253
No log 3.9231 204 0.6534 0.3426 0.6534 0.8083
No log 3.9615 206 0.6512 0.3258 0.6512 0.8069
No log 4.0 208 0.6446 0.3228 0.6446 0.8029
No log 4.0385 210 0.6470 0.3369 0.6470 0.8044
No log 4.0769 212 0.6659 0.3841 0.6659 0.8160
No log 4.1154 214 0.6738 0.3340 0.6738 0.8208
No log 4.1538 216 0.6876 0.3886 0.6876 0.8292
No log 4.1923 218 0.6844 0.3603 0.6844 0.8273
No log 4.2308 220 0.6900 0.2958 0.6900 0.8307
No log 4.2692 222 0.7443 0.3174 0.7443 0.8627
No log 4.3077 224 0.7443 0.2800 0.7443 0.8627
No log 4.3462 226 0.7177 0.3489 0.7177 0.8472
No log 4.3846 228 0.6926 0.3481 0.6926 0.8322
No log 4.4231 230 0.7066 0.2720 0.7066 0.8406
No log 4.4615 232 0.7235 0.2720 0.7235 0.8506
No log 4.5 234 0.7147 0.3211 0.7147 0.8454
No log 4.5385 236 0.7270 0.4149 0.7270 0.8526
No log 4.5769 238 0.7745 0.3506 0.7745 0.8800
No log 4.6154 240 0.7992 0.3590 0.7992 0.8940
No log 4.6538 242 0.7388 0.3832 0.7388 0.8595
No log 4.6923 244 0.6963 0.4696 0.6963 0.8345
No log 4.7308 246 0.6858 0.4346 0.6858 0.8282
No log 4.7692 248 0.7043 0.4595 0.7043 0.8392
No log 4.8077 250 0.7873 0.3967 0.7873 0.8873
No log 4.8462 252 0.7472 0.3967 0.7472 0.8644
No log 4.8846 254 0.6908 0.3813 0.6908 0.8312
No log 4.9231 256 0.6820 0.4268 0.6820 0.8259
No log 4.9615 258 0.6822 0.3953 0.6822 0.8260
No log 5.0 260 0.6742 0.3400 0.6742 0.8211
No log 5.0385 262 0.6883 0.3739 0.6883 0.8296
No log 5.0769 264 0.7361 0.4087 0.7361 0.8579
No log 5.1154 266 0.6811 0.4461 0.6811 0.8253
No log 5.1538 268 0.6668 0.4370 0.6668 0.8166
No log 5.1923 270 0.6785 0.4370 0.6785 0.8237
No log 5.2308 272 0.6257 0.4738 0.6257 0.7910
No log 5.2692 274 0.6156 0.3441 0.6156 0.7846
No log 5.3077 276 0.6269 0.4093 0.6269 0.7918
No log 5.3462 278 0.5842 0.4013 0.5842 0.7643
No log 5.3846 280 0.5905 0.4345 0.5905 0.7684
No log 5.4231 282 0.6728 0.3659 0.6728 0.8202
No log 5.4615 284 0.7163 0.4109 0.7163 0.8463
No log 5.5 286 0.6858 0.4353 0.6858 0.8282
No log 5.5385 288 0.6325 0.4295 0.6325 0.7953
No log 5.5769 290 0.6313 0.4640 0.6313 0.7946
No log 5.6154 292 0.6279 0.4362 0.6279 0.7924
No log 5.6538 294 0.6438 0.3856 0.6438 0.8024
No log 5.6923 296 0.6608 0.4193 0.6608 0.8129
No log 5.7308 298 0.6422 0.3656 0.6422 0.8014
No log 5.7692 300 0.6369 0.4397 0.6369 0.7981
No log 5.8077 302 0.6980 0.3526 0.6980 0.8354
No log 5.8462 304 0.7157 0.3526 0.7157 0.8460
No log 5.8846 306 0.6626 0.4700 0.6626 0.8140
No log 5.9231 308 0.6656 0.4555 0.6656 0.8159
No log 5.9615 310 0.6780 0.3495 0.6780 0.8234
No log 6.0 312 0.6789 0.4288 0.6789 0.8239
No log 6.0385 314 0.6769 0.3948 0.6769 0.8227
No log 6.0769 316 0.7101 0.3284 0.7101 0.8427
No log 6.1154 318 0.7248 0.3541 0.7248 0.8514
No log 6.1538 320 0.7784 0.3546 0.7784 0.8823
No log 6.1923 322 0.8616 0.2310 0.8616 0.9282
No log 6.2308 324 0.9005 0.2576 0.9005 0.9490
No log 6.2692 326 0.7736 0.2414 0.7736 0.8795
No log 6.3077 328 0.7229 0.3141 0.7229 0.8502
No log 6.3462 330 0.7688 0.3287 0.7688 0.8768
No log 6.3846 332 0.7767 0.1988 0.7767 0.8813
No log 6.4231 334 0.8157 0.2429 0.8157 0.9032
No log 6.4615 336 0.9519 0.2092 0.9519 0.9756
No log 6.5 338 0.9480 0.2289 0.9480 0.9736
No log 6.5385 340 0.8889 0.1816 0.8889 0.9428
No log 6.5769 342 0.9251 0.1982 0.9251 0.9618
No log 6.6154 344 0.9289 0.2570 0.9289 0.9638
No log 6.6538 346 0.9586 0.2086 0.9586 0.9791
No log 6.6923 348 0.8856 0.1203 0.8856 0.9411
No log 6.7308 350 0.7976 0.1834 0.7976 0.8931
No log 6.7692 352 0.7601 0.2502 0.7601 0.8718
No log 6.8077 354 0.7449 0.3340 0.7449 0.8631
No log 6.8462 356 0.7283 0.3375 0.7283 0.8534
No log 6.8846 358 0.7611 0.3761 0.7611 0.8724
No log 6.9231 360 0.7564 0.3716 0.7564 0.8697
No log 6.9615 362 0.7311 0.3856 0.7311 0.8550
No log 7.0 364 0.7294 0.3808 0.7294 0.8541
No log 7.0385 366 0.7392 0.4016 0.7392 0.8598
No log 7.0769 368 0.7300 0.3344 0.7300 0.8544
No log 7.1154 370 0.7169 0.3403 0.7169 0.8467
No log 7.1538 372 0.7352 0.3896 0.7352 0.8574
No log 7.1923 374 0.7104 0.4119 0.7104 0.8428
No log 7.2308 376 0.7447 0.4527 0.7447 0.8630
No log 7.2692 378 0.7306 0.4672 0.7306 0.8548
No log 7.3077 380 0.6819 0.3924 0.6819 0.8258
No log 7.3462 382 0.6576 0.4147 0.6576 0.8109
No log 7.3846 384 0.6963 0.3001 0.6963 0.8344
No log 7.4231 386 0.6861 0.4223 0.6861 0.8283
No log 7.4615 388 0.6262 0.4190 0.6262 0.7913
No log 7.5 390 0.5843 0.4378 0.5843 0.7644
No log 7.5385 392 0.6108 0.4126 0.6108 0.7816
No log 7.5769 394 0.6242 0.4384 0.6242 0.7900
No log 7.6154 396 0.6422 0.4617 0.6422 0.8014
No log 7.6538 398 0.6664 0.4295 0.6664 0.8163
No log 7.6923 400 0.7004 0.4099 0.7004 0.8369
No log 7.7308 402 0.7536 0.3528 0.7536 0.8681
No log 7.7692 404 0.7741 0.3469 0.7741 0.8798
No log 7.8077 406 0.7863 0.3224 0.7863 0.8867
No log 7.8462 408 0.8007 0.3224 0.8007 0.8948
No log 7.8846 410 0.8157 0.3353 0.8157 0.9031
No log 7.9231 412 0.7966 0.3364 0.7966 0.8925
No log 7.9615 414 0.8007 0.3418 0.8007 0.8948
No log 8.0 416 0.8774 0.3302 0.8774 0.9367
No log 8.0385 418 0.8694 0.2577 0.8694 0.9324
No log 8.0769 420 0.7507 0.3364 0.7507 0.8664
No log 8.1154 422 0.6889 0.4058 0.6889 0.8300
No log 8.1538 424 0.7430 0.4282 0.7430 0.8620
No log 8.1923 426 0.7474 0.4282 0.7474 0.8645
No log 8.2308 428 0.6880 0.4202 0.6880 0.8295
No log 8.2692 430 0.6808 0.4472 0.6808 0.8251
No log 8.3077 432 0.7327 0.3999 0.7327 0.8560
No log 8.3462 434 0.6983 0.4121 0.6983 0.8357
No log 8.3846 436 0.6710 0.4248 0.6710 0.8191
No log 8.4231 438 0.6626 0.4240 0.6626 0.8140
No log 8.4615 440 0.7116 0.4065 0.7116 0.8435
No log 8.5 442 0.7033 0.4212 0.7033 0.8386
No log 8.5385 444 0.6446 0.4201 0.6446 0.8028
No log 8.5769 446 0.6432 0.5379 0.6432 0.8020
No log 8.6154 448 0.6826 0.3612 0.6826 0.8262
No log 8.6538 450 0.6625 0.3920 0.6625 0.8139
No log 8.6923 452 0.6462 0.4342 0.6462 0.8039
No log 8.7308 454 0.6584 0.4777 0.6584 0.8114
No log 8.7692 456 0.6625 0.4941 0.6625 0.8139
No log 8.8077 458 0.6922 0.4221 0.6922 0.8320
No log 8.8462 460 0.7025 0.3964 0.7025 0.8381
No log 8.8846 462 0.6881 0.4031 0.6881 0.8295
No log 8.9231 464 0.6975 0.4085 0.6975 0.8352
No log 8.9615 466 0.7254 0.3955 0.7254 0.8517
No log 9.0 468 0.6914 0.4260 0.6914 0.8315
No log 9.0385 470 0.6654 0.3482 0.6654 0.8158
No log 9.0769 472 0.6534 0.3713 0.6534 0.8084
No log 9.1154 474 0.6510 0.3840 0.6510 0.8068
No log 9.1538 476 0.6326 0.4253 0.6326 0.7954
No log 9.1923 478 0.6655 0.4307 0.6655 0.8158
No log 9.2308 480 0.6783 0.4193 0.6783 0.8236
No log 9.2692 482 0.6680 0.4675 0.6680 0.8173
No log 9.3077 484 0.6684 0.4422 0.6684 0.8176
No log 9.3462 486 0.6647 0.4524 0.6647 0.8153
No log 9.3846 488 0.6557 0.4451 0.6557 0.8098
No log 9.4231 490 0.6493 0.4362 0.6493 0.8058
No log 9.4615 492 0.6443 0.4742 0.6443 0.8027
No log 9.5 494 0.6471 0.4190 0.6471 0.8044
No log 9.5385 496 0.6582 0.3774 0.6582 0.8113
No log 9.5769 498 0.6804 0.4025 0.6804 0.8249
0.4113 9.6154 500 0.7395 0.4783 0.7395 0.8600
0.4113 9.6538 502 0.7741 0.4321 0.7741 0.8798
0.4113 9.6923 504 0.6912 0.4149 0.6912 0.8314
0.4113 9.7308 506 0.6386 0.4432 0.6386 0.7991
0.4113 9.7692 508 0.6312 0.4605 0.6312 0.7945
0.4113 9.8077 510 0.6088 0.4106 0.6088 0.7803

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k10_task7_organization

Finetuned
(4019)
this model