ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k3_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8465
  • Qwk: 0.3798
  • Mse: 0.8465
  • Rmse: 0.9200

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1538 2 4.6676 -0.0132 4.6676 2.1605
No log 0.3077 4 2.8192 -0.0233 2.8192 1.6790
No log 0.4615 6 2.0097 -0.0370 2.0097 1.4176
No log 0.6154 8 1.4912 -0.0190 1.4912 1.2212
No log 0.7692 10 1.2517 0.0979 1.2517 1.1188
No log 0.9231 12 1.2974 0.0984 1.2974 1.1390
No log 1.0769 14 1.2047 0.1483 1.2047 1.0976
No log 1.2308 16 1.1178 0.3095 1.1178 1.0573
No log 1.3846 18 1.1032 0.3066 1.1032 1.0503
No log 1.5385 20 1.2659 0.1743 1.2659 1.1251
No log 1.6923 22 1.4707 0.2163 1.4707 1.2127
No log 1.8462 24 1.2117 0.1371 1.2117 1.1008
No log 2.0 26 1.1028 0.2813 1.1028 1.0502
No log 2.1538 28 1.1823 0.1081 1.1823 1.0873
No log 2.3077 30 1.1553 0.1773 1.1553 1.0749
No log 2.4615 32 1.1441 0.2935 1.1441 1.0696
No log 2.6154 34 1.2764 0.0898 1.2764 1.1298
No log 2.7692 36 1.3025 0.0575 1.3025 1.1413
No log 2.9231 38 1.1042 0.2337 1.1042 1.0508
No log 3.0769 40 0.9635 0.4608 0.9635 0.9816
No log 3.2308 42 0.9877 0.4273 0.9877 0.9938
No log 3.3846 44 0.9983 0.4273 0.9983 0.9992
No log 3.5385 46 0.9029 0.4541 0.9029 0.9502
No log 3.6923 48 0.8849 0.4534 0.8849 0.9407
No log 3.8462 50 0.8589 0.4804 0.8589 0.9268
No log 4.0 52 0.9089 0.4334 0.9089 0.9534
No log 4.1538 54 0.8937 0.4434 0.8937 0.9454
No log 4.3077 56 0.9754 0.4078 0.9754 0.9876
No log 4.4615 58 1.0384 0.3466 1.0384 1.0190
No log 4.6154 60 0.9194 0.4570 0.9194 0.9588
No log 4.7692 62 0.8285 0.4413 0.8285 0.9102
No log 4.9231 64 0.8375 0.4356 0.8375 0.9152
No log 5.0769 66 0.8450 0.4879 0.8450 0.9193
No log 5.2308 68 0.8630 0.4388 0.8630 0.9290
No log 5.3846 70 1.0089 0.3521 1.0089 1.0044
No log 5.5385 72 0.9235 0.4396 0.9235 0.9610
No log 5.6923 74 0.8507 0.5430 0.8507 0.9223
No log 5.8462 76 0.9400 0.5145 0.9400 0.9695
No log 6.0 78 0.8138 0.5543 0.8138 0.9021
No log 6.1538 80 1.0196 0.3642 1.0196 1.0097
No log 6.3077 82 1.2958 0.4655 1.2958 1.1383
No log 6.4615 84 1.0908 0.3676 1.0908 1.0444
No log 6.6154 86 0.7934 0.5184 0.7934 0.8907
No log 6.7692 88 0.8300 0.5346 0.8300 0.9110
No log 6.9231 90 0.7980 0.4979 0.7980 0.8933
No log 7.0769 92 0.8862 0.4098 0.8862 0.9414
No log 7.2308 94 0.9543 0.3152 0.9543 0.9769
No log 7.3846 96 0.8616 0.4158 0.8616 0.9282
No log 7.5385 98 0.8288 0.4608 0.8288 0.9104
No log 7.6923 100 0.9757 0.4119 0.9757 0.9878
No log 7.8462 102 1.0339 0.4410 1.0339 1.0168
No log 8.0 104 0.7967 0.5263 0.7967 0.8926
No log 8.1538 106 0.7527 0.4634 0.7527 0.8676
No log 8.3077 108 0.8113 0.5306 0.8113 0.9007
No log 8.4615 110 0.9996 0.4169 0.9996 0.9998
No log 8.6154 112 0.8656 0.5338 0.8656 0.9304
No log 8.7692 114 0.7214 0.5841 0.7214 0.8494
No log 8.9231 116 0.7304 0.5778 0.7304 0.8546
No log 9.0769 118 0.7736 0.4709 0.7736 0.8795
No log 9.2308 120 1.1079 0.3973 1.1079 1.0525
No log 9.3846 122 1.0304 0.4357 1.0304 1.0151
No log 9.5385 124 0.9893 0.4329 0.9893 0.9946
No log 9.6923 126 0.7550 0.4686 0.7550 0.8689
No log 9.8462 128 0.7232 0.5805 0.7232 0.8504
No log 10.0 130 0.7607 0.5528 0.7607 0.8722
No log 10.1538 132 0.7409 0.5618 0.7409 0.8608
No log 10.3077 134 0.7265 0.5905 0.7265 0.8523
No log 10.4615 136 0.8221 0.5688 0.8221 0.9067
No log 10.6154 138 0.8243 0.5688 0.8243 0.9079
No log 10.7692 140 0.7136 0.6874 0.7136 0.8448
No log 10.9231 142 0.6954 0.5925 0.6954 0.8339
No log 11.0769 144 0.7686 0.6533 0.7686 0.8767
No log 11.2308 146 1.1078 0.4108 1.1078 1.0525
No log 11.3846 148 1.2271 0.4055 1.2271 1.1077
No log 11.5385 150 0.9197 0.5408 0.9197 0.9590
No log 11.6923 152 0.7403 0.4826 0.7403 0.8604
No log 11.8462 154 0.7657 0.5202 0.7657 0.8750
No log 12.0 156 0.7568 0.5548 0.7568 0.8700
No log 12.1538 158 0.7744 0.5055 0.7744 0.8800
No log 12.3077 160 0.8383 0.5530 0.8383 0.9156
No log 12.4615 162 0.7272 0.5102 0.7272 0.8527
No log 12.6154 164 0.7406 0.5521 0.7406 0.8606
No log 12.7692 166 0.8126 0.5160 0.8126 0.9015
No log 12.9231 168 0.7330 0.5651 0.7330 0.8562
No log 13.0769 170 0.7075 0.5107 0.7075 0.8411
No log 13.2308 172 0.7587 0.5642 0.7587 0.8710
No log 13.3846 174 0.7323 0.5684 0.7323 0.8558
No log 13.5385 176 0.7206 0.5633 0.7206 0.8489
No log 13.6923 178 0.6991 0.5595 0.6991 0.8362
No log 13.8462 180 0.7340 0.5663 0.7340 0.8567
No log 14.0 182 0.7173 0.5213 0.7173 0.8469
No log 14.1538 184 0.7394 0.4977 0.7394 0.8599
No log 14.3077 186 0.8692 0.5555 0.8692 0.9323
No log 14.4615 188 1.0299 0.4638 1.0299 1.0148
No log 14.6154 190 0.9928 0.4705 0.9928 0.9964
No log 14.7692 192 0.8243 0.4624 0.8243 0.9079
No log 14.9231 194 0.7698 0.4769 0.7698 0.8774
No log 15.0769 196 0.7643 0.4769 0.7643 0.8742
No log 15.2308 198 0.7446 0.5296 0.7446 0.8629
No log 15.3846 200 0.7590 0.4879 0.7590 0.8712
No log 15.5385 202 0.9386 0.5402 0.9386 0.9688
No log 15.6923 204 1.0040 0.5101 1.0040 1.0020
No log 15.8462 206 0.8428 0.5075 0.8428 0.9180
No log 16.0 208 0.7383 0.5093 0.7383 0.8592
No log 16.1538 210 0.7599 0.5163 0.7599 0.8717
No log 16.3077 212 0.7848 0.5296 0.7848 0.8859
No log 16.4615 214 0.8025 0.4635 0.8025 0.8958
No log 16.6154 216 0.8652 0.4513 0.8652 0.9302
No log 16.7692 218 1.0175 0.3908 1.0175 1.0087
No log 16.9231 220 1.0503 0.4153 1.0503 1.0248
No log 17.0769 222 0.8972 0.4494 0.8972 0.9472
No log 17.2308 224 0.7763 0.4825 0.7763 0.8811
No log 17.3846 226 0.7768 0.4825 0.7768 0.8814
No log 17.5385 228 0.7846 0.4825 0.7846 0.8858
No log 17.6923 230 0.8656 0.4468 0.8656 0.9304
No log 17.8462 232 0.8106 0.4175 0.8106 0.9003
No log 18.0 234 0.7599 0.5357 0.7599 0.8717
No log 18.1538 236 0.7794 0.5118 0.7794 0.8829
No log 18.3077 238 0.7928 0.5118 0.7928 0.8904
No log 18.4615 240 0.7796 0.4579 0.7796 0.8830
No log 18.6154 242 0.7724 0.5237 0.7724 0.8789
No log 18.7692 244 0.8229 0.4570 0.8229 0.9071
No log 18.9231 246 0.8101 0.4681 0.8101 0.9001
No log 19.0769 248 0.8266 0.4915 0.8266 0.9092
No log 19.2308 250 0.8388 0.4703 0.8388 0.9159
No log 19.3846 252 0.8671 0.4604 0.8671 0.9312
No log 19.5385 254 0.8426 0.4650 0.8426 0.9179
No log 19.6923 256 0.8007 0.4681 0.8007 0.8948
No log 19.8462 258 0.8083 0.4570 0.8083 0.8991
No log 20.0 260 0.8219 0.4730 0.8219 0.9066
No log 20.1538 262 0.8033 0.4696 0.8033 0.8962
No log 20.3077 264 0.7962 0.4434 0.7962 0.8923
No log 20.4615 266 0.7920 0.4158 0.7920 0.8899
No log 20.6154 268 0.7822 0.4016 0.7822 0.8844
No log 20.7692 270 0.7672 0.4016 0.7672 0.8759
No log 20.9231 272 0.7575 0.4363 0.7575 0.8703
No log 21.0769 274 0.7743 0.4444 0.7743 0.8800
No log 21.2308 276 0.8342 0.4513 0.8342 0.9133
No log 21.3846 278 0.8250 0.4513 0.8250 0.9083
No log 21.5385 280 0.8272 0.4775 0.8272 0.9095
No log 21.6923 282 0.7729 0.4841 0.7729 0.8791
No log 21.8462 284 0.7373 0.4715 0.7373 0.8586
No log 22.0 286 0.7710 0.5467 0.7710 0.8781
No log 22.1538 288 0.9701 0.4949 0.9701 0.9849
No log 22.3077 290 1.3483 0.4017 1.3483 1.1611
No log 22.4615 292 1.4321 0.3850 1.4321 1.1967
No log 22.6154 294 1.2061 0.4460 1.2061 1.0982
No log 22.7692 296 0.9757 0.4500 0.9757 0.9878
No log 22.9231 298 0.8124 0.4570 0.8124 0.9014
No log 23.0769 300 0.7949 0.4563 0.7949 0.8916
No log 23.2308 302 0.8149 0.4158 0.8149 0.9027
No log 23.3846 304 0.9251 0.3879 0.9251 0.9618
No log 23.5385 306 1.0372 0.3494 1.0372 1.0184
No log 23.6923 308 0.9748 0.4154 0.9748 0.9873
No log 23.8462 310 0.8346 0.4344 0.8346 0.9135
No log 24.0 312 0.7621 0.4626 0.7621 0.8730
No log 24.1538 314 0.7535 0.4879 0.7535 0.8680
No log 24.3077 316 0.7871 0.5098 0.7871 0.8872
No log 24.4615 318 0.8165 0.5345 0.8165 0.9036
No log 24.6154 320 0.7895 0.5022 0.7895 0.8886
No log 24.7692 322 0.8113 0.4929 0.8113 0.9007
No log 24.9231 324 0.8119 0.5102 0.8119 0.9011
No log 25.0769 326 0.7970 0.5287 0.7970 0.8927
No log 25.2308 328 0.8001 0.4671 0.8001 0.8945
No log 25.3846 330 0.8012 0.4504 0.8012 0.8951
No log 25.5385 332 0.8240 0.4797 0.8240 0.9078
No log 25.6923 334 0.8738 0.4513 0.8738 0.9348
No log 25.8462 336 0.8383 0.4717 0.8383 0.9156
No log 26.0 338 0.8092 0.3914 0.8092 0.8995
No log 26.1538 340 0.7993 0.3914 0.7993 0.8940
No log 26.3077 342 0.8067 0.4570 0.8067 0.8981
No log 26.4615 344 0.8126 0.4775 0.8126 0.9015
No log 26.6154 346 0.8304 0.4530 0.8304 0.9112
No log 26.7692 348 0.8489 0.5291 0.8489 0.9213
No log 26.9231 350 0.9987 0.4813 0.9987 0.9993
No log 27.0769 352 0.9899 0.4813 0.9899 0.9949
No log 27.2308 354 0.8373 0.5580 0.8373 0.9150
No log 27.3846 356 0.7582 0.4364 0.7582 0.8707
No log 27.5385 358 0.7491 0.4804 0.7491 0.8655
No log 27.6923 360 0.7589 0.4804 0.7589 0.8712
No log 27.8462 362 0.7950 0.4197 0.7950 0.8917
No log 28.0 364 0.9242 0.4410 0.9242 0.9614
No log 28.1538 366 1.0258 0.4497 1.0258 1.0128
No log 28.3077 368 0.9981 0.4850 0.9981 0.9991
No log 28.4615 370 0.8694 0.4503 0.8694 0.9324
No log 28.6154 372 0.7924 0.4393 0.7924 0.8902
No log 28.7692 374 0.7716 0.4591 0.7716 0.8784
No log 28.9231 376 0.7652 0.5028 0.7652 0.8747
No log 29.0769 378 0.7659 0.4691 0.7659 0.8751
No log 29.2308 380 0.7836 0.4977 0.7836 0.8852
No log 29.3846 382 0.8759 0.4832 0.8759 0.9359
No log 29.5385 384 0.9409 0.5087 0.9409 0.9700
No log 29.6923 386 0.9225 0.4878 0.9225 0.9605
No log 29.8462 388 0.8296 0.4624 0.8296 0.9108
No log 30.0 390 0.7843 0.4826 0.7843 0.8856
No log 30.1538 392 0.8038 0.4726 0.8038 0.8966
No log 30.3077 394 0.8690 0.4308 0.8690 0.9322
No log 30.4615 396 0.9624 0.4581 0.9624 0.9810
No log 30.6154 398 0.9788 0.4581 0.9788 0.9893
No log 30.7692 400 0.9882 0.4556 0.9882 0.9941
No log 30.9231 402 0.9765 0.4023 0.9765 0.9882
No log 31.0769 404 0.9005 0.3942 0.9005 0.9489
No log 31.2308 406 0.8714 0.4098 0.8714 0.9335
No log 31.3846 408 0.8701 0.4331 0.8701 0.9328
No log 31.5385 410 0.8896 0.3942 0.8896 0.9432
No log 31.6923 412 0.8667 0.4271 0.8667 0.9310
No log 31.8462 414 0.8089 0.4626 0.8089 0.8994
No log 32.0 416 0.7754 0.4691 0.7754 0.8806
No log 32.1538 418 0.7773 0.4726 0.7773 0.8816
No log 32.3077 420 0.7992 0.4233 0.7992 0.8940
No log 32.4615 422 0.8177 0.3806 0.8177 0.9042
No log 32.6154 424 0.8529 0.4175 0.8529 0.9235
No log 32.7692 426 0.8678 0.4604 0.8678 0.9316
No log 32.9231 428 0.8397 0.4175 0.8397 0.9163
No log 33.0769 430 0.8155 0.3814 0.8155 0.9031
No log 33.2308 432 0.8127 0.3814 0.8127 0.9015
No log 33.3846 434 0.8263 0.3814 0.8263 0.9090
No log 33.5385 436 0.8426 0.3714 0.8426 0.9179
No log 33.6923 438 0.8503 0.3714 0.8503 0.9221
No log 33.8462 440 0.8345 0.3723 0.8345 0.9135
No log 34.0 442 0.8011 0.3965 0.8011 0.8951
No log 34.1538 444 0.8063 0.3965 0.8063 0.8980
No log 34.3077 446 0.8592 0.5133 0.8592 0.9269
No log 34.4615 448 0.9390 0.5582 0.9390 0.9690
No log 34.6154 450 0.9988 0.5445 0.9988 0.9994
No log 34.7692 452 1.0456 0.4673 1.0456 1.0226
No log 34.9231 454 1.0224 0.4118 1.0224 1.0111
No log 35.0769 456 1.0237 0.3601 1.0237 1.0118
No log 35.2308 458 0.9907 0.3920 0.9907 0.9954
No log 35.3846 460 0.9787 0.4154 0.9787 0.9893
No log 35.5385 462 0.9686 0.4155 0.9686 0.9842
No log 35.6923 464 0.9498 0.4344 0.9498 0.9746
No log 35.8462 466 0.9587 0.4344 0.9587 0.9791
No log 36.0 468 0.9437 0.3942 0.9437 0.9714
No log 36.1538 470 0.9380 0.3937 0.9380 0.9685
No log 36.3077 472 0.9272 0.3689 0.9272 0.9629
No log 36.4615 474 0.9376 0.3832 0.9376 0.9683
No log 36.6154 476 0.9624 0.3832 0.9624 0.9810
No log 36.7692 478 0.9873 0.3152 0.9873 0.9936
No log 36.9231 480 0.9956 0.3430 0.9956 0.9978
No log 37.0769 482 0.9697 0.3891 0.9697 0.9847
No log 37.2308 484 0.8825 0.4271 0.8825 0.9394
No log 37.3846 486 0.8602 0.3862 0.8602 0.9275
No log 37.5385 488 0.8815 0.3862 0.8815 0.9389
No log 37.6923 490 0.9313 0.3937 0.9313 0.9650
No log 37.8462 492 0.9816 0.3393 0.9816 0.9907
No log 38.0 494 1.0311 0.2870 1.0311 1.0154
No log 38.1538 496 1.0080 0.2870 1.0080 1.0040
No log 38.3077 498 0.9157 0.3937 0.9157 0.9569
0.2428 38.4615 500 0.8429 0.4138 0.8429 0.9181
0.2428 38.6154 502 0.8153 0.4061 0.8153 0.9029
0.2428 38.7692 504 0.8198 0.3960 0.8198 0.9055
0.2428 38.9231 506 0.8260 0.3956 0.8260 0.9089
0.2428 39.0769 508 0.8444 0.3798 0.8444 0.9189
0.2428 39.2308 510 0.8465 0.3798 0.8465 0.9200

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k3_task2_organization

Finetuned
(4019)
this model