ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k15_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5895
  • Qwk: 0.3551
  • Mse: 0.5895
  • Rmse: 0.7678

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.04 2 2.5243 -0.0593 2.5243 1.5888
No log 0.08 4 1.1961 0.1265 1.1961 1.0937
No log 0.12 6 0.9756 -0.0970 0.9756 0.9877
No log 0.16 8 0.8497 0.0717 0.8497 0.9218
No log 0.2 10 0.8649 0.1672 0.8649 0.9300
No log 0.24 12 0.7415 0.1236 0.7415 0.8611
No log 0.28 14 0.7796 -0.0027 0.7796 0.8830
No log 0.32 16 0.9014 -0.0426 0.9014 0.9494
No log 0.36 18 1.0370 -0.1111 1.0370 1.0183
No log 0.4 20 0.9503 -0.0354 0.9503 0.9749
No log 0.44 22 0.8827 0.0 0.8827 0.9395
No log 0.48 24 0.8542 0.0937 0.8542 0.9242
No log 0.52 26 0.8324 0.0893 0.8324 0.9124
No log 0.56 28 0.8141 0.0810 0.8141 0.9023
No log 0.6 30 0.8551 0.1815 0.8551 0.9247
No log 0.64 32 1.0405 0.1827 1.0405 1.0201
No log 0.68 34 0.9948 0.1911 0.9948 0.9974
No log 0.72 36 0.8430 0.2227 0.8430 0.9182
No log 0.76 38 0.8279 0.1228 0.8279 0.9099
No log 0.8 40 0.8648 0.1263 0.8648 0.9299
No log 0.84 42 0.8019 0.0408 0.8019 0.8955
No log 0.88 44 0.7794 0.1179 0.7794 0.8828
No log 0.92 46 0.8038 0.0866 0.8037 0.8965
No log 0.96 48 0.9299 0.1637 0.9299 0.9643
No log 1.0 50 1.1134 0.0715 1.1134 1.0552
No log 1.04 52 1.1694 0.0992 1.1694 1.0814
No log 1.08 54 1.0501 0.2202 1.0501 1.0247
No log 1.12 56 0.8452 0.2046 0.8452 0.9193
No log 1.16 58 0.7142 0.1529 0.7142 0.8451
No log 1.2 60 0.7337 0.2857 0.7337 0.8566
No log 1.24 62 0.7347 0.2857 0.7347 0.8572
No log 1.28 64 0.7380 0.1050 0.7380 0.8591
No log 1.32 66 0.7614 -0.0025 0.7614 0.8726
No log 1.3600 68 0.7655 -0.0025 0.7655 0.8749
No log 1.4 70 0.7753 0.0408 0.7753 0.8805
No log 1.44 72 0.7756 0.0347 0.7756 0.8807
No log 1.48 74 0.7885 0.0347 0.7885 0.8880
No log 1.52 76 0.8290 0.1575 0.8290 0.9105
No log 1.56 78 0.7814 0.0778 0.7814 0.8840
No log 1.6 80 0.7513 0.1050 0.7513 0.8668
No log 1.6400 82 0.7481 0.1400 0.7481 0.8650
No log 1.6800 84 0.7485 0.1400 0.7485 0.8652
No log 1.72 86 0.7762 0.3387 0.7762 0.8810
No log 1.76 88 0.7552 0.3123 0.7552 0.8690
No log 1.8 90 0.7835 0.3976 0.7835 0.8852
No log 1.8400 92 0.7857 0.2751 0.7857 0.8864
No log 1.88 94 0.7963 0.2121 0.7963 0.8924
No log 1.92 96 0.8244 0.3228 0.8244 0.9080
No log 1.96 98 0.9572 0.0781 0.9572 0.9784
No log 2.0 100 1.1662 -0.0333 1.1662 1.0799
No log 2.04 102 1.1798 -0.0333 1.1798 1.0862
No log 2.08 104 1.0851 -0.0693 1.0851 1.0417
No log 2.12 106 0.8272 0.2652 0.8272 0.9095
No log 2.16 108 0.7485 0.2392 0.7485 0.8652
No log 2.2 110 0.8120 0.2796 0.8120 0.9011
No log 2.24 112 0.7892 0.2550 0.7892 0.8883
No log 2.2800 114 0.7660 0.2126 0.7660 0.8752
No log 2.32 116 0.7621 0.2475 0.7621 0.8730
No log 2.36 118 0.8249 0.2784 0.8249 0.9082
No log 2.4 120 0.8091 0.3051 0.8091 0.8995
No log 2.44 122 0.7852 0.2591 0.7852 0.8861
No log 2.48 124 0.7981 0.2161 0.7981 0.8934
No log 2.52 126 0.8693 0.3195 0.8693 0.9323
No log 2.56 128 0.9362 0.3194 0.9362 0.9676
No log 2.6 130 0.9056 0.3344 0.9056 0.9516
No log 2.64 132 0.8654 0.3409 0.8654 0.9303
No log 2.68 134 0.8518 0.2899 0.8518 0.9229
No log 2.7200 136 0.8258 0.3314 0.8258 0.9088
No log 2.76 138 0.7714 0.2877 0.7714 0.8783
No log 2.8 140 0.7987 0.3011 0.7987 0.8937
No log 2.84 142 0.9038 0.2899 0.9038 0.9507
No log 2.88 144 0.9332 0.2232 0.9332 0.9660
No log 2.92 146 0.8829 0.3209 0.8829 0.9396
No log 2.96 148 0.8446 0.2893 0.8446 0.9190
No log 3.0 150 0.8024 0.3504 0.8024 0.8958
No log 3.04 152 0.8252 0.3618 0.8252 0.9084
No log 3.08 154 0.8820 0.3482 0.8820 0.9392
No log 3.12 156 0.8644 0.2747 0.8644 0.9297
No log 3.16 158 0.8710 0.3299 0.8710 0.9333
No log 3.2 160 0.9803 0.2886 0.9803 0.9901
No log 3.24 162 0.9953 0.2152 0.9953 0.9977
No log 3.2800 164 0.9214 0.3641 0.9214 0.9599
No log 3.32 166 0.7614 0.3746 0.7614 0.8726
No log 3.36 168 0.7279 0.3628 0.7279 0.8532
No log 3.4 170 0.7142 0.3556 0.7142 0.8451
No log 3.44 172 0.7049 0.3504 0.7049 0.8396
No log 3.48 174 0.8061 0.3183 0.8061 0.8978
No log 3.52 176 0.8756 0.3095 0.8756 0.9357
No log 3.56 178 0.8196 0.3544 0.8196 0.9053
No log 3.6 180 0.7168 0.3649 0.7168 0.8466
No log 3.64 182 0.7218 0.2234 0.7218 0.8496
No log 3.68 184 0.7489 0.1741 0.7489 0.8654
No log 3.7200 186 0.7338 0.1661 0.7338 0.8566
No log 3.76 188 0.7156 0.2279 0.7156 0.8459
No log 3.8 190 0.7184 0.2777 0.7184 0.8476
No log 3.84 192 0.7261 0.2747 0.7261 0.8521
No log 3.88 194 0.7244 0.2747 0.7244 0.8511
No log 3.92 196 0.7393 0.2652 0.7393 0.8598
No log 3.96 198 0.7339 0.2287 0.7339 0.8567
No log 4.0 200 0.7289 0.2505 0.7289 0.8537
No log 4.04 202 0.7361 0.3078 0.7361 0.8580
No log 4.08 204 0.7776 0.3569 0.7776 0.8818
No log 4.12 206 0.7799 0.3471 0.7799 0.8831
No log 4.16 208 0.7782 0.3966 0.7782 0.8822
No log 4.2 210 0.7509 0.2224 0.7509 0.8666
No log 4.24 212 0.7489 0.1592 0.7489 0.8654
No log 4.28 214 0.7336 0.2224 0.7336 0.8565
No log 4.32 216 0.7363 0.3640 0.7363 0.8581
No log 4.36 218 0.7795 0.3866 0.7795 0.8829
No log 4.4 220 0.7570 0.3688 0.7570 0.8700
No log 4.44 222 0.7048 0.2530 0.7048 0.8395
No log 4.48 224 0.6954 0.2622 0.6954 0.8339
No log 4.52 226 0.6930 0.2901 0.6930 0.8325
No log 4.5600 228 0.6886 0.3106 0.6886 0.8298
No log 4.6 230 0.7215 0.3656 0.7215 0.8494
No log 4.64 232 0.7807 0.4315 0.7807 0.8836
No log 4.68 234 0.7863 0.3891 0.7863 0.8867
No log 4.72 236 0.7613 0.3169 0.7613 0.8725
No log 4.76 238 0.6988 0.3099 0.6988 0.8359
No log 4.8 240 0.6681 0.3524 0.6681 0.8173
No log 4.84 242 0.6643 0.3478 0.6643 0.8151
No log 4.88 244 0.6753 0.3160 0.6753 0.8218
No log 4.92 246 0.6881 0.3160 0.6881 0.8295
No log 4.96 248 0.6686 0.3366 0.6686 0.8177
No log 5.0 250 0.6874 0.4451 0.6874 0.8291
No log 5.04 252 0.7115 0.3844 0.7115 0.8435
No log 5.08 254 0.6640 0.4451 0.6640 0.8149
No log 5.12 256 0.6234 0.3224 0.6234 0.7895
No log 5.16 258 0.6163 0.3524 0.6163 0.7850
No log 5.2 260 0.6212 0.3524 0.6212 0.7882
No log 5.24 262 0.6236 0.3551 0.6236 0.7897
No log 5.28 264 0.6238 0.4547 0.6238 0.7898
No log 5.32 266 0.6189 0.4459 0.6189 0.7867
No log 5.36 268 0.6108 0.4726 0.6108 0.7816
No log 5.4 270 0.6002 0.4866 0.6002 0.7747
No log 5.44 272 0.6171 0.4910 0.6171 0.7855
No log 5.48 274 0.6246 0.5352 0.6246 0.7903
No log 5.52 276 0.6147 0.4910 0.6147 0.7840
No log 5.5600 278 0.6315 0.4374 0.6315 0.7946
No log 5.6 280 0.7208 0.4666 0.7208 0.8490
No log 5.64 282 0.8569 0.3889 0.8569 0.9257
No log 5.68 284 0.8477 0.4064 0.8477 0.9207
No log 5.72 286 0.7660 0.4404 0.7660 0.8752
No log 5.76 288 0.7611 0.4723 0.7611 0.8724
No log 5.8 290 0.6968 0.4568 0.6968 0.8347
No log 5.84 292 0.6696 0.4374 0.6696 0.8183
No log 5.88 294 0.6779 0.4448 0.6779 0.8233
No log 5.92 296 0.6703 0.4958 0.6703 0.8187
No log 5.96 298 0.6720 0.3791 0.6720 0.8197
No log 6.0 300 0.6795 0.4125 0.6795 0.8243
No log 6.04 302 0.6912 0.3618 0.6912 0.8314
No log 6.08 304 0.6972 0.3864 0.6972 0.8350
No log 6.12 306 0.7048 0.4484 0.7048 0.8395
No log 6.16 308 0.6900 0.4484 0.6900 0.8307
No log 6.2 310 0.6693 0.4622 0.6693 0.8181
No log 6.24 312 0.6373 0.4337 0.6373 0.7983
No log 6.28 314 0.6324 0.4036 0.6324 0.7953
No log 6.32 316 0.6507 0.3762 0.6507 0.8067
No log 6.36 318 0.6863 0.3763 0.6863 0.8285
No log 6.4 320 0.7327 0.4350 0.7327 0.8560
No log 6.44 322 0.7222 0.3996 0.7222 0.8498
No log 6.48 324 0.7317 0.3996 0.7317 0.8554
No log 6.52 326 0.7999 0.4123 0.7999 0.8944
No log 6.5600 328 0.8802 0.4382 0.8802 0.9382
No log 6.6 330 0.8007 0.4369 0.8007 0.8948
No log 6.64 332 0.6988 0.2780 0.6988 0.8359
No log 6.68 334 0.6942 0.2593 0.6942 0.8332
No log 6.72 336 0.6912 0.2449 0.6912 0.8314
No log 6.76 338 0.7036 0.3891 0.7036 0.8388
No log 6.8 340 0.7651 0.3963 0.7651 0.8747
No log 6.84 342 0.7571 0.3963 0.7571 0.8701
No log 6.88 344 0.6877 0.4157 0.6877 0.8293
No log 6.92 346 0.6515 0.2419 0.6515 0.8071
No log 6.96 348 0.6384 0.2458 0.6384 0.7990
No log 7.0 350 0.6398 0.3450 0.6398 0.7999
No log 7.04 352 0.6230 0.3025 0.6230 0.7893
No log 7.08 354 0.6414 0.4315 0.6414 0.8009
No log 7.12 356 0.6997 0.4606 0.6997 0.8365
No log 7.16 358 0.6605 0.4408 0.6605 0.8127
No log 7.2 360 0.6212 0.4754 0.6212 0.7882
No log 7.24 362 0.6245 0.4970 0.6245 0.7903
No log 7.28 364 0.6216 0.5119 0.6216 0.7884
No log 7.32 366 0.6051 0.4975 0.6051 0.7779
No log 7.36 368 0.6695 0.4972 0.6695 0.8182
No log 7.4 370 0.7964 0.4584 0.7964 0.8924
No log 7.44 372 0.7518 0.4635 0.7518 0.8671
No log 7.48 374 0.6341 0.4672 0.6341 0.7963
No log 7.52 376 0.6049 0.4847 0.6049 0.7778
No log 7.5600 378 0.6875 0.4610 0.6875 0.8292
No log 7.6 380 0.7031 0.4887 0.7031 0.8385
No log 7.64 382 0.6448 0.4644 0.6448 0.8030
No log 7.68 384 0.6043 0.5050 0.6043 0.7774
No log 7.72 386 0.6089 0.3769 0.6089 0.7803
No log 7.76 388 0.6192 0.3724 0.6192 0.7869
No log 7.8 390 0.6175 0.4601 0.6175 0.7858
No log 7.84 392 0.6459 0.4302 0.6459 0.8037
No log 7.88 394 0.7328 0.4829 0.7328 0.8560
No log 7.92 396 0.7938 0.4815 0.7938 0.8909
No log 7.96 398 0.7658 0.4438 0.7658 0.8751
No log 8.0 400 0.6912 0.4212 0.6912 0.8314
No log 8.04 402 0.6248 0.4103 0.6248 0.7904
No log 8.08 404 0.6254 0.4190 0.6254 0.7908
No log 8.12 406 0.6266 0.4190 0.6266 0.7915
No log 8.16 408 0.6268 0.4036 0.6268 0.7917
No log 8.2 410 0.6443 0.4292 0.6443 0.8027
No log 8.24 412 0.7111 0.4294 0.7111 0.8433
No log 8.28 414 0.7627 0.4002 0.7627 0.8733
No log 8.32 416 0.7441 0.4072 0.7441 0.8626
No log 8.36 418 0.6903 0.4315 0.6903 0.8309
No log 8.4 420 0.6273 0.5034 0.6273 0.7920
No log 8.44 422 0.6097 0.4265 0.6097 0.7808
No log 8.48 424 0.6161 0.4295 0.6161 0.7849
No log 8.52 426 0.6232 0.5206 0.6232 0.7895
No log 8.56 428 0.6600 0.5275 0.6600 0.8124
No log 8.6 430 0.6563 0.5148 0.6563 0.8102
No log 8.64 432 0.6194 0.5434 0.6194 0.7870
No log 8.68 434 0.6263 0.5465 0.6263 0.7914
No log 8.72 436 0.7043 0.4492 0.7043 0.8392
No log 8.76 438 0.7803 0.4232 0.7803 0.8833
No log 8.8 440 0.7420 0.4261 0.7420 0.8614
No log 8.84 442 0.6472 0.3726 0.6472 0.8045
No log 8.88 444 0.6292 0.3995 0.6292 0.7932
No log 8.92 446 0.6294 0.4382 0.6294 0.7933
No log 8.96 448 0.6337 0.3598 0.6337 0.7961
No log 9.0 450 0.6641 0.3888 0.6641 0.8149
No log 9.04 452 0.7060 0.4218 0.7060 0.8402
No log 9.08 454 0.7150 0.3914 0.7150 0.8456
No log 9.12 456 0.6875 0.3985 0.6875 0.8291
No log 9.16 458 0.6835 0.4239 0.6835 0.8267
No log 9.2 460 0.6596 0.3417 0.6596 0.8122
No log 9.24 462 0.6540 0.3235 0.6540 0.8087
No log 9.28 464 0.6428 0.3594 0.6428 0.8017
No log 9.32 466 0.6449 0.3942 0.6449 0.8031
No log 9.36 468 0.6423 0.3942 0.6423 0.8015
No log 9.4 470 0.6216 0.3452 0.6216 0.7884
No log 9.44 472 0.5964 0.3738 0.5964 0.7723
No log 9.48 474 0.5871 0.3452 0.5871 0.7662
No log 9.52 476 0.5951 0.4393 0.5951 0.7714
No log 9.56 478 0.6125 0.4542 0.6125 0.7826
No log 9.6 480 0.6123 0.4997 0.6123 0.7825
No log 9.64 482 0.5846 0.4662 0.5846 0.7646
No log 9.68 484 0.5950 0.4970 0.5950 0.7713
No log 9.72 486 0.6364 0.4937 0.6364 0.7977
No log 9.76 488 0.6119 0.4951 0.6119 0.7823
No log 9.8 490 0.5642 0.5107 0.5642 0.7511
No log 9.84 492 0.5698 0.4615 0.5698 0.7549
No log 9.88 494 0.5699 0.5061 0.5699 0.7549
No log 9.92 496 0.5578 0.4637 0.5578 0.7469
No log 9.96 498 0.5830 0.4155 0.5830 0.7635
0.3327 10.0 500 0.6153 0.4282 0.6153 0.7844
0.3327 10.04 502 0.6432 0.4371 0.6432 0.8020
0.3327 10.08 504 0.6357 0.3885 0.6357 0.7973
0.3327 10.12 506 0.5912 0.3931 0.5912 0.7689
0.3327 10.16 508 0.5777 0.3691 0.5777 0.7601
0.3327 10.2 510 0.5772 0.3239 0.5772 0.7598
0.3327 10.24 512 0.5895 0.3551 0.5895 0.7678

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k15_task7_organization

Finetuned
(4019)
this model