ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k7_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8647
  • Qwk: 0.0784
  • Mse: 0.8647
  • Rmse: 0.9299

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0556 2 3.7442 -0.0073 3.7442 1.9350
No log 0.1111 4 2.0569 -0.0066 2.0569 1.4342
No log 0.1667 6 1.8401 0.0235 1.8401 1.3565
No log 0.2222 8 1.5022 0.0 1.5022 1.2256
No log 0.2778 10 0.9637 -0.0269 0.9637 0.9817
No log 0.3333 12 0.7931 -0.0390 0.7931 0.8906
No log 0.3889 14 1.1817 0.0493 1.1817 1.0871
No log 0.4444 16 1.7315 -0.0028 1.7315 1.3158
No log 0.5 18 1.2926 0.0048 1.2926 1.1369
No log 0.5556 20 1.0164 -0.0164 1.0164 1.0081
No log 0.6111 22 0.8020 0.0759 0.8020 0.8955
No log 0.6667 24 0.7969 0.0296 0.7969 0.8927
No log 0.7222 26 0.7963 0.1202 0.7963 0.8923
No log 0.7778 28 0.7699 0.1318 0.7699 0.8774
No log 0.8333 30 0.8554 -0.0801 0.8554 0.9249
No log 0.8889 32 0.7926 -0.0753 0.7926 0.8903
No log 0.9444 34 0.6914 0.0506 0.6914 0.8315
No log 1.0 36 0.6999 0.0460 0.6999 0.8366
No log 1.0556 38 0.9474 0.0282 0.9474 0.9734
No log 1.1111 40 1.0664 -0.0178 1.0664 1.0327
No log 1.1667 42 0.8704 0.0839 0.8704 0.9330
No log 1.2222 44 0.7203 0.0416 0.7203 0.8487
No log 1.2778 46 0.6905 0.0460 0.6905 0.8310
No log 1.3333 48 0.7227 0.0374 0.7227 0.8501
No log 1.3889 50 0.7542 0.0374 0.7542 0.8684
No log 1.4444 52 0.8326 -0.1722 0.8326 0.9125
No log 1.5 54 0.8944 -0.1249 0.8944 0.9457
No log 1.5556 56 0.8591 -0.0252 0.8591 0.9269
No log 1.6111 58 0.8076 0.0807 0.8076 0.8987
No log 1.6667 60 0.7983 0.0909 0.7983 0.8935
No log 1.7222 62 0.8770 0.1701 0.8770 0.9365
No log 1.7778 64 1.0215 0.0217 1.0215 1.0107
No log 1.8333 66 0.9208 0.0362 0.9208 0.9596
No log 1.8889 68 0.8240 0.1199 0.8240 0.9077
No log 1.9444 70 0.7933 0.1249 0.7933 0.8907
No log 2.0 72 0.7508 0.0454 0.7508 0.8665
No log 2.0556 74 0.8299 0.0297 0.8299 0.9110
No log 2.1111 76 0.7392 0.0541 0.7392 0.8598
No log 2.1667 78 0.7626 0.0513 0.7626 0.8733
No log 2.2222 80 0.8143 0.0543 0.8143 0.9024
No log 2.2778 82 0.8630 0.1189 0.8630 0.9290
No log 2.3333 84 0.8913 0.1181 0.8913 0.9441
No log 2.3889 86 0.9735 0.0340 0.9735 0.9867
No log 2.4444 88 0.9254 0.0856 0.9254 0.9620
No log 2.5 90 1.0095 0.0903 1.0095 1.0048
No log 2.5556 92 1.0674 0.0875 1.0674 1.0331
No log 2.6111 94 1.4413 0.0545 1.4413 1.2005
No log 2.6667 96 1.5445 0.0303 1.5445 1.2428
No log 2.7222 98 1.0876 -0.0051 1.0876 1.0429
No log 2.7778 100 0.9662 0.1011 0.9662 0.9829
No log 2.8333 102 0.9797 0.0890 0.9797 0.9898
No log 2.8889 104 1.1713 0.0160 1.1713 1.0823
No log 2.9444 106 1.0820 -0.0608 1.0820 1.0402
No log 3.0 108 0.9693 0.0192 0.9693 0.9845
No log 3.0556 110 0.9116 -0.0409 0.9116 0.9548
No log 3.1111 112 1.0341 0.0367 1.0341 1.0169
No log 3.1667 114 1.1214 0.0719 1.1214 1.0589
No log 3.2222 116 0.8672 0.0573 0.8672 0.9312
No log 3.2778 118 0.8989 -0.0767 0.8989 0.9481
No log 3.3333 120 1.0046 0.0974 1.0046 1.0023
No log 3.3889 122 1.2473 -0.0623 1.2473 1.1168
No log 3.4444 124 1.1525 -0.0423 1.1525 1.0735
No log 3.5 126 0.9296 0.0866 0.9296 0.9641
No log 3.5556 128 0.9418 -0.0355 0.9418 0.9705
No log 3.6111 130 0.8563 0.1282 0.8563 0.9254
No log 3.6667 132 0.9858 -0.0899 0.9858 0.9929
No log 3.7222 134 0.9693 -0.1278 0.9693 0.9845
No log 3.7778 136 0.8186 0.1674 0.8186 0.9048
No log 3.8333 138 0.9110 -0.0336 0.9110 0.9545
No log 3.8889 140 0.9022 0.1267 0.9022 0.9498
No log 3.9444 142 1.2628 0.0296 1.2628 1.1237
No log 4.0 144 1.1959 -0.0058 1.1959 1.0936
No log 4.0556 146 0.9906 -0.0137 0.9906 0.9953
No log 4.1111 148 0.9622 0.1597 0.9622 0.9809
No log 4.1667 150 0.9840 0.0223 0.9840 0.9919
No log 4.2222 152 1.2769 0.0308 1.2769 1.1300
No log 4.2778 154 1.2478 0.0558 1.2478 1.1171
No log 4.3333 156 0.9668 -0.0424 0.9668 0.9832
No log 4.3889 158 0.9025 0.1604 0.9025 0.9500
No log 4.4444 160 0.8793 0.1287 0.8793 0.9377
No log 4.5 162 0.8238 0.1751 0.8238 0.9076
No log 4.5556 164 0.7564 0.1311 0.7564 0.8697
No log 4.6111 166 0.7261 0.1856 0.7261 0.8521
No log 4.6667 168 0.8103 0.0279 0.8103 0.9001
No log 4.7222 170 0.8445 -0.0089 0.8445 0.9190
No log 4.7778 172 0.8242 0.2128 0.8242 0.9078
No log 4.8333 174 0.9899 0.0419 0.9899 0.9949
No log 4.8889 176 1.0420 0.0668 1.0420 1.0208
No log 4.9444 178 0.9636 0.0569 0.9636 0.9816
No log 5.0 180 1.0571 0.0322 1.0571 1.0281
No log 5.0556 182 0.9955 0.0569 0.9955 0.9977
No log 5.1111 184 0.9250 0.1221 0.9250 0.9618
No log 5.1667 186 0.8808 0.1673 0.8808 0.9385
No log 5.2222 188 0.9623 -0.1471 0.9623 0.9810
No log 5.2778 190 1.0402 -0.0504 1.0402 1.0199
No log 5.3333 192 0.9027 -0.0107 0.9027 0.9501
No log 5.3889 194 0.8438 0.1327 0.8438 0.9186
No log 5.4444 196 0.8534 0.1408 0.8534 0.9238
No log 5.5 198 1.1142 -0.0058 1.1142 1.0555
No log 5.5556 200 1.2675 0.0330 1.2675 1.1258
No log 5.6111 202 1.0871 0.0219 1.0871 1.0426
No log 5.6667 204 0.8577 0.1673 0.8577 0.9261
No log 5.7222 206 0.8185 0.1718 0.8185 0.9047
No log 5.7778 208 0.8176 0.1761 0.8176 0.9042
No log 5.8333 210 0.9561 -0.0905 0.9561 0.9778
No log 5.8889 212 0.9801 -0.0198 0.9801 0.9900
No log 5.9444 214 0.8430 0.1379 0.8430 0.9181
No log 6.0 216 0.8427 0.0359 0.8427 0.9180
No log 6.0556 218 0.8516 0.1786 0.8516 0.9228
No log 6.1111 220 1.0862 0.1045 1.0862 1.0422
No log 6.1667 222 1.1385 0.1046 1.1385 1.0670
No log 6.2222 224 0.9877 0.0912 0.9877 0.9938
No log 6.2778 226 0.8648 0.1591 0.8648 0.9299
No log 6.3333 228 0.8699 0.0842 0.8699 0.9327
No log 6.3889 230 0.9948 -0.0504 0.9948 0.9974
No log 6.4444 232 0.9432 0.0366 0.9432 0.9712
No log 6.5 234 0.8548 0.1255 0.8548 0.9246
No log 6.5556 236 0.9164 0.0931 0.9164 0.9573
No log 6.6111 238 0.9547 -0.0033 0.9547 0.9771
No log 6.6667 240 0.9962 -0.0316 0.9962 0.9981
No log 6.7222 242 0.8776 0.0822 0.8776 0.9368
No log 6.7778 244 0.8741 0.1048 0.8741 0.9349
No log 6.8333 246 0.8530 0.0723 0.8530 0.9236
No log 6.8889 248 0.8595 0.0506 0.8595 0.9271
No log 6.9444 250 0.9133 0.0307 0.9133 0.9557
No log 7.0 252 0.8303 0.1277 0.8303 0.9112
No log 7.0556 254 0.8323 0.1277 0.8323 0.9123
No log 7.1111 256 0.8956 -0.0148 0.8956 0.9464
No log 7.1667 258 1.0097 -0.0253 1.0097 1.0048
No log 7.2222 260 0.8919 0.0134 0.8919 0.9444
No log 7.2778 262 0.8649 0.1277 0.8649 0.9300
No log 7.3333 264 0.9357 -0.0438 0.9357 0.9673
No log 7.3889 266 1.1150 -0.0462 1.1150 1.0559
No log 7.4444 268 0.9748 0.0606 0.9748 0.9873
No log 7.5 270 0.8874 0.1635 0.8874 0.9420
No log 7.5556 272 0.8335 0.1277 0.8335 0.9130
No log 7.6111 274 0.8416 0.1310 0.8416 0.9174
No log 7.6667 276 0.9210 -0.0037 0.9210 0.9597
No log 7.7222 278 0.8114 0.0583 0.8114 0.9008
No log 7.7778 280 0.7810 0.1318 0.7810 0.8838
No log 7.8333 282 0.7987 0.1311 0.7987 0.8937
No log 7.8889 284 0.8367 0.1282 0.8367 0.9147
No log 7.9444 286 0.9046 0.0909 0.9046 0.9511
No log 8.0 288 0.9645 -0.0119 0.9645 0.9821
No log 8.0556 290 0.9371 0.1135 0.9371 0.9680
No log 8.1111 292 0.9456 0.1135 0.9456 0.9724
No log 8.1667 294 1.0242 -0.0316 1.0242 1.0120
No log 8.2222 296 1.0909 -0.0170 1.0909 1.0445
No log 8.2778 298 0.9156 -0.0033 0.9156 0.9569
No log 8.3333 300 0.8150 0.1292 0.8150 0.9028
No log 8.3889 302 0.8299 -0.0195 0.8299 0.9110
No log 8.4444 304 0.8329 0.0359 0.8329 0.9126
No log 8.5 306 1.0541 0.0366 1.0541 1.0267
No log 8.5556 308 1.2454 0.0496 1.2454 1.1160
No log 8.6111 310 1.1435 0.0145 1.1435 1.0694
No log 8.6667 312 1.1373 0.0455 1.1373 1.0664
No log 8.7222 314 1.0822 0.0402 1.0822 1.0403
No log 8.7778 316 1.0840 0.0719 1.0840 1.0412
No log 8.8333 318 0.8891 0.0613 0.8891 0.9429
No log 8.8889 320 0.8493 0.1770 0.8493 0.9216
No log 8.9444 322 0.8323 0.1228 0.8323 0.9123
No log 9.0 324 0.8465 0.1228 0.8465 0.9200
No log 9.0556 326 0.8942 0.0265 0.8942 0.9456
No log 9.1111 328 0.8345 0.1352 0.8345 0.9135
No log 9.1667 330 0.7985 0.0791 0.7985 0.8936
No log 9.2222 332 0.7922 0.1236 0.7922 0.8900
No log 9.2778 334 0.8013 0.1425 0.8013 0.8951
No log 9.3333 336 0.8187 0.1425 0.8187 0.9048
No log 9.3889 338 0.8244 0.1372 0.8244 0.9079
No log 9.4444 340 0.9332 0.0326 0.9332 0.9660
No log 9.5 342 0.9210 0.0996 0.9210 0.9597
No log 9.5556 344 0.9030 0.0145 0.9030 0.9503
No log 9.6111 346 0.8987 0.1133 0.8987 0.9480
No log 9.6667 348 0.9234 0.0165 0.9234 0.9609
No log 9.7222 350 1.1845 -0.0077 1.1845 1.0884
No log 9.7778 352 1.2505 0.0534 1.2505 1.1183
No log 9.8333 354 0.9633 0.0687 0.9633 0.9815
No log 9.8889 356 0.8350 0.1049 0.8350 0.9138
No log 9.9444 358 0.8947 -0.0008 0.8947 0.9459
No log 10.0 360 0.7963 0.1202 0.7963 0.8924
No log 10.0556 362 0.7727 0.1906 0.7727 0.8790
No log 10.1111 364 1.0537 -0.0327 1.0537 1.0265
No log 10.1667 366 1.1662 0.0308 1.1662 1.0799
No log 10.2222 368 1.0046 0.0175 1.0046 1.0023
No log 10.2778 370 0.8109 0.0840 0.8109 0.9005
No log 10.3333 372 0.8438 0.0622 0.8438 0.9186
No log 10.3889 374 0.8324 0.0670 0.8324 0.9123
No log 10.4444 376 0.8802 0.1002 0.8802 0.9382
No log 10.5 378 1.0858 0.0190 1.0858 1.0420
No log 10.5556 380 1.0531 0.0190 1.0531 1.0262
No log 10.6111 382 0.9250 0.0680 0.9250 0.9618
No log 10.6667 384 0.7995 0.1660 0.7995 0.8942
No log 10.7222 386 0.7711 0.0323 0.7711 0.8782
No log 10.7778 388 0.7807 0.2153 0.7807 0.8836
No log 10.8333 390 0.8109 0.2194 0.8109 0.9005
No log 10.8889 392 0.7934 0.2666 0.7934 0.8907
No log 10.9444 394 0.7969 0.2594 0.7969 0.8927
No log 11.0 396 0.7675 0.2628 0.7675 0.8760
No log 11.0556 398 0.7891 0.2593 0.7891 0.8883
No log 11.1111 400 0.7714 0.2628 0.7714 0.8783
No log 11.1667 402 0.8060 0.2594 0.8060 0.8978
No log 11.2222 404 0.9107 0.1042 0.9107 0.9543
No log 11.2778 406 0.9107 0.1385 0.9107 0.9543
No log 11.3333 408 0.8437 0.0632 0.8437 0.9185
No log 11.3889 410 0.7866 0.0449 0.7866 0.8869
No log 11.4444 412 0.8045 0.1434 0.8045 0.8969
No log 11.5 414 0.8360 0.1827 0.8360 0.9144
No log 11.5556 416 0.8428 0.1845 0.8428 0.9181
No log 11.6111 418 0.8383 0.1138 0.8383 0.9156
No log 11.6667 420 0.8830 0.0470 0.8830 0.9397
No log 11.7222 422 0.8315 0.1199 0.8315 0.9118
No log 11.7778 424 0.8470 0.1315 0.8470 0.9203
No log 11.8333 426 1.0342 -0.0504 1.0342 1.0170
No log 11.8889 428 1.0058 -0.0204 1.0058 1.0029
No log 11.9444 430 0.8190 0.0545 0.8190 0.9050
No log 12.0 432 0.7762 0.0768 0.7762 0.8810
No log 12.0556 434 0.7829 0.0768 0.7829 0.8848
No log 12.1111 436 0.7877 0.1244 0.7877 0.8875
No log 12.1667 438 0.9314 0.0311 0.9314 0.9651
No log 12.2222 440 1.0088 0.0365 1.0088 1.0044
No log 12.2778 442 0.9099 -0.0295 0.9099 0.9539
No log 12.3333 444 0.8439 0.0393 0.8439 0.9187
No log 12.3889 446 0.8168 0.1244 0.8168 0.9038
No log 12.4444 448 0.7915 0.0776 0.7915 0.8897
No log 12.5 450 0.7798 0.1244 0.7798 0.8831
No log 12.5556 452 0.7886 0.0791 0.7886 0.8880
No log 12.6111 454 0.8832 -0.0099 0.8832 0.9398
No log 12.6667 456 0.9155 0.0340 0.9155 0.9568
No log 12.7222 458 0.8592 0.0955 0.8592 0.9269
No log 12.7778 460 0.8140 0.0791 0.8140 0.9022
No log 12.8333 462 0.8170 0.0269 0.8170 0.9039
No log 12.8889 464 0.7884 0.1236 0.7884 0.8879
No log 12.9444 466 0.8929 0.0029 0.8929 0.9449
No log 13.0 468 0.9144 -0.0306 0.9144 0.9563
No log 13.0556 470 0.8177 0.2036 0.8177 0.9043
No log 13.1111 472 0.7884 0.0791 0.7884 0.8879
No log 13.1667 474 0.7787 0.1244 0.7787 0.8824
No log 13.2222 476 0.7964 0.1744 0.7964 0.8924
No log 13.2778 478 0.8161 -0.0226 0.8161 0.9034
No log 13.3333 480 0.7984 0.0902 0.7984 0.8935
No log 13.3889 482 0.7858 0.1244 0.7858 0.8865
No log 13.4444 484 0.8130 0.0723 0.8130 0.9017
No log 13.5 486 0.8261 0.0611 0.8261 0.9089
No log 13.5556 488 0.8146 0.0376 0.8146 0.9026
No log 13.6111 490 0.8377 0.1339 0.8377 0.9153
No log 13.6667 492 0.7884 0.1786 0.7884 0.8879
No log 13.7222 494 0.7240 0.1304 0.7240 0.8509
No log 13.7778 496 0.7016 0.1311 0.7016 0.8376
No log 13.8333 498 0.7180 0.1024 0.7180 0.8474
0.3253 13.8889 500 0.7629 0.1395 0.7629 0.8735
0.3253 13.9444 502 0.7924 0.0798 0.7924 0.8902
0.3253 14.0 504 0.8634 0.0045 0.8634 0.9292
0.3253 14.0556 506 0.9608 0.0215 0.9608 0.9802
0.3253 14.1111 508 0.9618 0.0238 0.9618 0.9807
0.3253 14.1667 510 0.8647 0.0784 0.8647 0.9299

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k7_task3_organization

Finetuned
(4019)
this model