ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k7_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8196
  • Qwk: 0.0762
  • Mse: 0.8196
  • Rmse: 0.9053

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0556 2 3.6216 0.0017 3.6216 1.9031
No log 0.1111 4 1.8102 0.0213 1.8102 1.3454
No log 0.1667 6 1.0845 -0.0695 1.0845 1.0414
No log 0.2222 8 0.7672 -0.1230 0.7672 0.8759
No log 0.2778 10 0.7138 0.0555 0.7138 0.8449
No log 0.3333 12 0.7136 0.0555 0.7136 0.8447
No log 0.3889 14 0.8352 0.0711 0.8352 0.9139
No log 0.4444 16 1.0873 -0.0067 1.0873 1.0428
No log 0.5 18 0.7966 0.0953 0.7966 0.8925
No log 0.5556 20 0.7904 0.1449 0.7904 0.8890
No log 0.6111 22 0.7987 0.0999 0.7987 0.8937
No log 0.6667 24 0.7536 -0.0160 0.7536 0.8681
No log 0.7222 26 0.8310 0.1097 0.8310 0.9116
No log 0.7778 28 1.1431 -0.0198 1.1431 1.0692
No log 0.8333 30 0.9159 0.0196 0.9159 0.9570
No log 0.8889 32 0.8703 0.0617 0.8703 0.9329
No log 0.9444 34 0.7991 0.0680 0.7991 0.8939
No log 1.0 36 0.7997 0.0525 0.7997 0.8943
No log 1.0556 38 0.7740 0.0183 0.7740 0.8798
No log 1.1111 40 0.9294 0.1402 0.9294 0.9641
No log 1.1667 42 0.9313 0.1414 0.9313 0.9650
No log 1.2222 44 0.8902 0.0037 0.8902 0.9435
No log 1.2778 46 0.9752 0.1551 0.9752 0.9875
No log 1.3333 48 0.9392 0.0301 0.9392 0.9691
No log 1.3889 50 1.1364 -0.0337 1.1364 1.0660
No log 1.4444 52 1.0098 0.0134 1.0098 1.0049
No log 1.5 54 0.9612 0.1500 0.9612 0.9804
No log 1.5556 56 0.9698 0.0600 0.9698 0.9848
No log 1.6111 58 1.2530 0.0305 1.2530 1.1194
No log 1.6667 60 1.0587 0.1042 1.0587 1.0289
No log 1.7222 62 0.8841 0.0272 0.8841 0.9403
No log 1.7778 64 0.8593 0.0947 0.8593 0.9270
No log 1.8333 66 1.0676 0.1390 1.0676 1.0333
No log 1.8889 68 0.8630 0.2206 0.8630 0.9290
No log 1.9444 70 0.7861 0.0175 0.7861 0.8866
No log 2.0 72 0.8336 0.1315 0.8336 0.9130
No log 2.0556 74 0.8756 0.0919 0.8756 0.9358
No log 2.1111 76 0.8705 0.2041 0.8705 0.9330
No log 2.1667 78 0.9143 0.0875 0.9143 0.9562
No log 2.2222 80 1.0664 -0.0013 1.0664 1.0327
No log 2.2778 82 1.0169 0.0281 1.0169 1.0084
No log 2.3333 84 0.9399 0.0552 0.9399 0.9695
No log 2.3889 86 0.9815 0.0350 0.9815 0.9907
No log 2.4444 88 0.8637 -0.0238 0.8637 0.9294
No log 2.5 90 0.8698 -0.0217 0.8698 0.9326
No log 2.5556 92 0.8541 0.0095 0.8541 0.9242
No log 2.6111 94 1.0694 0.0631 1.0694 1.0341
No log 2.6667 96 0.8419 -0.1126 0.8419 0.9175
No log 2.7222 98 1.0092 0.0748 1.0092 1.0046
No log 2.7778 100 0.9865 0.0714 0.9865 0.9932
No log 2.8333 102 0.9251 0.0888 0.9251 0.9618
No log 2.8889 104 0.9967 0.0352 0.9967 0.9983
No log 2.9444 106 1.0617 0.0687 1.0617 1.0304
No log 3.0 108 1.2013 -0.0318 1.2013 1.0960
No log 3.0556 110 0.9953 0.0975 0.9953 0.9976
No log 3.1111 112 0.9256 0.2375 0.9256 0.9621
No log 3.1667 114 0.9373 0.1685 0.9373 0.9682
No log 3.2222 116 0.9010 0.0804 0.9010 0.9492
No log 3.2778 118 0.8896 0.0846 0.8896 0.9432
No log 3.3333 120 0.8657 -0.0303 0.8657 0.9304
No log 3.3889 122 0.7782 -0.0170 0.7782 0.8822
No log 3.4444 124 0.7489 -0.0179 0.7489 0.8654
No log 3.5 126 0.7525 -0.0204 0.7525 0.8675
No log 3.5556 128 0.7898 -0.0449 0.7898 0.8887
No log 3.6111 130 0.8030 0.0089 0.8030 0.8961
No log 3.6667 132 0.8045 -0.0614 0.8045 0.8970
No log 3.7222 134 0.8089 -0.0583 0.8089 0.8994
No log 3.7778 136 0.8606 0.0606 0.8606 0.9277
No log 3.8333 138 0.8272 0.1030 0.8272 0.9095
No log 3.8889 140 0.8394 -0.0735 0.8394 0.9162
No log 3.9444 142 0.8497 -0.0355 0.8497 0.9218
No log 4.0 144 0.8008 0.0983 0.8008 0.8949
No log 4.0556 146 0.8897 0.1854 0.8897 0.9432
No log 4.1111 148 0.8581 0.0101 0.8581 0.9263
No log 4.1667 150 0.8575 0.2254 0.8575 0.9260
No log 4.2222 152 0.8985 0.0996 0.8985 0.9479
No log 4.2778 154 1.1212 -0.0385 1.1212 1.0589
No log 4.3333 156 0.9851 0.0767 0.9851 0.9925
No log 4.3889 158 0.7835 -0.0550 0.7835 0.8852
No log 4.4444 160 0.7859 0.0338 0.7859 0.8865
No log 4.5 162 0.8135 0.1549 0.8135 0.9020
No log 4.5556 164 0.8652 0.1950 0.8652 0.9302
No log 4.6111 166 0.9058 0.1466 0.9058 0.9517
No log 4.6667 168 0.8327 0.1372 0.8327 0.9125
No log 4.7222 170 0.8379 0.0341 0.8379 0.9153
No log 4.7778 172 0.8154 -0.0163 0.8154 0.9030
No log 4.8333 174 0.8226 -0.0195 0.8226 0.9069
No log 4.8889 176 0.7934 0.0930 0.7934 0.8907
No log 4.9444 178 0.8472 0.1494 0.8472 0.9204
No log 5.0 180 0.8049 0.0503 0.8049 0.8972
No log 5.0556 182 1.0732 -0.0760 1.0732 1.0359
No log 5.1111 184 1.1371 0.0735 1.1371 1.0663
No log 5.1667 186 0.8622 -0.0209 0.8622 0.9285
No log 5.2222 188 0.9286 0.1458 0.9286 0.9637
No log 5.2778 190 0.9119 0.1412 0.9119 0.9549
No log 5.3333 192 0.8430 0.1904 0.8430 0.9181
No log 5.3889 194 0.8305 0.0697 0.8305 0.9113
No log 5.4444 196 0.7964 0.0690 0.7964 0.8924
No log 5.5 198 0.7346 0.0282 0.7346 0.8571
No log 5.5556 200 0.6972 -0.0065 0.6972 0.8350
No log 5.6111 202 0.7125 0.0759 0.7125 0.8441
No log 5.6667 204 0.7295 0.0670 0.7295 0.8541
No log 5.7222 206 0.6886 -0.0062 0.6886 0.8298
No log 5.7778 208 0.7130 -0.0030 0.7130 0.8444
No log 5.8333 210 0.7331 0.0338 0.7331 0.8562
No log 5.8889 212 0.8300 0.0065 0.8300 0.9111
No log 5.9444 214 0.7754 0.0670 0.7754 0.8806
No log 6.0 216 0.7653 0.1292 0.7653 0.8748
No log 6.0556 218 0.7838 0.0289 0.7838 0.8853
No log 6.1111 220 0.7901 0.0709 0.7901 0.8889
No log 6.1667 222 0.7993 0.0583 0.7993 0.8940
No log 6.2222 224 0.7824 0.0058 0.7824 0.8845
No log 6.2778 226 0.7937 0.1006 0.7937 0.8909
No log 6.3333 228 1.0010 0.0783 1.0010 1.0005
No log 6.3889 230 0.8896 0.0684 0.8896 0.9432
No log 6.4444 232 0.8260 0.1358 0.8260 0.9088
No log 6.5 234 1.0866 0.0542 1.0866 1.0424
No log 6.5556 236 1.0371 0.0808 1.0371 1.0184
No log 6.6111 238 0.7748 -0.0406 0.7748 0.8802
No log 6.6667 240 0.7343 0.0374 0.7343 0.8569
No log 6.7222 242 0.7692 0.0152 0.7692 0.8770
No log 6.7778 244 0.7500 -0.0030 0.7500 0.8660
No log 6.8333 246 0.8386 0.1032 0.8386 0.9158
No log 6.8889 248 0.9086 0.1800 0.9086 0.9532
No log 6.9444 250 0.8660 0.1263 0.8660 0.9306
No log 7.0 252 0.8559 0.2036 0.8559 0.9251
No log 7.0556 254 0.8536 0.1977 0.8536 0.9239
No log 7.1111 256 0.8045 0.0495 0.8045 0.8969
No log 7.1667 258 0.7771 -0.0056 0.7771 0.8815
No log 7.2222 260 0.7581 -0.0056 0.7581 0.8707
No log 7.2778 262 0.7643 0.0643 0.7643 0.8743
No log 7.3333 264 0.7934 0.0966 0.7934 0.8907
No log 7.3889 266 0.7704 0.1550 0.7704 0.8777
No log 7.4444 268 0.7656 0.0981 0.7656 0.8750
No log 7.5 270 0.7248 0.1249 0.7248 0.8513
No log 7.5556 272 0.8071 0.1286 0.8071 0.8984
No log 7.6111 274 0.7199 0.1758 0.7199 0.8485
No log 7.6667 276 0.7163 -0.0541 0.7163 0.8464
No log 7.7222 278 0.7252 -0.0062 0.7252 0.8516
No log 7.7778 280 0.7585 0.1585 0.7585 0.8709
No log 7.8333 282 0.7863 0.1983 0.7863 0.8868
No log 7.8889 284 0.8426 0.1498 0.8426 0.9179
No log 7.9444 286 0.8513 0.1604 0.8513 0.9226
No log 8.0 288 0.8460 0.1807 0.8460 0.9198
No log 8.0556 290 0.8162 0.1863 0.8162 0.9035
No log 8.1111 292 0.8247 0.2123 0.8247 0.9082
No log 8.1667 294 0.7990 0.0606 0.7990 0.8938
No log 8.2222 296 0.7306 0.0394 0.7306 0.8548
No log 8.2778 298 0.7386 0.0357 0.7386 0.8594
No log 8.3333 300 0.7307 0.0357 0.7307 0.8548
No log 8.3889 302 0.7367 0.0357 0.7367 0.8583
No log 8.4444 304 0.7683 0.0503 0.7683 0.8765
No log 8.5 306 0.7641 0.0543 0.7641 0.8741
No log 8.5556 308 0.7249 -0.0029 0.7249 0.8514
No log 8.6111 310 0.7077 -0.0059 0.7077 0.8412
No log 8.6667 312 0.7159 0.0355 0.7159 0.8461
No log 8.7222 314 0.7127 0.0395 0.7127 0.8442
No log 8.7778 316 0.7853 0.1611 0.7853 0.8862
No log 8.8333 318 0.8434 0.1536 0.8434 0.9184
No log 8.8889 320 0.7214 0.0058 0.7214 0.8494
No log 8.9444 322 0.7943 0.1342 0.7943 0.8913
No log 9.0 324 0.8206 0.0346 0.8206 0.9059
No log 9.0556 326 0.6945 0.0807 0.6945 0.8334
No log 9.1111 328 0.7137 0.0058 0.7137 0.8448
No log 9.1667 330 0.7333 0.0058 0.7333 0.8563
No log 9.2222 332 0.7166 0.0269 0.7166 0.8465
No log 9.2778 334 0.8273 0.0988 0.8273 0.9095
No log 9.3333 336 0.7451 0.1817 0.7451 0.8632
No log 9.3889 338 0.7103 -0.0446 0.7103 0.8428
No log 9.4444 340 0.7115 -0.0446 0.7115 0.8435
No log 9.5 342 0.6990 0.0759 0.6990 0.8361
No log 9.5556 344 0.7150 0.1553 0.7150 0.8456
No log 9.6111 346 0.6967 0.0918 0.6967 0.8347
No log 9.6667 348 0.7318 0.0058 0.7318 0.8555
No log 9.7222 350 0.7561 0.0058 0.7561 0.8696
No log 9.7778 352 0.7089 0.0503 0.7089 0.8420
No log 9.8333 354 0.8240 0.0831 0.8240 0.9078
No log 9.8889 356 0.9743 0.0547 0.9743 0.9871
No log 9.9444 358 0.8538 0.0876 0.8538 0.9240
No log 10.0 360 0.8051 0.1031 0.8051 0.8973
No log 10.0556 362 0.8573 0.0200 0.8573 0.9259
No log 10.1111 364 0.8102 0.0583 0.8102 0.9001
No log 10.1667 366 0.7327 0.0412 0.7327 0.8560
No log 10.2222 368 0.8109 0.1336 0.8109 0.9005
No log 10.2778 370 0.8054 0.0956 0.8054 0.8975
No log 10.3333 372 0.7578 0.0465 0.7578 0.8705
No log 10.3889 374 0.7985 0.1030 0.7985 0.8936
No log 10.4444 376 0.7566 -0.0029 0.7566 0.8698
No log 10.5 378 0.7759 0.1440 0.7759 0.8808
No log 10.5556 380 0.7565 0.1146 0.7565 0.8698
No log 10.6111 382 0.7533 -0.0469 0.7533 0.8679
No log 10.6667 384 0.8320 0.0673 0.8320 0.9121
No log 10.7222 386 0.7895 0.0598 0.7895 0.8885
No log 10.7778 388 0.7206 -0.0032 0.7206 0.8489
No log 10.8333 390 0.7264 0.1244 0.7264 0.8523
No log 10.8889 392 0.7296 -0.0032 0.7296 0.8541
No log 10.9444 394 0.7576 -0.0469 0.7576 0.8704
No log 11.0 396 0.8419 0.0633 0.8419 0.9175
No log 11.0556 398 0.8735 0.0654 0.8735 0.9346
No log 11.1111 400 0.7916 0.0503 0.7916 0.8897
No log 11.1667 402 0.8169 0.1841 0.8169 0.9038
No log 11.2222 404 0.8104 0.1379 0.8104 0.9002
No log 11.2778 406 0.7526 0.0432 0.7526 0.8675
No log 11.3333 408 0.8046 0.0110 0.8046 0.8970
No log 11.3889 410 0.7835 0.0557 0.7835 0.8851
No log 11.4444 412 0.7457 0.0432 0.7457 0.8635
No log 11.5 414 0.8360 0.2181 0.8360 0.9144
No log 11.5556 416 0.9083 0.0596 0.9083 0.9531
No log 11.6111 418 0.8319 0.0888 0.8319 0.9121
No log 11.6667 420 0.8028 0.2103 0.8028 0.8960
No log 11.7222 422 0.8881 0.0957 0.8881 0.9424
No log 11.7778 424 0.8551 0.0955 0.8551 0.9247
No log 11.8333 426 0.7860 0.1236 0.7860 0.8866
No log 11.8889 428 0.7992 0.1050 0.7992 0.8940
No log 11.9444 430 0.7977 0.1187 0.7977 0.8932
No log 12.0 432 0.8233 0.2041 0.8233 0.9073
No log 12.0556 434 0.8102 0.1277 0.8102 0.9001
No log 12.1111 436 0.7586 0.1189 0.7586 0.8710
No log 12.1667 438 0.7316 -0.0059 0.7316 0.8553
No log 12.2222 440 0.7347 -0.0030 0.7347 0.8572
No log 12.2778 442 0.7444 -0.0513 0.7444 0.8628
No log 12.3333 444 0.7535 0.0828 0.7535 0.8680
No log 12.3889 446 0.7691 -0.0113 0.7691 0.8770
No log 12.4444 448 0.7652 -0.0469 0.7652 0.8748
No log 12.5 450 0.7530 -0.0469 0.7530 0.8678
No log 12.5556 452 0.7364 -0.0513 0.7364 0.8581
No log 12.6111 454 0.7396 0.1254 0.7396 0.8600
No log 12.6667 456 0.7499 0.0723 0.7499 0.8660
No log 12.7222 458 0.7308 -0.0118 0.7308 0.8549
No log 12.7778 460 0.7608 0.0085 0.7608 0.8722
No log 12.8333 462 0.7673 0.0570 0.7673 0.8760
No log 12.8889 464 0.7461 0.0412 0.7461 0.8638
No log 12.9444 466 0.7498 0.0840 0.7498 0.8659
No log 13.0 468 0.7636 0.0 0.7636 0.8738
No log 13.0556 470 0.7478 -0.0550 0.7478 0.8648
No log 13.1111 472 0.7777 -0.0186 0.7777 0.8819
No log 13.1667 474 0.7878 -0.0209 0.7878 0.8876
No log 13.2222 476 0.8038 0.0522 0.8038 0.8965
No log 13.2778 478 0.8416 0.0623 0.8416 0.9174
No log 13.3333 480 0.7873 0.0840 0.7873 0.8873
No log 13.3889 482 0.9142 0.0250 0.9142 0.9561
No log 13.4444 484 1.0589 0.1414 1.0589 1.0290
No log 13.5 486 0.9447 0.0175 0.9447 0.9719
No log 13.5556 488 0.7835 0.0670 0.7835 0.8851
No log 13.6111 490 0.8595 0.1419 0.8595 0.9271
No log 13.6667 492 0.9423 0.1471 0.9423 0.9707
No log 13.7222 494 0.8183 0.1127 0.8183 0.9046
No log 13.7778 496 0.7008 -0.0513 0.7008 0.8371
No log 13.8333 498 0.7111 0.1444 0.7111 0.8432
0.2478 13.8889 500 0.7312 0.1196 0.7312 0.8551
0.2478 13.9444 502 0.7403 0.0323 0.7403 0.8604
0.2478 14.0 504 0.7573 0.0341 0.7573 0.8702
0.2478 14.0556 506 0.7800 0.0341 0.7800 0.8832
0.2478 14.1111 508 0.8193 0.0377 0.8193 0.9052
0.2478 14.1667 510 0.8196 0.0762 0.8196 0.9053

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k7_task3_organization

Finetuned
(4019)
this model