ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k6_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8457
  • Qwk: 0.1502
  • Mse: 0.8457
  • Rmse: 0.9196

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0645 2 3.9178 0.0017 3.9178 1.9793
No log 0.1290 4 1.8956 0.0810 1.8956 1.3768
No log 0.1935 6 1.3164 -0.0183 1.3164 1.1474
No log 0.2581 8 1.1543 -0.0648 1.1543 1.0744
No log 0.3226 10 1.1486 0.0006 1.1486 1.0717
No log 0.3871 12 1.0683 0.0067 1.0683 1.0336
No log 0.4516 14 0.9359 -0.0894 0.9359 0.9674
No log 0.5161 16 0.8138 0.0409 0.8138 0.9021
No log 0.5806 18 0.7204 0.0555 0.7204 0.8487
No log 0.6452 20 0.7555 0.0296 0.7555 0.8692
No log 0.7097 22 0.8763 0.0909 0.8763 0.9361
No log 0.7742 24 0.8850 0.0152 0.8850 0.9407
No log 0.8387 26 1.0326 0.0747 1.0326 1.0162
No log 0.9032 28 0.9725 0.0260 0.9725 0.9862
No log 0.9677 30 0.9406 -0.0035 0.9406 0.9699
No log 1.0323 32 0.9742 0.1356 0.9742 0.9870
No log 1.0968 34 1.1858 0.1045 1.1858 1.0890
No log 1.1613 36 0.9011 0.0154 0.9011 0.9493
No log 1.2258 38 0.9093 0.0007 0.9093 0.9536
No log 1.2903 40 0.8418 0.1334 0.8418 0.9175
No log 1.3548 42 0.8470 0.0856 0.8470 0.9203
No log 1.4194 44 0.8453 0.1277 0.8453 0.9194
No log 1.4839 46 0.9575 0.0180 0.9575 0.9785
No log 1.5484 48 1.0712 -0.0023 1.0712 1.0350
No log 1.6129 50 0.9788 0.0233 0.9788 0.9893
No log 1.6774 52 0.7902 0.1007 0.7902 0.8889
No log 1.7419 54 0.9088 -0.0097 0.9088 0.9533
No log 1.8065 56 0.7856 0.0927 0.7856 0.8864
No log 1.8710 58 0.7882 0.1080 0.7882 0.8878
No log 1.9355 60 0.7655 0.1702 0.7655 0.8749
No log 2.0 62 0.9086 0.0644 0.9086 0.9532
No log 2.0645 64 0.9466 0.0519 0.9466 0.9730
No log 2.1290 66 0.7670 0.1863 0.7670 0.8758
No log 2.1935 68 0.7995 0.1929 0.7995 0.8942
No log 2.2581 70 0.8201 0.1929 0.8201 0.9056
No log 2.3226 72 0.7729 0.2743 0.7729 0.8792
No log 2.3871 74 1.1043 0.0735 1.1043 1.0508
No log 2.4516 76 1.3058 0.1555 1.3058 1.1427
No log 2.5161 78 0.9589 0.1548 0.9589 0.9792
No log 2.5806 80 1.3128 0.1103 1.3128 1.1458
No log 2.6452 82 1.3892 -0.0154 1.3892 1.1787
No log 2.7097 84 1.0483 0.0836 1.0483 1.0239
No log 2.7742 86 0.8153 0.0650 0.8153 0.9029
No log 2.8387 88 1.0255 -0.0558 1.0255 1.0126
No log 2.9032 90 0.8324 0.0953 0.8324 0.9123
No log 2.9677 92 0.7232 0.1644 0.7232 0.8504
No log 3.0323 94 0.7486 0.1143 0.7486 0.8652
No log 3.0968 96 0.8376 0.1281 0.8376 0.9152
No log 3.1613 98 0.8475 0.1660 0.8475 0.9206
No log 3.2258 100 0.8584 0.1228 0.8584 0.9265
No log 3.2903 102 0.8239 0.1660 0.8239 0.9077
No log 3.3548 104 0.8710 0.1609 0.8710 0.9333
No log 3.4194 106 0.8421 0.1538 0.8421 0.9177
No log 3.4839 108 1.1024 0.2210 1.1024 1.0499
No log 3.5484 110 1.0162 0.1009 1.0162 1.0081
No log 3.6129 112 0.8759 0.1509 0.8759 0.9359
No log 3.6774 114 0.8667 0.1432 0.8667 0.9310
No log 3.7419 116 0.8664 0.2345 0.8664 0.9308
No log 3.8065 118 0.7647 0.1221 0.7647 0.8744
No log 3.8710 120 0.8128 0.1037 0.8128 0.9015
No log 3.9355 122 0.7733 0.0129 0.7733 0.8794
No log 4.0 124 0.8450 0.0762 0.8450 0.9192
No log 4.0645 126 1.4435 0.0359 1.4435 1.2015
No log 4.1290 128 1.3781 0.0115 1.3781 1.1739
No log 4.1935 130 0.8904 0.0843 0.8904 0.9436
No log 4.2581 132 0.7880 0.0633 0.7880 0.8877
No log 4.3226 134 0.9449 0.1116 0.9449 0.9720
No log 4.3871 136 0.8402 0.1484 0.8402 0.9166
No log 4.4516 138 0.8294 0.0871 0.8294 0.9107
No log 4.5161 140 1.1373 0.0795 1.1373 1.0664
No log 4.5806 142 1.0771 0.1187 1.0771 1.0378
No log 4.6452 144 0.7826 0.0871 0.7826 0.8846
No log 4.7097 146 0.7461 0.0879 0.7461 0.8638
No log 4.7742 148 0.7702 0.1003 0.7702 0.8776
No log 4.8387 150 0.8891 0.0727 0.8891 0.9429
No log 4.9032 152 0.9180 -0.0029 0.9180 0.9581
No log 4.9677 154 0.8546 0.0930 0.8546 0.9244
No log 5.0323 156 0.8196 0.1051 0.8196 0.9053
No log 5.0968 158 0.8109 0.1767 0.8109 0.9005
No log 5.1613 160 0.8946 0.0377 0.8946 0.9458
No log 5.2258 162 0.9482 -0.0194 0.9482 0.9737
No log 5.2903 164 0.7520 0.1047 0.7520 0.8672
No log 5.3548 166 0.6982 0.1576 0.6982 0.8356
No log 5.4194 168 0.7153 0.1080 0.7153 0.8458
No log 5.4839 170 0.7106 0.0869 0.7106 0.8430
No log 5.5484 172 0.7835 0.0175 0.7835 0.8852
No log 5.6129 174 0.7981 -0.0209 0.7981 0.8934
No log 5.6774 176 0.7831 -0.0427 0.7831 0.8849
No log 5.7419 178 0.7944 -0.0387 0.7944 0.8913
No log 5.8065 180 0.7818 -0.0086 0.7818 0.8842
No log 5.8710 182 0.8258 -0.0303 0.8258 0.9088
No log 5.9355 184 0.7771 0.1144 0.7771 0.8815
No log 6.0 186 0.7356 -0.0488 0.7356 0.8577
No log 6.0645 188 0.7589 0.0981 0.7589 0.8711
No log 6.1290 190 0.8451 0.1416 0.8451 0.9193
No log 6.1935 192 0.9106 0.1440 0.9106 0.9543
No log 6.2581 194 0.9965 0.1138 0.9965 0.9982
No log 6.3226 196 1.0527 0.1101 1.0527 1.0260
No log 6.3871 198 0.9626 0.1249 0.9626 0.9811
No log 6.4516 200 0.9427 0.0632 0.9427 0.9709
No log 6.5161 202 0.8449 0.0989 0.8449 0.9192
No log 6.5806 204 0.7709 0.0236 0.7709 0.8780
No log 6.6452 206 0.8214 0.0068 0.8214 0.9063
No log 6.7097 208 0.7510 0.0650 0.7510 0.8666
No log 6.7742 210 0.7587 0.1029 0.7587 0.8710
No log 6.8387 212 0.8631 0.1121 0.8631 0.9290
No log 6.9032 214 0.8417 0.1079 0.8417 0.9174
No log 6.9677 216 0.7762 0.0216 0.7762 0.8810
No log 7.0323 218 0.7678 -0.0230 0.7678 0.8763
No log 7.0968 220 0.7356 0.0783 0.7356 0.8577
No log 7.1613 222 0.7699 -0.0307 0.7699 0.8774
No log 7.2258 224 0.7430 0.0471 0.7430 0.8620
No log 7.2903 226 0.7303 0.0741 0.7303 0.8546
No log 7.3548 228 0.7486 0.0639 0.7486 0.8652
No log 7.4194 230 0.7785 0.0867 0.7785 0.8823
No log 7.4839 232 0.7371 0.0289 0.7371 0.8586
No log 7.5484 234 0.7493 0.2481 0.7493 0.8656
No log 7.6129 236 0.7627 0.2553 0.7627 0.8733
No log 7.6774 238 0.7420 0.1049 0.7420 0.8614
No log 7.7419 240 0.7277 0.2138 0.7277 0.8531
No log 7.8065 242 0.7311 0.1718 0.7311 0.8551
No log 7.8710 244 0.7176 0.1689 0.7176 0.8471
No log 7.9355 246 0.7200 0.1199 0.7200 0.8485
No log 8.0 248 0.7091 0.0776 0.7091 0.8421
No log 8.0645 250 0.7494 0.1028 0.7494 0.8657
No log 8.1290 252 0.8465 0.0285 0.8465 0.9200
No log 8.1935 254 0.8090 0.0110 0.8090 0.8994
No log 8.2581 256 0.8012 0.1529 0.8012 0.8951
No log 8.3226 258 0.8043 -0.0541 0.8043 0.8968
No log 8.3871 260 0.8242 0.0595 0.8242 0.9078
No log 8.4516 262 0.7732 0.0828 0.7732 0.8793
No log 8.5161 264 0.8077 0.0622 0.8077 0.8987
No log 8.5806 266 0.7889 0.1095 0.7889 0.8882
No log 8.6452 268 0.7368 0.1244 0.7368 0.8584
No log 8.7097 270 0.7210 0.1244 0.7210 0.8491
No log 8.7742 272 0.7211 0.1244 0.7211 0.8492
No log 8.8387 274 0.7336 0.0741 0.7336 0.8565
No log 8.9032 276 0.7391 0.1240 0.7391 0.8597
No log 8.9677 278 0.7553 0.0741 0.7553 0.8691
No log 9.0323 280 0.7816 0.1143 0.7816 0.8841
No log 9.0968 282 0.7823 0.0376 0.7823 0.8845
No log 9.1613 284 0.7926 0.0810 0.7926 0.8903
No log 9.2258 286 0.7685 0.1240 0.7685 0.8767
No log 9.2903 288 0.7555 0.0834 0.7555 0.8692
No log 9.3548 290 0.7350 0.0828 0.7350 0.8573
No log 9.4194 292 0.7291 0.0723 0.7291 0.8539
No log 9.4839 294 0.7155 0.0269 0.7155 0.8459
No log 9.5484 296 0.7238 0.0749 0.7238 0.8508
No log 9.6129 298 0.7634 0.1004 0.7634 0.8737
No log 9.6774 300 0.7729 0.0749 0.7729 0.8792
No log 9.7419 302 0.8227 0.1647 0.8227 0.9070
No log 9.8065 304 0.8910 0.1506 0.8910 0.9439
No log 9.8710 306 0.8191 0.0595 0.8191 0.9050
No log 9.9355 308 0.7578 0.0840 0.7578 0.8705
No log 10.0 310 0.7954 0.1495 0.7954 0.8919
No log 10.0645 312 0.8037 0.1001 0.8037 0.8965
No log 10.1290 314 0.7293 0.1146 0.7293 0.8540
No log 10.1935 316 0.7676 -0.0837 0.7676 0.8761
No log 10.2581 318 0.7853 -0.0837 0.7853 0.8862
No log 10.3226 320 0.7257 0.0741 0.7257 0.8519
No log 10.3871 322 0.8285 0.1291 0.8285 0.9102
No log 10.4516 324 0.8180 0.1342 0.8180 0.9044
No log 10.5161 326 0.7648 0.0313 0.7648 0.8746
No log 10.5806 328 0.7880 -0.0798 0.7880 0.8877
No log 10.6452 330 0.7344 0.0376 0.7344 0.8570
No log 10.7097 332 0.7327 0.1095 0.7327 0.8560
No log 10.7742 334 0.7433 0.1146 0.7433 0.8622
No log 10.8387 336 0.7403 -0.0599 0.7403 0.8604
No log 10.9032 338 0.7712 -0.1398 0.7712 0.8782
No log 10.9677 340 0.7854 -0.1823 0.7854 0.8863
No log 11.0323 342 0.7648 -0.0967 0.7648 0.8745
No log 11.0968 344 0.7567 0.0814 0.7567 0.8699
No log 11.1613 346 0.7526 0.0814 0.7526 0.8675
No log 11.2258 348 0.7380 0.0814 0.7380 0.8591
No log 11.2903 350 0.7070 0.0814 0.7070 0.8409
No log 11.3548 352 0.7011 0.0 0.7011 0.8373
No log 11.4194 354 0.7442 -0.1551 0.7442 0.8627
No log 11.4839 356 0.7538 -0.2006 0.7538 0.8682
No log 11.5484 358 0.6918 -0.0062 0.6918 0.8317
No log 11.6129 360 0.7392 0.1449 0.7392 0.8598
No log 11.6774 362 0.8614 0.0786 0.8614 0.9281
No log 11.7419 364 0.8127 0.0826 0.8127 0.9015
No log 11.8065 366 0.7262 0.0289 0.7262 0.8522
No log 11.8710 368 0.7418 0.0884 0.7418 0.8613
No log 11.9355 370 0.7418 0.0884 0.7418 0.8613
No log 12.0 372 0.7177 0.0323 0.7177 0.8472
No log 12.0645 374 0.7548 0.1387 0.7548 0.8688
No log 12.1290 376 0.7749 0.0871 0.7749 0.8803
No log 12.1935 378 0.7443 0.1095 0.7443 0.8628
No log 12.2581 380 0.7367 0.1585 0.7367 0.8583
No log 12.3226 382 0.7544 0.0650 0.7544 0.8686
No log 12.3871 384 0.7509 0.0327 0.7509 0.8666
No log 12.4516 386 0.7470 -0.0079 0.7470 0.8643
No log 12.5161 388 0.7460 0.0732 0.7460 0.8637
No log 12.5806 390 0.7492 -0.0132 0.7492 0.8656
No log 12.6452 392 0.7392 -0.0557 0.7392 0.8598
No log 12.7097 394 0.7632 -0.0406 0.7632 0.8736
No log 12.7742 396 0.7634 -0.0406 0.7634 0.8737
No log 12.8387 398 0.7444 -0.0082 0.7444 0.8628
No log 12.9032 400 0.7668 -0.0138 0.7668 0.8757
No log 12.9677 402 0.8201 0.0205 0.8201 0.9056
No log 13.0323 404 0.8744 0.0409 0.8744 0.9351
No log 13.0968 406 0.8384 0.0095 0.8384 0.9157
No log 13.1613 408 0.7953 0.0660 0.7953 0.8918
No log 13.2258 410 0.7687 -0.0533 0.7687 0.8768
No log 13.2903 412 0.7609 0.0214 0.7609 0.8723
No log 13.3548 414 0.7463 0.0670 0.7463 0.8639
No log 13.4194 416 0.7269 -0.0170 0.7269 0.8526
No log 13.4839 418 0.7207 -0.0612 0.7207 0.8489
No log 13.5484 420 0.6976 0.0807 0.6976 0.8352
No log 13.6129 422 0.6924 0.0260 0.6924 0.8321
No log 13.6774 424 0.7139 0.0318 0.7139 0.8449
No log 13.7419 426 0.7494 -0.0195 0.7494 0.8657
No log 13.8065 428 0.7943 0.0277 0.7943 0.8913
No log 13.8710 430 0.8489 0.0670 0.8489 0.9214
No log 13.9355 432 0.8787 0.0265 0.8787 0.9374
No log 14.0 434 0.8737 0.0265 0.8737 0.9347
No log 14.0645 436 0.8311 -0.0127 0.8311 0.9116
No log 14.1290 438 0.7910 -0.0079 0.7910 0.8894
No log 14.1935 440 0.7444 0.0289 0.7444 0.8628
No log 14.2581 442 0.7535 0.1449 0.7535 0.8680
No log 14.3226 444 0.7517 0.1836 0.7517 0.8670
No log 14.3871 446 0.7165 0.1146 0.7165 0.8465
No log 14.4516 448 0.7126 -0.0976 0.7126 0.8442
No log 14.5161 450 0.7340 -0.0427 0.7340 0.8568
No log 14.5806 452 0.7208 0.0257 0.7208 0.8490
No log 14.6452 454 0.7456 0.1879 0.7456 0.8635
No log 14.7097 456 0.7365 0.1495 0.7365 0.8582
No log 14.7742 458 0.7245 0.0410 0.7245 0.8511
No log 14.8387 460 0.7353 -0.0426 0.7353 0.8575
No log 14.9032 462 0.7320 -0.0976 0.7320 0.8556
No log 14.9677 464 0.7067 -0.0560 0.7067 0.8407
No log 15.0323 466 0.6867 0.0416 0.6867 0.8287
No log 15.0968 468 0.6865 0.1379 0.6865 0.8286
No log 15.1613 470 0.7032 0.1627 0.7032 0.8386
No log 15.2258 472 0.7110 0.2034 0.7110 0.8432
No log 15.2903 474 0.7146 0.1627 0.7146 0.8454
No log 15.3548 476 0.7306 0.0700 0.7306 0.8548
No log 15.4194 478 0.7609 0.0749 0.7609 0.8723
No log 15.4839 480 0.7704 0.0717 0.7704 0.8777
No log 15.5484 482 0.8060 0.0024 0.8060 0.8978
No log 15.6129 484 0.8046 0.0313 0.8046 0.8970
No log 15.6774 486 0.8219 0.1004 0.8219 0.9066
No log 15.7419 488 0.9013 0.0293 0.9013 0.9493
No log 15.8065 490 0.8597 0.1379 0.8597 0.9272
No log 15.8710 492 0.7706 0.0277 0.7706 0.8779
No log 15.9355 494 0.7575 -0.0138 0.7575 0.8703
No log 16.0 496 0.7901 0.1449 0.7901 0.8889
No log 16.0645 498 0.8288 0.1775 0.8288 0.9104
0.2516 16.1290 500 0.8166 0.1775 0.8166 0.9036
0.2516 16.1935 502 0.7547 0.0680 0.7547 0.8687
0.2516 16.2581 504 0.7468 -0.0541 0.7468 0.8642
0.2516 16.3226 506 0.7465 -0.0082 0.7465 0.8640
0.2516 16.3871 508 0.7719 0.1286 0.7719 0.8786
0.2516 16.4516 510 0.8457 0.1502 0.8457 0.9196

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k6_task3_organization

Finetuned
(4019)
this model