ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k4_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7555
  • Qwk: 0.1518
  • Mse: 0.7555
  • Rmse: 0.8692

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0952 2 3.8621 0.0029 3.8621 1.9652
No log 0.1905 4 1.8680 0.0704 1.8680 1.3667
No log 0.2857 6 0.8596 0.0316 0.8596 0.9271
No log 0.3810 8 0.8057 0.0549 0.8057 0.8976
No log 0.4762 10 1.2310 0.0176 1.2310 1.1095
No log 0.5714 12 0.7359 0.0260 0.7359 0.8579
No log 0.6667 14 0.6846 0.1021 0.6846 0.8274
No log 0.7619 16 0.8373 -0.0833 0.8373 0.9150
No log 0.8571 18 0.7408 -0.0725 0.7408 0.8607
No log 0.9524 20 0.8186 -0.0390 0.8186 0.9048
No log 1.0476 22 0.7394 -0.0739 0.7394 0.8599
No log 1.1429 24 0.7329 0.1582 0.7329 0.8561
No log 1.2381 26 0.7416 0.1082 0.7416 0.8612
No log 1.3333 28 0.7873 0.0562 0.7873 0.8873
No log 1.4286 30 1.3880 0.0584 1.3880 1.1782
No log 1.5238 32 1.1801 0.0366 1.1801 1.0863
No log 1.6190 34 0.7837 0.0449 0.7837 0.8853
No log 1.7143 36 0.9018 -0.1421 0.9018 0.9497
No log 1.8095 38 0.8332 0.0535 0.8332 0.9128
No log 1.9048 40 0.8364 0.1221 0.8364 0.9145
No log 2.0 42 0.8669 0.0656 0.8669 0.9311
No log 2.0952 44 0.8926 0.0377 0.8926 0.9448
No log 2.1905 46 0.8227 0.0656 0.8227 0.9070
No log 2.2857 48 0.8308 0.0656 0.8308 0.9115
No log 2.3810 50 0.8412 0.2121 0.8412 0.9172
No log 2.4762 52 0.8506 0.2955 0.8506 0.9223
No log 2.5714 54 0.9942 0.0747 0.9942 0.9971
No log 2.6667 56 0.8879 0.0649 0.8879 0.9423
No log 2.7619 58 0.7655 0.2063 0.7655 0.8749
No log 2.8571 60 0.7554 0.1674 0.7554 0.8691
No log 2.9524 62 0.8155 0.0068 0.8155 0.9030
No log 3.0476 64 0.8049 0.1358 0.8049 0.8972
No log 3.1429 66 0.9900 0.0147 0.9900 0.9950
No log 3.2381 68 0.8359 0.1872 0.8359 0.9143
No log 3.3333 70 0.9064 -0.0408 0.9064 0.9520
No log 3.4286 72 0.8788 0.1519 0.8788 0.9375
No log 3.5238 74 0.9295 0.1005 0.9295 0.9641
No log 3.6190 76 0.8839 0.0230 0.8839 0.9401
No log 3.7143 78 0.9014 0.1884 0.9014 0.9494
No log 3.8095 80 0.8373 0.1203 0.8373 0.9150
No log 3.9048 82 0.8522 0.0856 0.8522 0.9232
No log 4.0 84 0.8748 0.1309 0.8748 0.9353
No log 4.0952 86 0.8518 0.0962 0.8518 0.9229
No log 4.1905 88 0.8284 0.2041 0.8284 0.9101
No log 4.2857 90 0.9164 0.1001 0.9164 0.9573
No log 4.3810 92 0.9154 0.0999 0.9154 0.9567
No log 4.4762 94 0.9855 0.0452 0.9855 0.9927
No log 4.5714 96 0.9719 0.0113 0.9719 0.9859
No log 4.6667 98 0.8774 0.0998 0.8774 0.9367
No log 4.7619 100 0.8361 0.0211 0.8361 0.9144
No log 4.8571 102 0.7805 0.1463 0.7805 0.8835
No log 4.9524 104 0.8565 0.0068 0.8565 0.9255
No log 5.0476 106 0.7865 0.1705 0.7865 0.8868
No log 5.1429 108 0.8659 0.0681 0.8659 0.9306
No log 5.2381 110 0.8043 0.0996 0.8043 0.8968
No log 5.3333 112 1.0384 0.0426 1.0384 1.0190
No log 5.4286 114 1.2278 -0.0011 1.2278 1.1081
No log 5.5238 116 0.8631 0.2366 0.8631 0.9290
No log 5.6190 118 0.8801 0.1661 0.8801 0.9381
No log 5.7143 120 0.8038 0.0916 0.8038 0.8965
No log 5.8095 122 0.8619 0.0831 0.8619 0.9284
No log 5.9048 124 0.7791 0.1675 0.7791 0.8827
No log 6.0 126 0.8236 0.1243 0.8236 0.9075
No log 6.0952 128 0.7845 0.0913 0.7845 0.8857
No log 6.1905 130 0.7547 0.1254 0.7547 0.8687
No log 6.2857 132 0.7781 0.1659 0.7781 0.8821
No log 6.3810 134 0.8983 0.1188 0.8983 0.9478
No log 6.4762 136 0.8795 0.2319 0.8795 0.9378
No log 6.5714 138 0.9228 0.1339 0.9228 0.9606
No log 6.6667 140 0.9805 0.1009 0.9805 0.9902
No log 6.7619 142 0.7721 0.2374 0.7721 0.8787
No log 6.8571 144 0.9865 0.0402 0.9865 0.9932
No log 6.9524 146 0.8193 0.1392 0.8193 0.9051
No log 7.0476 148 0.6828 0.2477 0.6828 0.8263
No log 7.1429 150 0.7547 0.1123 0.7547 0.8687
No log 7.2381 152 0.6947 0.2877 0.6947 0.8335
No log 7.3333 154 0.7753 0.1862 0.7753 0.8805
No log 7.4286 156 0.9513 0.1111 0.9513 0.9753
No log 7.5238 158 0.8494 0.1551 0.8494 0.9216
No log 7.6190 160 0.6974 0.0821 0.6974 0.8351
No log 7.7143 162 0.7475 0.1778 0.7475 0.8646
No log 7.8095 164 0.7277 0.2588 0.7277 0.8531
No log 7.9048 166 0.9696 0.0794 0.9696 0.9847
No log 8.0 168 0.9379 0.0824 0.9379 0.9684
No log 8.0952 170 0.7392 0.1311 0.7392 0.8598
No log 8.1905 172 0.7161 0.1311 0.7161 0.8462
No log 8.2857 174 0.7042 0.0914 0.7042 0.8392
No log 8.3810 176 0.7734 0.1336 0.7734 0.8795
No log 8.4762 178 0.7640 0.1836 0.7640 0.8741
No log 8.5714 180 0.7069 0.0863 0.7069 0.8408
No log 8.6667 182 0.7524 0.1906 0.7524 0.8674
No log 8.7619 184 0.8026 0.2002 0.8026 0.8959
No log 8.8571 186 0.9744 0.1683 0.9744 0.9871
No log 8.9524 188 1.3316 0.1030 1.3316 1.1539
No log 9.0476 190 1.3262 0.0802 1.3262 1.1516
No log 9.1429 192 1.0495 0.0933 1.0495 1.0244
No log 9.2381 194 0.9973 0.2010 0.9973 0.9987
No log 9.3333 196 0.9063 0.1331 0.9063 0.9520
No log 9.4286 198 0.8172 0.2103 0.8172 0.9040
No log 9.5238 200 0.7605 0.0922 0.7605 0.8721
No log 9.6190 202 0.7775 0.1675 0.7775 0.8818
No log 9.7143 204 0.8071 0.0490 0.8071 0.8984
No log 9.8095 206 0.8185 0.0956 0.8185 0.9047
No log 9.9048 208 0.8703 0.0734 0.8703 0.9329
No log 10.0 210 0.8480 0.1277 0.8480 0.9209
No log 10.0952 212 0.8146 0.1423 0.8146 0.9026
No log 10.1905 214 0.8452 0.1649 0.8452 0.9193
No log 10.2857 216 0.8350 0.1649 0.8350 0.9138
No log 10.3810 218 0.7949 0.1199 0.7949 0.8916
No log 10.4762 220 0.8469 -0.0393 0.8469 0.9203
No log 10.5714 222 0.8223 0.0680 0.8223 0.9068
No log 10.6667 224 0.8619 0.0879 0.8619 0.9284
No log 10.7619 226 1.0360 0.0741 1.0360 1.0178
No log 10.8571 228 0.8999 0.1581 0.8999 0.9487
No log 10.9524 230 0.8389 0.1249 0.8389 0.9159
No log 11.0476 232 0.8471 0.1143 0.8471 0.9204
No log 11.1429 234 0.8543 0.0884 0.8543 0.9243
No log 11.2381 236 0.8014 0.1465 0.8014 0.8952
No log 11.3333 238 0.8184 -0.0228 0.8184 0.9046
No log 11.4286 240 0.8261 0.0248 0.8261 0.9089
No log 11.5238 242 0.7571 0.2046 0.7571 0.8701
No log 11.6190 244 0.9873 0.0852 0.9873 0.9936
No log 11.7143 246 1.0728 0.0707 1.0728 1.0358
No log 11.8095 248 0.8353 0.1475 0.8353 0.9139
No log 11.9048 250 0.7435 0.2138 0.7435 0.8622
No log 12.0 252 0.7851 0.0947 0.7851 0.8860
No log 12.0952 254 0.7562 0.2063 0.7562 0.8696
No log 12.1905 256 0.8033 0.1687 0.8033 0.8963
No log 12.2857 258 0.7948 0.1742 0.7948 0.8915
No log 12.3810 260 0.8013 0.1742 0.8013 0.8952
No log 12.4762 262 0.7652 0.2063 0.7652 0.8748
No log 12.5714 264 0.7956 0.1573 0.7956 0.8920
No log 12.6667 266 0.9143 0.0918 0.9143 0.9562
No log 12.7619 268 0.9371 0.0492 0.9371 0.9680
No log 12.8571 270 0.7885 0.1196 0.7885 0.8880
No log 12.9524 272 0.7740 0.1718 0.7740 0.8797
No log 13.0476 274 0.8026 0.0944 0.8026 0.8959
No log 13.1429 276 0.7741 0.0937 0.7741 0.8798
No log 13.2381 278 0.7550 0.1249 0.7550 0.8689
No log 13.3333 280 0.8191 0.1687 0.8191 0.9050
No log 13.4286 282 0.8016 0.1423 0.8016 0.8953
No log 13.5238 284 0.7883 0.2431 0.7883 0.8879
No log 13.6190 286 0.7496 0.1249 0.7496 0.8658
No log 13.7143 288 0.7492 0.2053 0.7492 0.8655
No log 13.8095 290 0.7900 0.2259 0.7900 0.8888
No log 13.9048 292 0.8114 0.2259 0.8114 0.9008
No log 14.0 294 0.8127 0.2484 0.8127 0.9015
No log 14.0952 296 0.8565 0.2066 0.8565 0.9255
No log 14.1905 298 0.8511 0.2513 0.8511 0.9226
No log 14.2857 300 0.8315 0.1797 0.8315 0.9118
No log 14.3810 302 0.8708 -0.0029 0.8708 0.9332
No log 14.4762 304 0.8627 0.0986 0.8627 0.9288
No log 14.5714 306 0.9047 0.0498 0.9047 0.9512
No log 14.6667 308 0.7889 0.0476 0.7889 0.8882
No log 14.7619 310 0.6798 0.1371 0.6798 0.8245
No log 14.8571 312 0.6973 -0.0541 0.6973 0.8350
No log 14.9524 314 0.7095 0.1807 0.7095 0.8423
No log 15.0476 316 0.8846 -0.0094 0.8846 0.9405
No log 15.1429 318 0.9704 0.0481 0.9704 0.9851
No log 15.2381 320 0.8432 0.0917 0.8432 0.9182
No log 15.3333 322 0.7391 0.1304 0.7391 0.8597
No log 15.4286 324 0.6997 0.0869 0.6997 0.8365
No log 15.5238 326 0.6901 0.0414 0.6901 0.8307
No log 15.6190 328 0.7091 0.1807 0.7091 0.8421
No log 15.7143 330 0.7868 0.0956 0.7868 0.8870
No log 15.8095 332 0.7614 0.1675 0.7614 0.8726
No log 15.9048 334 0.7685 0.1287 0.7685 0.8767
No log 16.0 336 0.8725 0.1001 0.8725 0.9341
No log 16.0952 338 0.8976 0.1003 0.8976 0.9474
No log 16.1905 340 0.7850 0.1244 0.7850 0.8860
No log 16.2857 342 0.9172 0.0951 0.9172 0.9577
No log 16.3810 344 0.9505 0.0556 0.9505 0.9749
No log 16.4762 346 0.7768 0.0909 0.7768 0.8813
No log 16.5714 348 0.6794 0.1828 0.6794 0.8242
No log 16.6667 350 0.6878 0.2258 0.6878 0.8293
No log 16.7619 352 0.7125 0.2258 0.7125 0.8441
No log 16.8571 354 0.7533 0.1096 0.7533 0.8680
No log 16.9524 356 0.7629 0.1196 0.7629 0.8734
No log 17.0476 358 0.7685 0.1249 0.7685 0.8766
No log 17.1429 360 0.7721 0.1249 0.7721 0.8787
No log 17.2381 362 0.7993 0.0175 0.7993 0.8940
No log 17.3333 364 0.8518 0.0755 0.8518 0.9229
No log 17.4286 366 0.7865 0.0600 0.7865 0.8868
No log 17.5238 368 0.7315 0.0732 0.7315 0.8553
No log 17.6190 370 0.7252 0.0732 0.7252 0.8516
No log 17.7143 372 0.7573 0.1146 0.7573 0.8702
No log 17.8095 374 0.7889 0.0562 0.7889 0.8882
No log 17.9048 376 0.8654 0.1026 0.8654 0.9303
No log 18.0 378 0.8354 0.0917 0.8354 0.9140
No log 18.0952 380 0.8134 0.2229 0.8134 0.9019
No log 18.1905 382 0.8083 0.0807 0.8083 0.8991
No log 18.2857 384 0.7869 0.1903 0.7869 0.8871
No log 18.3810 386 0.8459 0.1145 0.8459 0.9198
No log 18.4762 388 0.9633 0.1353 0.9633 0.9815
No log 18.5714 390 0.8999 0.0956 0.8999 0.9486
No log 18.6667 392 0.7557 0.1986 0.7557 0.8693
No log 18.7619 394 0.7564 0.1729 0.7564 0.8697
No log 18.8571 396 0.7945 0.0580 0.7945 0.8914
No log 18.9524 398 0.7609 0.1761 0.7609 0.8723
No log 19.0476 400 0.7742 0.1921 0.7742 0.8799
No log 19.1429 402 0.9589 0.1182 0.9589 0.9792
No log 19.2381 404 0.9990 0.1111 0.9990 0.9995
No log 19.3333 406 0.8984 0.1379 0.8984 0.9478
No log 19.4286 408 0.7873 0.1922 0.7873 0.8873
No log 19.5238 410 0.8128 0.1131 0.8128 0.9015
No log 19.6190 412 0.7862 0.1561 0.7862 0.8867
No log 19.7143 414 0.7866 0.1141 0.7866 0.8869
No log 19.8095 416 0.7937 0.1141 0.7937 0.8909
No log 19.9048 418 0.7964 0.1901 0.7964 0.8924
No log 20.0 420 0.7767 0.1529 0.7767 0.8813
No log 20.0952 422 0.7599 0.0783 0.7599 0.8717
No log 20.1905 424 0.7580 0.1196 0.7580 0.8706
No log 20.2857 426 0.8445 0.0799 0.8445 0.9189
No log 20.3810 428 0.9396 0.1354 0.9396 0.9693
No log 20.4762 430 0.8652 0.0762 0.8652 0.9302
No log 20.5714 432 0.7719 0.1675 0.7719 0.8786
No log 20.6667 434 0.8015 0.0838 0.8015 0.8952
No log 20.7619 436 0.8513 0.1323 0.8513 0.9227
No log 20.8571 438 0.7880 0.1132 0.7880 0.8877
No log 20.9524 440 0.7673 0.1986 0.7673 0.8760
No log 21.0476 442 0.8940 0.1027 0.8940 0.9455
No log 21.1429 444 0.9481 0.1882 0.9481 0.9737
No log 21.2381 446 0.9672 0.1882 0.9672 0.9835
No log 21.3333 448 0.8638 0.1935 0.8638 0.9294
No log 21.4286 450 0.7785 0.1415 0.7785 0.8823
No log 21.5238 452 0.7717 0.0804 0.7717 0.8785
No log 21.6190 454 0.7370 0.1244 0.7370 0.8585
No log 21.7143 456 0.7281 0.1901 0.7281 0.8533
No log 21.8095 458 0.7810 0.2288 0.7810 0.8837
No log 21.9048 460 0.7801 0.2288 0.7801 0.8832
No log 22.0 462 0.7328 0.2053 0.7328 0.8560
No log 22.0952 464 0.7350 0.2053 0.7350 0.8573
No log 22.1905 466 0.7940 0.2194 0.7940 0.8911
No log 22.2857 468 0.8725 0.0769 0.8725 0.9341
No log 22.3810 470 0.9603 0.0519 0.9603 0.9800
No log 22.4762 472 0.8925 0.0734 0.8925 0.9447
No log 22.5714 474 0.7949 0.1456 0.7949 0.8916
No log 22.6667 476 0.8750 0.1893 0.8750 0.9354
No log 22.7619 478 0.8410 0.1281 0.8410 0.9170
No log 22.8571 480 0.7539 0.1139 0.7539 0.8683
No log 22.9524 482 0.7717 0.1901 0.7717 0.8785
No log 23.0476 484 0.8981 0.0692 0.8981 0.9477
No log 23.1429 486 0.8883 0.0692 0.8883 0.9425
No log 23.2381 488 0.7763 0.1783 0.7763 0.8811
No log 23.3333 490 0.6953 0.1807 0.6953 0.8339
No log 23.4286 492 0.7256 0.1080 0.7256 0.8518
No log 23.5238 494 0.7316 0.1080 0.7316 0.8554
No log 23.6190 496 0.7045 0.0479 0.7045 0.8393
No log 23.7143 498 0.6838 0.1371 0.6838 0.8269
0.2266 23.8095 500 0.7542 0.2288 0.7542 0.8685
0.2266 23.9048 502 0.8900 0.0267 0.8900 0.9434
0.2266 24.0 504 0.9062 0.0241 0.9062 0.9520
0.2266 24.0952 506 0.7965 0.2288 0.7965 0.8925
0.2266 24.1905 508 0.7607 0.1901 0.7607 0.8722
0.2266 24.2857 510 0.7555 0.1518 0.7555 0.8692

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
3
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k4_task3_organization

Finetuned
(4019)
this model