ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k6_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6460
  • Qwk: 0.5684
  • Mse: 0.6460
  • Rmse: 0.8038

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0625 2 3.9931 0.0130 3.9931 1.9983
No log 0.125 4 2.1465 0.0175 2.1465 1.4651
No log 0.1875 6 1.5532 -0.0209 1.5532 1.2463
No log 0.25 8 1.4583 -0.0365 1.4583 1.2076
No log 0.3125 10 1.2790 0.0821 1.2790 1.1309
No log 0.375 12 1.0192 0.1810 1.0192 1.0095
No log 0.4375 14 1.1202 0.1821 1.1202 1.0584
No log 0.5 16 1.2112 0.1240 1.2112 1.1005
No log 0.5625 18 1.2789 -0.0311 1.2789 1.1309
No log 0.625 20 1.3459 -0.0148 1.3459 1.1601
No log 0.6875 22 1.2002 0.1658 1.2002 1.0955
No log 0.75 24 1.0701 0.1341 1.0701 1.0344
No log 0.8125 26 1.0356 0.1493 1.0356 1.0177
No log 0.875 28 1.0212 0.1465 1.0212 1.0105
No log 0.9375 30 1.0221 0.1189 1.0221 1.0110
No log 1.0 32 1.0711 0.2171 1.0711 1.0350
No log 1.0625 34 1.0946 0.2465 1.0946 1.0462
No log 1.125 36 1.1472 0.1591 1.1472 1.0711
No log 1.1875 38 1.3028 0.0760 1.3028 1.1414
No log 1.25 40 1.2848 0.1142 1.2848 1.1335
No log 1.3125 42 1.0322 0.2100 1.0322 1.0160
No log 1.375 44 0.9069 0.3435 0.9069 0.9523
No log 1.4375 46 0.9158 0.3145 0.9158 0.9570
No log 1.5 48 0.9162 0.3663 0.9162 0.9572
No log 1.5625 50 0.9694 0.3363 0.9694 0.9846
No log 1.625 52 0.9766 0.3985 0.9766 0.9882
No log 1.6875 54 0.8881 0.3474 0.8881 0.9424
No log 1.75 56 0.8430 0.3704 0.8430 0.9182
No log 1.8125 58 0.8593 0.3765 0.8593 0.9270
No log 1.875 60 1.1360 0.3415 1.1360 1.0658
No log 1.9375 62 1.0510 0.2947 1.0510 1.0252
No log 2.0 64 0.8675 0.3540 0.8675 0.9314
No log 2.0625 66 0.8609 0.3652 0.8609 0.9278
No log 2.125 68 1.0102 0.3724 1.0102 1.0051
No log 2.1875 70 0.9943 0.4417 0.9943 0.9972
No log 2.25 72 0.7775 0.4163 0.7775 0.8818
No log 2.3125 74 0.8088 0.5138 0.8088 0.8993
No log 2.375 76 0.7603 0.5993 0.7603 0.8720
No log 2.4375 78 0.7793 0.4388 0.7793 0.8828
No log 2.5 80 0.7993 0.3642 0.7993 0.8940
No log 2.5625 82 0.7960 0.4594 0.7960 0.8922
No log 2.625 84 0.7702 0.5887 0.7702 0.8776
No log 2.6875 86 0.7699 0.5674 0.7699 0.8775
No log 2.75 88 0.7266 0.5759 0.7266 0.8524
No log 2.8125 90 0.7569 0.5566 0.7569 0.8700
No log 2.875 92 0.6917 0.5905 0.6917 0.8317
No log 2.9375 94 0.7979 0.5898 0.7979 0.8933
No log 3.0 96 0.9445 0.5365 0.9445 0.9718
No log 3.0625 98 0.8590 0.5036 0.8590 0.9268
No log 3.125 100 0.7082 0.5890 0.7082 0.8415
No log 3.1875 102 0.8217 0.5571 0.8217 0.9065
No log 3.25 104 1.1640 0.4285 1.1640 1.0789
No log 3.3125 106 0.9986 0.4773 0.9986 0.9993
No log 3.375 108 0.7277 0.5945 0.7277 0.8531
No log 3.4375 110 0.7439 0.5388 0.7439 0.8625
No log 3.5 112 0.7921 0.5222 0.7921 0.8900
No log 3.5625 114 0.8252 0.5222 0.8252 0.9084
No log 3.625 116 0.8031 0.4870 0.8031 0.8962
No log 3.6875 118 0.7755 0.5124 0.7755 0.8806
No log 3.75 120 0.8440 0.5584 0.8440 0.9187
No log 3.8125 122 0.8165 0.5279 0.8165 0.9036
No log 3.875 124 0.7678 0.5199 0.7678 0.8762
No log 3.9375 126 0.7657 0.5676 0.7657 0.8751
No log 4.0 128 0.7359 0.5260 0.7359 0.8579
No log 4.0625 130 0.7721 0.4777 0.7721 0.8787
No log 4.125 132 0.9788 0.5177 0.9788 0.9893
No log 4.1875 134 0.8718 0.5261 0.8718 0.9337
No log 4.25 136 0.7274 0.5530 0.7274 0.8529
No log 4.3125 138 0.7405 0.5427 0.7405 0.8605
No log 4.375 140 0.7242 0.5425 0.7242 0.8510
No log 4.4375 142 0.7217 0.6057 0.7217 0.8495
No log 4.5 144 0.7166 0.5425 0.7166 0.8466
No log 4.5625 146 0.7120 0.5955 0.7120 0.8438
No log 4.625 148 0.7096 0.6096 0.7096 0.8424
No log 4.6875 150 0.7247 0.5328 0.7247 0.8513
No log 4.75 152 0.8606 0.4831 0.8606 0.9277
No log 4.8125 154 0.7855 0.4789 0.7855 0.8863
No log 4.875 156 0.6849 0.6442 0.6849 0.8276
No log 4.9375 158 0.6974 0.6634 0.6974 0.8351
No log 5.0 160 0.7335 0.5777 0.7335 0.8564
No log 5.0625 162 0.7087 0.6955 0.7087 0.8418
No log 5.125 164 0.7226 0.5501 0.7226 0.8501
No log 5.1875 166 0.9145 0.5041 0.9145 0.9563
No log 5.25 168 0.9091 0.5036 0.9091 0.9535
No log 5.3125 170 0.7672 0.5383 0.7672 0.8759
No log 5.375 172 0.8544 0.5230 0.8544 0.9243
No log 5.4375 174 0.8965 0.4318 0.8965 0.9468
No log 5.5 176 0.8049 0.4310 0.8049 0.8972
No log 5.5625 178 0.7377 0.4692 0.7377 0.8589
No log 5.625 180 0.7559 0.4857 0.7559 0.8694
No log 5.6875 182 0.7178 0.4975 0.7178 0.8472
No log 5.75 184 0.6970 0.5375 0.6970 0.8349
No log 5.8125 186 0.7498 0.5266 0.7498 0.8659
No log 5.875 188 0.7252 0.5948 0.7252 0.8516
No log 5.9375 190 0.7179 0.5774 0.7179 0.8473
No log 6.0 192 0.7018 0.5359 0.7018 0.8377
No log 6.0625 194 0.7059 0.5373 0.7059 0.8402
No log 6.125 196 0.7115 0.5480 0.7115 0.8435
No log 6.1875 198 0.7030 0.6076 0.7030 0.8385
No log 6.25 200 0.7532 0.6118 0.7532 0.8679
No log 6.3125 202 0.8004 0.5875 0.8004 0.8947
No log 6.375 204 0.7449 0.6247 0.7449 0.8631
No log 6.4375 206 0.7249 0.6649 0.7249 0.8514
No log 6.5 208 0.7467 0.4984 0.7467 0.8641
No log 6.5625 210 0.7717 0.4291 0.7717 0.8784
No log 6.625 212 0.7633 0.4640 0.7633 0.8737
No log 6.6875 214 0.7435 0.4540 0.7435 0.8623
No log 6.75 216 0.7728 0.4898 0.7728 0.8791
No log 6.8125 218 0.7735 0.4762 0.7735 0.8795
No log 6.875 220 0.7496 0.4893 0.7496 0.8658
No log 6.9375 222 0.7443 0.5700 0.7443 0.8627
No log 7.0 224 0.7614 0.5220 0.7614 0.8726
No log 7.0625 226 0.8071 0.5416 0.8071 0.8984
No log 7.125 228 0.8081 0.4958 0.8081 0.8990
No log 7.1875 230 0.8036 0.4958 0.8036 0.8964
No log 7.25 232 0.7675 0.4234 0.7675 0.8760
No log 7.3125 234 0.7433 0.4416 0.7433 0.8621
No log 7.375 236 0.7396 0.4416 0.7396 0.8600
No log 7.4375 238 0.7729 0.4227 0.7729 0.8791
No log 7.5 240 0.8505 0.4450 0.8505 0.9222
No log 7.5625 242 0.7906 0.5048 0.7906 0.8892
No log 7.625 244 0.7479 0.5002 0.7479 0.8648
No log 7.6875 246 0.7663 0.4988 0.7663 0.8754
No log 7.75 248 0.7333 0.4659 0.7333 0.8563
No log 7.8125 250 0.7256 0.4953 0.7256 0.8518
No log 7.875 252 0.7318 0.4953 0.7318 0.8555
No log 7.9375 254 0.7349 0.4807 0.7349 0.8573
No log 8.0 256 0.7448 0.4807 0.7448 0.8630
No log 8.0625 258 0.7567 0.4388 0.7567 0.8699
No log 8.125 260 0.7483 0.4643 0.7483 0.8651
No log 8.1875 262 0.7495 0.5117 0.7495 0.8657
No log 8.25 264 0.7316 0.5346 0.7316 0.8554
No log 8.3125 266 0.7125 0.5701 0.7125 0.8441
No log 8.375 268 0.7089 0.5386 0.7089 0.8420
No log 8.4375 270 0.7116 0.5552 0.7116 0.8436
No log 8.5 272 0.7085 0.5891 0.7085 0.8417
No log 8.5625 274 0.7046 0.5199 0.7046 0.8394
No log 8.625 276 0.7147 0.4841 0.7147 0.8454
No log 8.6875 278 0.7121 0.4398 0.7121 0.8438
No log 8.75 280 0.7131 0.4416 0.7131 0.8444
No log 8.8125 282 0.7678 0.4656 0.7678 0.8762
No log 8.875 284 0.7853 0.4686 0.7853 0.8861
No log 8.9375 286 0.7153 0.4707 0.7153 0.8458
No log 9.0 288 0.7828 0.5610 0.7828 0.8847
No log 9.0625 290 0.8754 0.4560 0.8754 0.9356
No log 9.125 292 0.8096 0.5042 0.8096 0.8998
No log 9.1875 294 0.7307 0.5302 0.7307 0.8548
No log 9.25 296 0.7637 0.4841 0.7637 0.8739
No log 9.3125 298 0.8028 0.4489 0.8028 0.8960
No log 9.375 300 0.7711 0.4944 0.7711 0.8781
No log 9.4375 302 0.7531 0.4903 0.7531 0.8678
No log 9.5 304 0.7740 0.4980 0.7740 0.8798
No log 9.5625 306 0.8610 0.4799 0.8610 0.9279
No log 9.625 308 0.8404 0.5137 0.8404 0.9167
No log 9.6875 310 0.7391 0.5577 0.7391 0.8597
No log 9.75 312 0.8315 0.3816 0.8315 0.9119
No log 9.8125 314 0.9836 0.4556 0.9836 0.9917
No log 9.875 316 0.9260 0.4303 0.9260 0.9623
No log 9.9375 318 0.7457 0.4609 0.7457 0.8635
No log 10.0 320 0.7564 0.4847 0.7565 0.8697
No log 10.0625 322 0.9256 0.3648 0.9256 0.9621
No log 10.125 324 0.9908 0.4197 0.9908 0.9954
No log 10.1875 326 0.9223 0.5134 0.9223 0.9604
No log 10.25 328 0.7920 0.4727 0.7920 0.8899
No log 10.3125 330 0.8022 0.4919 0.8022 0.8956
No log 10.375 332 0.8351 0.4951 0.8351 0.9138
No log 10.4375 334 0.8241 0.4402 0.8241 0.9078
No log 10.5 336 0.7544 0.4919 0.7544 0.8685
No log 10.5625 338 0.8001 0.5387 0.8001 0.8945
No log 10.625 340 0.8751 0.5198 0.8751 0.9355
No log 10.6875 342 0.8198 0.5510 0.8198 0.9055
No log 10.75 344 0.7055 0.5707 0.7055 0.8399
No log 10.8125 346 0.6452 0.6307 0.6452 0.8032
No log 10.875 348 0.6512 0.5555 0.6512 0.8070
No log 10.9375 350 0.6524 0.5329 0.6524 0.8077
No log 11.0 352 0.6465 0.6013 0.6465 0.8041
No log 11.0625 354 0.6448 0.6427 0.6448 0.8030
No log 11.125 356 0.6400 0.6616 0.6400 0.8000
No log 11.1875 358 0.6354 0.6699 0.6354 0.7971
No log 11.25 360 0.6744 0.5323 0.6744 0.8212
No log 11.3125 362 0.7246 0.5558 0.7246 0.8513
No log 11.375 364 0.7060 0.5547 0.7060 0.8402
No log 11.4375 366 0.6502 0.5961 0.6502 0.8064
No log 11.5 368 0.6575 0.6649 0.6575 0.8108
No log 11.5625 370 0.6672 0.6282 0.6672 0.8168
No log 11.625 372 0.6878 0.5759 0.6878 0.8294
No log 11.6875 374 0.7441 0.4640 0.7441 0.8626
No log 11.75 376 0.7907 0.4439 0.7907 0.8892
No log 11.8125 378 0.7896 0.4418 0.7896 0.8886
No log 11.875 380 0.7247 0.5343 0.7247 0.8513
No log 11.9375 382 0.6985 0.5288 0.6985 0.8358
No log 12.0 384 0.7416 0.5070 0.7416 0.8612
No log 12.0625 386 0.7613 0.5254 0.7613 0.8725
No log 12.125 388 0.7070 0.4742 0.7070 0.8409
No log 12.1875 390 0.6763 0.5402 0.6763 0.8224
No log 12.25 392 0.7006 0.5441 0.7006 0.8370
No log 12.3125 394 0.7555 0.4913 0.7555 0.8692
No log 12.375 396 0.6982 0.5370 0.6982 0.8356
No log 12.4375 398 0.6512 0.5771 0.6512 0.8070
No log 12.5 400 0.6708 0.5361 0.6708 0.8190
No log 12.5625 402 0.6853 0.5127 0.6853 0.8278
No log 12.625 404 0.6906 0.5018 0.6906 0.8310
No log 12.6875 406 0.6852 0.4764 0.6852 0.8278
No log 12.75 408 0.7125 0.4588 0.7125 0.8441
No log 12.8125 410 0.7140 0.4857 0.7140 0.8450
No log 12.875 412 0.6782 0.5656 0.6782 0.8235
No log 12.9375 414 0.6558 0.5822 0.6558 0.8098
No log 13.0 416 0.6501 0.5597 0.6501 0.8063
No log 13.0625 418 0.6505 0.5635 0.6505 0.8065
No log 13.125 420 0.7145 0.5370 0.7145 0.8453
No log 13.1875 422 0.7918 0.4811 0.7918 0.8899
No log 13.25 424 0.7755 0.4922 0.7755 0.8807
No log 13.3125 426 0.6911 0.5112 0.6911 0.8313
No log 13.375 428 0.6842 0.5786 0.6842 0.8272
No log 13.4375 430 0.7837 0.5572 0.7837 0.8852
No log 13.5 432 0.8054 0.5695 0.8054 0.8975
No log 13.5625 434 0.7545 0.5181 0.7545 0.8686
No log 13.625 436 0.6975 0.5010 0.6975 0.8351
No log 13.6875 438 0.7020 0.5202 0.7020 0.8378
No log 13.75 440 0.7373 0.5230 0.7373 0.8587
No log 13.8125 442 0.7240 0.5204 0.7240 0.8509
No log 13.875 444 0.7063 0.5060 0.7063 0.8404
No log 13.9375 446 0.7310 0.4960 0.7310 0.8550
No log 14.0 448 0.7623 0.5707 0.7623 0.8731
No log 14.0625 450 0.7612 0.5470 0.7612 0.8725
No log 14.125 452 0.7036 0.5844 0.7036 0.8388
No log 14.1875 454 0.6658 0.5724 0.6658 0.8160
No log 14.25 456 0.6614 0.5724 0.6614 0.8133
No log 14.3125 458 0.6662 0.6102 0.6662 0.8162
No log 14.375 460 0.6984 0.5062 0.6984 0.8357
No log 14.4375 462 0.7320 0.5266 0.7320 0.8556
No log 14.5 464 0.7291 0.5675 0.7291 0.8539
No log 14.5625 466 0.6907 0.5516 0.6907 0.8311
No log 14.625 468 0.6557 0.5774 0.6557 0.8097
No log 14.6875 470 0.6327 0.6322 0.6327 0.7954
No log 14.75 472 0.6302 0.5771 0.6302 0.7939
No log 14.8125 474 0.6305 0.5771 0.6305 0.7940
No log 14.875 476 0.6362 0.5536 0.6362 0.7976
No log 14.9375 478 0.6420 0.5536 0.6420 0.8012
No log 15.0 480 0.6362 0.5536 0.6362 0.7977
No log 15.0625 482 0.6342 0.5666 0.6342 0.7963
No log 15.125 484 0.6451 0.5704 0.6451 0.8032
No log 15.1875 486 0.6570 0.5463 0.6570 0.8105
No log 15.25 488 0.6469 0.5329 0.6469 0.8043
No log 15.3125 490 0.6430 0.6118 0.6430 0.8019
No log 15.375 492 0.6414 0.6317 0.6414 0.8009
No log 15.4375 494 0.6406 0.6107 0.6406 0.8004
No log 15.5 496 0.6414 0.6006 0.6414 0.8009
No log 15.5625 498 0.6398 0.6006 0.6398 0.7999
0.2986 15.625 500 0.6381 0.6118 0.6381 0.7988
0.2986 15.6875 502 0.6378 0.5759 0.6378 0.7986
0.2986 15.75 504 0.6306 0.5735 0.6306 0.7941
0.2986 15.8125 506 0.6283 0.5614 0.6283 0.7927
0.2986 15.875 508 0.6459 0.5684 0.6459 0.8037
0.2986 15.9375 510 0.6460 0.5684 0.6460 0.8038

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k6_task5_organization

Finetuned
(4019)
this model