ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k5_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8475
  • Qwk: 0.6074
  • Mse: 0.8475
  • Rmse: 0.9206

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0870 2 6.7650 0.0 6.7650 2.6010
No log 0.1739 4 4.1813 0.0779 4.1813 2.0448
No log 0.2609 6 3.1679 0.0124 3.1679 1.7799
No log 0.3478 8 2.2355 0.1045 2.2355 1.4952
No log 0.4348 10 1.7567 0.2342 1.7567 1.3254
No log 0.5217 12 1.6618 0.1835 1.6618 1.2891
No log 0.6087 14 1.7238 0.1121 1.7238 1.3129
No log 0.6957 16 1.7642 0.1333 1.7642 1.3282
No log 0.7826 18 2.0496 0.0870 2.0496 1.4316
No log 0.8696 20 2.0483 0.1653 2.0483 1.4312
No log 0.9565 22 2.1151 0.1231 2.1151 1.4543
No log 1.0435 24 1.9973 0.2185 1.9973 1.4133
No log 1.1304 26 1.8770 0.2523 1.8770 1.3700
No log 1.2174 28 1.9424 0.2521 1.9424 1.3937
No log 1.3043 30 2.0452 0.1860 2.0452 1.4301
No log 1.3913 32 2.3021 0.1418 2.3021 1.5173
No log 1.4783 34 2.5694 0.1103 2.5694 1.6029
No log 1.5652 36 2.4455 0.1259 2.4455 1.5638
No log 1.6522 38 2.7063 0.1625 2.7063 1.6451
No log 1.7391 40 2.9047 0.1091 2.9047 1.7043
No log 1.8261 42 2.1700 0.2041 2.1700 1.4731
No log 1.9130 44 1.2497 0.4839 1.2497 1.1179
No log 2.0 46 1.3586 0.3866 1.3586 1.1656
No log 2.0870 48 1.1978 0.4463 1.1978 1.0944
No log 2.1739 50 1.2681 0.4409 1.2681 1.1261
No log 2.2609 52 2.0429 0.3684 2.0429 1.4293
No log 2.3478 54 3.0325 0.1693 3.0325 1.7414
No log 2.4348 56 3.1003 0.1474 3.1003 1.7608
No log 2.5217 58 2.3569 0.2312 2.3569 1.5352
No log 2.6087 60 1.5743 0.4412 1.5743 1.2547
No log 2.6957 62 1.2031 0.4715 1.2031 1.0969
No log 2.7826 64 1.1924 0.5 1.1924 1.0920
No log 2.8696 66 1.3227 0.5197 1.3227 1.1501
No log 2.9565 68 1.6020 0.3971 1.6020 1.2657
No log 3.0435 70 1.8323 0.2639 1.8323 1.3536
No log 3.1304 72 1.6261 0.4225 1.6261 1.2752
No log 3.2174 74 1.1227 0.6377 1.1227 1.0596
No log 3.3043 76 0.8691 0.6763 0.8691 0.9323
No log 3.3913 78 0.9102 0.6176 0.9102 0.9540
No log 3.4783 80 1.0089 0.6087 1.0089 1.0045
No log 3.5652 82 1.0343 0.65 1.0343 1.0170
No log 3.6522 84 1.1621 0.6 1.1621 1.0780
No log 3.7391 86 1.1706 0.5868 1.1706 1.0819
No log 3.8261 88 1.1902 0.5890 1.1902 1.0910
No log 3.9130 90 1.1113 0.6122 1.1113 1.0542
No log 4.0 92 1.1448 0.5077 1.1448 1.0700
No log 4.0870 94 1.0838 0.5606 1.0838 1.0410
No log 4.1739 96 1.0788 0.6232 1.0788 1.0386
No log 4.2609 98 1.4993 0.4737 1.4993 1.2244
No log 4.3478 100 1.8104 0.4432 1.8104 1.3455
No log 4.4348 102 1.5464 0.5581 1.5464 1.2436
No log 4.5217 104 0.8606 0.6892 0.8606 0.9277
No log 4.6087 106 0.7387 0.7092 0.7387 0.8595
No log 4.6957 108 0.7467 0.6618 0.7467 0.8641
No log 4.7826 110 0.7040 0.7324 0.7040 0.8390
No log 4.8696 112 0.7647 0.6809 0.7647 0.8745
No log 4.9565 114 0.7713 0.6944 0.7713 0.8783
No log 5.0435 116 0.7685 0.7034 0.7685 0.8767
No log 5.1304 118 0.6507 0.75 0.6507 0.8067
No log 5.2174 120 0.6686 0.7273 0.6686 0.8177
No log 5.3043 122 0.6561 0.7383 0.6561 0.8100
No log 5.3913 124 0.7323 0.7296 0.7323 0.8558
No log 5.4783 126 0.9327 0.7176 0.9327 0.9657
No log 5.5652 128 1.2376 0.5989 1.2376 1.1125
No log 5.6522 130 1.1405 0.6369 1.1405 1.0680
No log 5.7391 132 0.8716 0.6883 0.8716 0.9336
No log 5.8261 134 0.8590 0.6857 0.8590 0.9268
No log 5.9130 136 0.8759 0.6857 0.8759 0.9359
No log 6.0 138 0.9554 0.6479 0.9554 0.9775
No log 6.0870 140 1.5033 0.5667 1.5033 1.2261
No log 6.1739 142 1.8687 0.4742 1.8687 1.3670
No log 6.2609 144 1.7179 0.5236 1.7179 1.3107
No log 6.3478 146 1.1434 0.6747 1.1434 1.0693
No log 6.4348 148 0.7913 0.7133 0.7913 0.8895
No log 6.5217 150 0.7950 0.7123 0.7950 0.8916
No log 6.6087 152 0.7449 0.7222 0.7449 0.8631
No log 6.6957 154 0.7286 0.7237 0.7286 0.8536
No log 6.7826 156 0.9499 0.6951 0.9499 0.9746
No log 6.8696 158 1.1011 0.6310 1.1011 1.0493
No log 6.9565 160 0.9500 0.6879 0.9500 0.9747
No log 7.0435 162 0.7621 0.7383 0.7621 0.8730
No log 7.1304 164 0.7314 0.7310 0.7314 0.8552
No log 7.2174 166 0.7484 0.7361 0.7484 0.8651
No log 7.3043 168 0.7171 0.7397 0.7171 0.8468
No log 7.3913 170 0.6751 0.7517 0.6751 0.8217
No log 7.4783 172 0.6906 0.7467 0.6906 0.8310
No log 7.5652 174 0.7159 0.7347 0.7159 0.8461
No log 7.6522 176 0.7414 0.7397 0.7414 0.8611
No log 7.7391 178 0.8189 0.6928 0.8189 0.9049
No log 7.8261 180 0.8502 0.6753 0.8502 0.9221
No log 7.9130 182 0.9208 0.6460 0.9208 0.9596
No log 8.0 184 0.8986 0.6582 0.8986 0.9480
No log 8.0870 186 0.8446 0.6711 0.8446 0.9190
No log 8.1739 188 0.8065 0.6429 0.8065 0.8980
No log 8.2609 190 0.7771 0.6377 0.7771 0.8815
No log 8.3478 192 0.7258 0.7183 0.7258 0.8519
No log 8.4348 194 0.6998 0.6906 0.6998 0.8366
No log 8.5217 196 0.6813 0.7273 0.6813 0.8254
No log 8.6087 198 0.6806 0.7248 0.6806 0.8250
No log 8.6957 200 0.7635 0.6933 0.7635 0.8738
No log 8.7826 202 0.9094 0.6887 0.9094 0.9536
No log 8.8696 204 0.9531 0.6711 0.9531 0.9763
No log 8.9565 206 0.9506 0.6711 0.9506 0.9750
No log 9.0435 208 0.9751 0.6887 0.9751 0.9875
No log 9.1304 210 0.8793 0.6887 0.8793 0.9377
No log 9.2174 212 0.7340 0.7152 0.7340 0.8567
No log 9.3043 214 0.6656 0.7297 0.6656 0.8158
No log 9.3913 216 0.6350 0.7383 0.6350 0.7969
No log 9.4783 218 0.6378 0.7083 0.6378 0.7986
No log 9.5652 220 0.6684 0.7123 0.6684 0.8175
No log 9.6522 222 0.6420 0.7133 0.6420 0.8012
No log 9.7391 224 0.6199 0.7397 0.6199 0.7873
No log 9.8261 226 0.6187 0.7448 0.6187 0.7866
No log 9.9130 228 0.6273 0.7671 0.6273 0.7920
No log 10.0 230 0.6334 0.7260 0.6334 0.7959
No log 10.0870 232 0.6453 0.7133 0.6453 0.8033
No log 10.1739 234 0.6563 0.7448 0.6563 0.8101
No log 10.2609 236 0.7059 0.7517 0.7059 0.8402
No log 10.3478 238 0.6949 0.7483 0.6949 0.8336
No log 10.4348 240 0.6380 0.7347 0.6380 0.7988
No log 10.5217 242 0.6644 0.7172 0.6644 0.8151
No log 10.6087 244 0.7013 0.7092 0.7013 0.8374
No log 10.6957 246 0.7063 0.7133 0.7063 0.8404
No log 10.7826 248 0.7189 0.7211 0.7189 0.8479
No log 10.8696 250 0.7688 0.7347 0.7688 0.8768
No log 10.9565 252 0.8090 0.6712 0.8090 0.8994
No log 11.0435 254 0.8514 0.6294 0.8514 0.9227
No log 11.1304 256 0.8124 0.6939 0.8124 0.9014
No log 11.2174 258 0.7766 0.7162 0.7766 0.8813
No log 11.3043 260 0.7542 0.7260 0.7542 0.8684
No log 11.3913 262 0.7801 0.7075 0.7801 0.8832
No log 11.4783 264 0.8103 0.6849 0.8103 0.9002
No log 11.5652 266 0.8197 0.6761 0.8197 0.9054
No log 11.6522 268 0.8439 0.6471 0.8439 0.9186
No log 11.7391 270 0.8183 0.6853 0.8183 0.9046
No log 11.8261 272 0.7544 0.7172 0.7544 0.8686
No log 11.9130 274 0.7147 0.7632 0.7147 0.8454
No log 12.0 276 0.7326 0.7742 0.7326 0.8559
No log 12.0870 278 0.8298 0.6974 0.8298 0.9110
No log 12.1739 280 0.9080 0.6790 0.9080 0.9529
No log 12.2609 282 0.8577 0.6923 0.8577 0.9261
No log 12.3478 284 0.7662 0.7436 0.7662 0.8753
No log 12.4348 286 0.7085 0.7742 0.7085 0.8417
No log 12.5217 288 0.7166 0.7742 0.7166 0.8465
No log 12.6087 290 0.7500 0.7742 0.7500 0.8660
No log 12.6957 292 0.7760 0.7333 0.7760 0.8809
No log 12.7826 294 0.8389 0.7134 0.8389 0.9159
No log 12.8696 296 0.8839 0.6962 0.8839 0.9402
No log 12.9565 298 0.8038 0.7237 0.8038 0.8966
No log 13.0435 300 0.7541 0.7568 0.7541 0.8684
No log 13.1304 302 0.7425 0.7467 0.7425 0.8617
No log 13.2174 304 0.8579 0.7097 0.8579 0.9262
No log 13.3043 306 1.0625 0.6509 1.0625 1.0308
No log 13.3913 308 1.0485 0.6509 1.0485 1.0240
No log 13.4783 310 0.8158 0.7170 0.8158 0.9032
No log 13.5652 312 0.6532 0.76 0.6532 0.8082
No log 13.6522 314 0.6457 0.7222 0.6457 0.8036
No log 13.7391 316 0.6604 0.7183 0.6604 0.8126
No log 13.8261 318 0.7060 0.7260 0.7060 0.8403
No log 13.9130 320 0.7773 0.7248 0.7773 0.8816
No log 14.0 322 0.8063 0.6897 0.8063 0.8979
No log 14.0870 324 0.8335 0.6803 0.8335 0.9130
No log 14.1739 326 0.7563 0.7034 0.7563 0.8697
No log 14.2609 328 0.7059 0.7183 0.7059 0.8402
No log 14.3478 330 0.7190 0.7042 0.7190 0.8479
No log 14.4348 332 0.7113 0.7123 0.7113 0.8434
No log 14.5217 334 0.7626 0.7625 0.7626 0.8733
No log 14.6087 336 0.8560 0.7073 0.8560 0.9252
No log 14.6957 338 1.1169 0.6746 1.1169 1.0568
No log 14.7826 340 1.1370 0.6548 1.1370 1.0663
No log 14.8696 342 0.8650 0.7160 0.8650 0.9301
No log 14.9565 344 0.7139 0.7875 0.7139 0.8449
No log 15.0435 346 0.6808 0.7662 0.6808 0.8251
No log 15.1304 348 0.7291 0.75 0.7291 0.8539
No log 15.2174 350 0.7982 0.7190 0.7982 0.8934
No log 15.3043 352 0.8040 0.7190 0.8040 0.8966
No log 15.3913 354 0.7417 0.7347 0.7417 0.8612
No log 15.4783 356 0.7204 0.7324 0.7204 0.8488
No log 15.5652 358 0.7197 0.7324 0.7197 0.8484
No log 15.6522 360 0.7392 0.7222 0.7392 0.8598
No log 15.7391 362 0.7885 0.6294 0.7885 0.8880
No log 15.8261 364 0.7984 0.6294 0.7984 0.8935
No log 15.9130 366 0.8001 0.5839 0.8001 0.8945
No log 16.0 368 0.8229 0.6418 0.8229 0.9071
No log 16.0870 370 0.8908 0.5397 0.8908 0.9438
No log 16.1739 372 0.9094 0.5354 0.9094 0.9536
No log 16.2609 374 0.8504 0.5846 0.8504 0.9222
No log 16.3478 376 0.7952 0.6418 0.7952 0.8918
No log 16.4348 378 0.7301 0.7361 0.7301 0.8544
No log 16.5217 380 0.7133 0.6944 0.7133 0.8446
No log 16.6087 382 0.7119 0.7075 0.7119 0.8438
No log 16.6957 384 0.7692 0.6939 0.7692 0.8770
No log 16.7826 386 0.7595 0.6573 0.7595 0.8715
No log 16.8696 388 0.7479 0.7 0.7479 0.8648
No log 16.9565 390 0.8014 0.6569 0.8014 0.8952
No log 17.0435 392 0.8446 0.5970 0.8446 0.9190
No log 17.1304 394 0.7946 0.6618 0.7946 0.8914
No log 17.2174 396 0.8151 0.6087 0.8151 0.9028
No log 17.3043 398 0.8461 0.6131 0.8461 0.9198
No log 17.3913 400 0.8184 0.6277 0.8184 0.9046
No log 17.4783 402 0.7882 0.6316 0.7882 0.8878
No log 17.5652 404 0.8153 0.5891 0.8153 0.9030
No log 17.6522 406 0.8530 0.5649 0.8530 0.9236
No log 17.7391 408 0.8673 0.6131 0.8673 0.9313
No log 17.8261 410 0.8610 0.6286 0.8610 0.9279
No log 17.9130 412 0.7903 0.6479 0.7903 0.8890
No log 18.0 414 0.7302 0.6857 0.7302 0.8545
No log 18.0870 416 0.7768 0.6277 0.7768 0.8814
No log 18.1739 418 0.8729 0.6377 0.8729 0.9343
No log 18.2609 420 0.9138 0.5926 0.9138 0.9559
No log 18.3478 422 0.8938 0.5821 0.8938 0.9454
No log 18.4348 424 0.8277 0.5985 0.8277 0.9098
No log 18.5217 426 0.7716 0.6286 0.7716 0.8784
No log 18.6087 428 0.7973 0.6849 0.7973 0.8929
No log 18.6957 430 0.8693 0.6533 0.8693 0.9324
No log 18.7826 432 0.8871 0.6622 0.8871 0.9419
No log 18.8696 434 0.8451 0.6241 0.8451 0.9193
No log 18.9565 436 0.8261 0.6286 0.8261 0.9089
No log 19.0435 438 0.8008 0.6620 0.8008 0.8949
No log 19.1304 440 0.7924 0.6479 0.7924 0.8901
No log 19.2174 442 0.8275 0.6043 0.8275 0.9097
No log 19.3043 444 0.8342 0.6338 0.8342 0.9134
No log 19.3913 446 0.8369 0.6434 0.8369 0.9148
No log 19.4783 448 0.8458 0.6759 0.8458 0.9197
No log 19.5652 450 0.8881 0.6623 0.8881 0.9424
No log 19.6522 452 0.8954 0.6923 0.8954 0.9463
No log 19.7391 454 0.8901 0.7134 0.8901 0.9435
No log 19.8261 456 0.8623 0.7067 0.8623 0.9286
No log 19.9130 458 0.8162 0.6897 0.8162 0.9034
No log 20.0 460 0.8129 0.6429 0.8129 0.9016
No log 20.0870 462 0.8267 0.6222 0.8267 0.9093
No log 20.1739 464 0.8353 0.6119 0.8353 0.9139
No log 20.2609 466 0.8550 0.6471 0.8550 0.9247
No log 20.3478 468 0.8563 0.6429 0.8563 0.9254
No log 20.4348 470 0.8505 0.6892 0.8505 0.9222
No log 20.5217 472 0.8110 0.6892 0.8110 0.9006
No log 20.6087 474 0.7384 0.6713 0.7384 0.8593
No log 20.6957 476 0.7232 0.6950 0.7232 0.8504
No log 20.7826 478 0.7606 0.6571 0.7606 0.8721
No log 20.8696 480 0.8318 0.6131 0.8318 0.9120
No log 20.9565 482 0.9223 0.6197 0.9223 0.9603
No log 21.0435 484 1.0611 0.5957 1.0611 1.0301
No log 21.1304 486 1.0210 0.5957 1.0210 1.0104
No log 21.2174 488 0.8705 0.6131 0.8705 0.9330
No log 21.3043 490 0.7588 0.6475 0.7588 0.8711
No log 21.3913 492 0.7584 0.6377 0.7584 0.8708
No log 21.4783 494 0.7575 0.6571 0.7575 0.8703
No log 21.5652 496 0.7752 0.6475 0.7752 0.8805
No log 21.6522 498 0.8010 0.6475 0.8010 0.8950
0.3751 21.7391 500 0.8073 0.6619 0.8073 0.8985
0.3751 21.8261 502 0.8072 0.6571 0.8072 0.8985
0.3751 21.9130 504 0.8242 0.6569 0.8242 0.9079
0.3751 22.0 506 0.8223 0.6377 0.8223 0.9068
0.3751 22.0870 508 0.7747 0.6571 0.7747 0.8802
0.3751 22.1739 510 0.7901 0.6620 0.7901 0.8889
0.3751 22.2609 512 0.8862 0.6486 0.8862 0.9414
0.3751 22.3478 514 0.8811 0.6197 0.8811 0.9387
0.3751 22.4348 516 0.8531 0.6176 0.8531 0.9236
0.3751 22.5217 518 0.8551 0.6015 0.8551 0.9247
0.3751 22.6087 520 0.8475 0.6074 0.8475 0.9206

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k5_task1_organization

Finetuned
(4019)
this model