ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k14_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6983
  • Qwk: 0.7194
  • Mse: 0.6983
  • Rmse: 0.8356

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0190 2 6.7739 0.0308 6.7739 2.6027
No log 0.0381 4 5.1846 0.0072 5.1846 2.2770
No log 0.0571 6 2.9936 0.0848 2.9936 1.7302
No log 0.0762 8 2.1278 0.1364 2.1278 1.4587
No log 0.0952 10 1.8352 0.2581 1.8352 1.3547
No log 0.1143 12 1.5715 0.1455 1.5715 1.2536
No log 0.1333 14 1.8025 0.2000 1.8025 1.3426
No log 0.1524 16 1.9851 -0.0370 1.9851 1.4089
No log 0.1714 18 2.5745 -0.1212 2.5745 1.6045
No log 0.1905 20 2.6569 -0.0576 2.6569 1.6300
No log 0.2095 22 2.0708 0.0992 2.0708 1.4390
No log 0.2286 24 1.5655 0.2075 1.5655 1.2512
No log 0.2476 26 1.3968 0.2752 1.3968 1.1819
No log 0.2667 28 1.3438 0.2752 1.3438 1.1592
No log 0.2857 30 1.3179 0.3793 1.3179 1.1480
No log 0.3048 32 1.3345 0.3717 1.3345 1.1552
No log 0.3238 34 1.3879 0.2407 1.3879 1.1781
No log 0.3429 36 1.3083 0.3091 1.3083 1.1438
No log 0.3619 38 1.1811 0.3717 1.1811 1.0868
No log 0.3810 40 1.1146 0.4 1.1146 1.0557
No log 0.4 42 1.1763 0.4516 1.1763 1.0846
No log 0.4190 44 1.1676 0.4496 1.1676 1.0806
No log 0.4381 46 1.0495 0.5116 1.0495 1.0244
No log 0.4571 48 0.9677 0.6165 0.9677 0.9837
No log 0.4762 50 0.9508 0.6667 0.9508 0.9751
No log 0.4952 52 0.9663 0.6377 0.9663 0.9830
No log 0.5143 54 0.9811 0.6131 0.9811 0.9905
No log 0.5333 56 1.0553 0.5970 1.0553 1.0273
No log 0.5524 58 0.9676 0.6241 0.9676 0.9837
No log 0.5714 60 0.9226 0.6528 0.9226 0.9605
No log 0.5905 62 0.9553 0.6667 0.9553 0.9774
No log 0.6095 64 1.0980 0.6203 1.0980 1.0478
No log 0.6286 66 1.1146 0.6364 1.1146 1.0558
No log 0.6476 68 0.7391 0.7843 0.7391 0.8597
No log 0.6667 70 0.7380 0.7027 0.7380 0.8591
No log 0.6857 72 0.7792 0.6531 0.7792 0.8827
No log 0.7048 74 0.6749 0.7778 0.6749 0.8215
No log 0.7238 76 0.6950 0.7722 0.6950 0.8336
No log 0.7429 78 0.6760 0.7662 0.6760 0.8222
No log 0.7619 80 0.6606 0.7662 0.6606 0.8128
No log 0.7810 82 0.7952 0.7261 0.7952 0.8917
No log 0.8 84 1.1062 0.6154 1.1062 1.0518
No log 0.8190 86 0.8865 0.7024 0.8865 0.9416
No log 0.8381 88 0.8262 0.6866 0.8262 0.9090
No log 0.8571 90 0.9765 0.6324 0.9765 0.9882
No log 0.8762 92 0.8021 0.6815 0.8021 0.8956
No log 0.8952 94 0.6500 0.7733 0.6500 0.8062
No log 0.9143 96 0.8998 0.6854 0.8998 0.9486
No log 0.9333 98 0.9279 0.6711 0.9279 0.9633
No log 0.9524 100 0.8809 0.7172 0.8809 0.9386
No log 0.9714 102 0.7668 0.7211 0.7668 0.8757
No log 0.9905 104 0.5865 0.7805 0.5865 0.7658
No log 1.0095 106 0.4958 0.8293 0.4958 0.7041
No log 1.0286 108 0.5021 0.8242 0.5021 0.7086
No log 1.0476 110 0.5351 0.8221 0.5351 0.7315
No log 1.0667 112 0.6230 0.7882 0.6230 0.7893
No log 1.0857 114 0.6676 0.7765 0.6676 0.8171
No log 1.1048 116 0.6547 0.8024 0.6547 0.8091
No log 1.1238 118 0.6882 0.7550 0.6882 0.8296
No log 1.1429 120 0.7623 0.7436 0.7623 0.8731
No log 1.1619 122 0.8607 0.7108 0.8607 0.9277
No log 1.1810 124 0.8639 0.6914 0.8639 0.9295
No log 1.2 126 0.8787 0.7030 0.8787 0.9374
No log 1.2190 128 0.9051 0.6941 0.9051 0.9514
No log 1.2381 130 0.9020 0.6951 0.9020 0.9497
No log 1.2571 132 0.8554 0.7215 0.8554 0.9249
No log 1.2762 134 0.8773 0.6887 0.8773 0.9366
No log 1.2952 136 0.8891 0.6968 0.8891 0.9429
No log 1.3143 138 0.7790 0.6980 0.7790 0.8826
No log 1.3333 140 0.6656 0.7417 0.6656 0.8158
No log 1.3524 142 0.5962 0.7778 0.5962 0.7721
No log 1.3714 144 0.6069 0.8075 0.6069 0.7791
No log 1.3905 146 0.6592 0.8 0.6592 0.8119
No log 1.4095 148 0.7214 0.7665 0.7214 0.8494
No log 1.4286 150 0.8823 0.7174 0.8823 0.9393
No log 1.4476 152 0.9377 0.7128 0.9377 0.9683
No log 1.4667 154 0.8635 0.7403 0.8635 0.9293
No log 1.4857 156 0.6564 0.7453 0.6564 0.8102
No log 1.5048 158 0.6607 0.7742 0.6607 0.8128
No log 1.5238 160 0.7973 0.7027 0.7973 0.8929
No log 1.5429 162 1.1800 0.5939 1.1800 1.0863
No log 1.5619 164 1.2359 0.5556 1.2359 1.1117
No log 1.5810 166 0.9320 0.5985 0.9320 0.9654
No log 1.6 168 0.8218 0.6667 0.8218 0.9065
No log 1.6190 170 0.8211 0.6767 0.8211 0.9061
No log 1.6381 172 0.9487 0.6324 0.9487 0.9740
No log 1.6571 174 1.2007 0.5442 1.2007 1.0958
No log 1.6762 176 1.1893 0.5503 1.1893 1.0906
No log 1.6952 178 1.0337 0.5860 1.0337 1.0167
No log 1.7143 180 0.7646 0.7590 0.7646 0.8744
No log 1.7333 182 0.6687 0.8023 0.6687 0.8177
No log 1.7524 184 0.6413 0.8161 0.6413 0.8008
No log 1.7714 186 0.6141 0.7955 0.6141 0.7836
No log 1.7905 188 0.6070 0.7811 0.6070 0.7791
No log 1.8095 190 0.6050 0.7975 0.6050 0.7778
No log 1.8286 192 0.7951 0.7024 0.7951 0.8917
No log 1.8476 194 1.0900 0.6243 1.0900 1.0440
No log 1.8667 196 1.1595 0.6034 1.1595 1.0768
No log 1.8857 198 1.0051 0.6587 1.0051 1.0025
No log 1.9048 200 0.8329 0.7105 0.8329 0.9126
No log 1.9238 202 0.8260 0.7421 0.8260 0.9089
No log 1.9429 204 0.9616 0.6667 0.9616 0.9806
No log 1.9619 206 1.2294 0.6364 1.2294 1.1088
No log 1.9810 208 1.4076 0.6066 1.4076 1.1864
No log 2.0 210 1.2854 0.6321 1.2854 1.1338
No log 2.0190 212 1.0277 0.6190 1.0277 1.0138
No log 2.0381 214 0.9405 0.6619 0.9405 0.9698
No log 2.0571 216 0.8785 0.7007 0.8785 0.9373
No log 2.0762 218 0.6205 0.8158 0.6205 0.7877
No log 2.0952 220 0.5777 0.8050 0.5777 0.7600
No log 2.1143 222 0.6337 0.7632 0.6337 0.7960
No log 2.1333 224 0.7578 0.7248 0.7578 0.8705
No log 2.1524 226 0.8356 0.6993 0.8356 0.9141
No log 2.1714 228 0.7066 0.6957 0.7066 0.8406
No log 2.1905 230 0.7754 0.7574 0.7754 0.8806
No log 2.2095 232 0.8914 0.7386 0.8914 0.9441
No log 2.2286 234 0.7597 0.6897 0.7597 0.8716
No log 2.2476 236 0.6355 0.7586 0.6355 0.7972
No log 2.2667 238 0.5860 0.7755 0.5860 0.7655
No log 2.2857 240 0.5857 0.7712 0.5857 0.7653
No log 2.3048 242 0.7522 0.7574 0.7522 0.8673
No log 2.3238 244 0.8824 0.7579 0.8824 0.9393
No log 2.3429 246 0.8424 0.7835 0.8424 0.9178
No log 2.3619 248 0.6466 0.8172 0.6466 0.8041
No log 2.3810 250 0.5654 0.8214 0.5654 0.7519
No log 2.4 252 0.5481 0.8121 0.5481 0.7403
No log 2.4190 254 0.6072 0.8144 0.6072 0.7793
No log 2.4381 256 0.6996 0.7711 0.6996 0.8364
No log 2.4571 258 0.7749 0.7614 0.7749 0.8803
No log 2.4762 260 0.6561 0.8072 0.6561 0.8100
No log 2.4952 262 0.5161 0.7867 0.5161 0.7184
No log 2.5143 264 0.5107 0.7922 0.5107 0.7146
No log 2.5333 266 0.5834 0.8025 0.5834 0.7638
No log 2.5524 268 0.6996 0.8023 0.6996 0.8364
No log 2.5714 270 0.7510 0.7836 0.7510 0.8666
No log 2.5905 272 0.7363 0.8 0.7363 0.8581
No log 2.6095 274 0.7804 0.7831 0.7804 0.8834
No log 2.6286 276 0.7170 0.8049 0.7170 0.8467
No log 2.6476 278 0.6873 0.8049 0.6873 0.8290
No log 2.6667 280 0.6396 0.8121 0.6396 0.7998
No log 2.6857 282 0.6569 0.8121 0.6569 0.8105
No log 2.7048 284 0.8056 0.8022 0.8056 0.8976
No log 2.7238 286 0.8801 0.7576 0.8801 0.9381
No log 2.7429 288 0.7092 0.8177 0.7092 0.8421
No log 2.7619 290 0.5892 0.8095 0.5892 0.7676
No log 2.7810 292 0.5885 0.8144 0.5885 0.7671
No log 2.8 294 0.6494 0.775 0.6494 0.8058
No log 2.8190 296 0.6685 0.7643 0.6685 0.8176
No log 2.8381 298 0.6832 0.7368 0.6832 0.8266
No log 2.8571 300 0.6408 0.7632 0.6408 0.8005
No log 2.8762 302 0.6027 0.8025 0.6027 0.7764
No log 2.8952 304 0.7045 0.7929 0.7045 0.8394
No log 2.9143 306 0.9268 0.7582 0.9268 0.9627
No log 2.9333 308 0.9289 0.7650 0.9289 0.9638
No log 2.9524 310 0.8015 0.7709 0.8015 0.8953
No log 2.9714 312 0.5884 0.8049 0.5884 0.7670
No log 2.9905 314 0.5506 0.7925 0.5506 0.7421
No log 3.0095 316 0.6292 0.7742 0.6292 0.7932
No log 3.0286 318 0.6740 0.7333 0.6740 0.8210
No log 3.0476 320 0.6563 0.7550 0.6563 0.8101
No log 3.0667 322 0.6499 0.76 0.6499 0.8061
No log 3.0857 324 0.6931 0.7333 0.6931 0.8325
No log 3.1048 326 0.7024 0.7451 0.7024 0.8381
No log 3.1238 328 0.6990 0.7564 0.6990 0.8361
No log 3.1429 330 0.6545 0.7673 0.6545 0.8090
No log 3.1619 332 0.5774 0.7925 0.5774 0.7599
No log 3.1810 334 0.5807 0.8187 0.5807 0.7621
No log 3.2 336 0.6640 0.8249 0.6640 0.8149
No log 3.2190 338 0.7556 0.8023 0.7556 0.8693
No log 3.2381 340 0.8633 0.7368 0.8633 0.9291
No log 3.2571 342 0.9015 0.7066 0.9015 0.9495
No log 3.2762 344 0.8678 0.7052 0.8678 0.9315
No log 3.2952 346 0.8372 0.7337 0.8372 0.9150
No log 3.3143 348 0.7754 0.7545 0.7754 0.8806
No log 3.3333 350 0.7667 0.7545 0.7667 0.8756
No log 3.3524 352 0.7867 0.7470 0.7867 0.8869
No log 3.3714 354 0.7348 0.7545 0.7348 0.8572
No log 3.3905 356 0.7855 0.7674 0.7855 0.8863
No log 3.4095 358 0.7450 0.7417 0.7450 0.8631
No log 3.4286 360 0.6911 0.7651 0.6911 0.8313
No log 3.4476 362 0.6634 0.7671 0.6634 0.8145
No log 3.4667 364 0.6705 0.7763 0.6705 0.8188
No log 3.4857 366 0.6574 0.7871 0.6574 0.8108
No log 3.5048 368 0.6113 0.7871 0.6113 0.7819
No log 3.5238 370 0.6146 0.7821 0.6146 0.7840
No log 3.5429 372 0.7095 0.8045 0.7095 0.8423
No log 3.5619 374 0.7734 0.8065 0.7734 0.8794
No log 3.5810 376 0.6777 0.8249 0.6777 0.8232
No log 3.6 378 0.5755 0.7871 0.5755 0.7586
No log 3.6190 380 0.5793 0.7785 0.5793 0.7611
No log 3.6381 382 0.6663 0.7712 0.6663 0.8163
No log 3.6571 384 0.7162 0.7857 0.7162 0.8463
No log 3.6762 386 0.6593 0.8023 0.6593 0.8119
No log 3.6952 388 0.5399 0.7925 0.5399 0.7348
No log 3.7143 390 0.4892 0.8049 0.4892 0.6995
No log 3.7333 392 0.4808 0.8072 0.4808 0.6934
No log 3.7524 394 0.4692 0.8121 0.4692 0.6850
No log 3.7714 396 0.5120 0.7949 0.5120 0.7155
No log 3.7905 398 0.7133 0.7976 0.7133 0.8445
No log 3.8095 400 0.9336 0.7182 0.9336 0.9662
No log 3.8286 402 0.8793 0.7529 0.8793 0.9377
No log 3.8476 404 0.6771 0.8025 0.6771 0.8229
No log 3.8667 406 0.5395 0.7871 0.5395 0.7345
No log 3.8857 408 0.5321 0.8242 0.5321 0.7295
No log 3.9048 410 0.5421 0.8072 0.5421 0.7363
No log 3.9238 412 0.6152 0.8092 0.6152 0.7843
No log 3.9429 414 0.8146 0.7556 0.8146 0.9025
No log 3.9619 416 1.0051 0.7150 1.0051 1.0025
No log 3.9810 418 1.1254 0.6915 1.1254 1.0608
No log 4.0 420 1.0117 0.6905 1.0117 1.0058
No log 4.0190 422 0.8284 0.6901 0.8284 0.9101
No log 4.0381 424 0.7428 0.6917 0.7428 0.8619
No log 4.0571 426 0.7488 0.6917 0.7488 0.8653
No log 4.0762 428 0.7354 0.7015 0.7354 0.8575
No log 4.0952 430 0.6903 0.7059 0.6903 0.8308
No log 4.1143 432 0.6249 0.7299 0.6249 0.7905
No log 4.1333 434 0.5625 0.7843 0.5625 0.7500
No log 4.1524 436 0.5970 0.8070 0.5970 0.7726
No log 4.1714 438 0.7091 0.7933 0.7091 0.8421
No log 4.1905 440 0.7358 0.7892 0.7358 0.8578
No log 4.2095 442 0.6641 0.7933 0.6641 0.8149
No log 4.2286 444 0.7216 0.7574 0.7216 0.8495
No log 4.2476 446 0.7741 0.7067 0.7741 0.8798
No log 4.2667 448 0.7903 0.6809 0.7903 0.8890
No log 4.2857 450 0.7517 0.7 0.7517 0.8670
No log 4.3048 452 0.6672 0.7050 0.6672 0.8168
No log 4.3238 454 0.5733 0.7651 0.5733 0.7572
No log 4.3429 456 0.5376 0.8 0.5376 0.7332
No log 4.3619 458 0.5732 0.7821 0.5732 0.7571
No log 4.3810 460 0.6989 0.7692 0.6989 0.8360
No log 4.4 462 0.8526 0.7442 0.8526 0.9233
No log 4.4190 464 0.8012 0.7514 0.8012 0.8951
No log 4.4381 466 0.6991 0.7619 0.6991 0.8361
No log 4.4571 468 0.5469 0.8098 0.5469 0.7395
No log 4.4762 470 0.5006 0.8 0.5006 0.7075
No log 4.4952 472 0.5074 0.8 0.5074 0.7123
No log 4.5143 474 0.5974 0.8202 0.5974 0.7729
No log 4.5333 476 0.7794 0.7444 0.7794 0.8828
No log 4.5524 478 0.8452 0.7303 0.8452 0.9193
No log 4.5714 480 0.7266 0.7439 0.7266 0.8524
No log 4.5905 482 0.6170 0.7194 0.6170 0.7855
No log 4.6095 484 0.5991 0.7619 0.5991 0.7740
No log 4.6286 486 0.6179 0.7376 0.6179 0.7860
No log 4.6476 488 0.6433 0.7517 0.6433 0.8021
No log 4.6667 490 0.7028 0.7453 0.7028 0.8383
No log 4.6857 492 0.6853 0.7531 0.6853 0.8278
No log 4.7048 494 0.6558 0.7805 0.6558 0.8098
No log 4.7238 496 0.5913 0.7875 0.5913 0.7690
No log 4.7429 498 0.5397 0.7843 0.5397 0.7346
0.4509 4.7619 500 0.5532 0.7733 0.5532 0.7438
0.4509 4.7810 502 0.6232 0.7771 0.6232 0.7894
0.4509 4.8 504 0.8210 0.7485 0.8210 0.9061
0.4509 4.8190 506 0.9191 0.7386 0.9191 0.9587
0.4509 4.8381 508 0.8581 0.7375 0.8581 0.9263
0.4509 4.8571 510 0.7361 0.6993 0.7361 0.8580
0.4509 4.8762 512 0.6983 0.7194 0.6983 0.8356

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k14_task1_organization

Finetuned
(4023)
this model