ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k10_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6430
  • Qwk: 0.5510
  • Mse: 0.6430
  • Rmse: 0.8019

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.04 2 4.0514 -0.0323 4.0514 2.0128
No log 0.08 4 2.7174 0.0035 2.7174 1.6485
No log 0.12 6 1.3984 0.0310 1.3984 1.1825
No log 0.16 8 1.1221 0.2140 1.1221 1.0593
No log 0.2 10 1.1106 0.2314 1.1106 1.0539
No log 0.24 12 1.1546 0.1240 1.1546 1.0745
No log 0.28 14 1.5389 0.0759 1.5389 1.2405
No log 0.32 16 1.6085 0.1744 1.6085 1.2683
No log 0.36 18 1.1791 0.2468 1.1791 1.0859
No log 0.4 20 0.7876 0.4109 0.7876 0.8875
No log 0.44 22 0.7594 0.4966 0.7594 0.8714
No log 0.48 24 0.7569 0.4845 0.7569 0.8700
No log 0.52 26 0.7505 0.4861 0.7505 0.8663
No log 0.56 28 0.8058 0.3927 0.8058 0.8977
No log 0.6 30 0.7040 0.5217 0.7040 0.8390
No log 0.64 32 0.6778 0.5217 0.6778 0.8233
No log 0.68 34 0.6383 0.6414 0.6383 0.7990
No log 0.72 36 0.6474 0.6374 0.6474 0.8046
No log 0.76 38 0.6447 0.6095 0.6447 0.8029
No log 0.8 40 0.7840 0.6109 0.7840 0.8854
No log 0.84 42 1.1395 0.4768 1.1395 1.0675
No log 0.88 44 0.9586 0.6086 0.9586 0.9791
No log 0.92 46 0.7023 0.5916 0.7023 0.8380
No log 0.96 48 0.7377 0.6622 0.7377 0.8589
No log 1.0 50 0.6818 0.6374 0.6818 0.8257
No log 1.04 52 0.6397 0.5825 0.6397 0.7998
No log 1.08 54 0.7952 0.6175 0.7952 0.8917
No log 1.12 56 1.3722 0.4027 1.3722 1.1714
No log 1.16 58 1.6334 0.3087 1.6334 1.2780
No log 1.2 60 1.3402 0.3830 1.3402 1.1577
No log 1.24 62 0.7757 0.6061 0.7757 0.8808
No log 1.28 64 0.6793 0.6497 0.6793 0.8242
No log 1.32 66 1.1577 0.4237 1.1577 1.0759
No log 1.3600 68 1.2644 0.3969 1.2644 1.1245
No log 1.4 70 0.9773 0.5277 0.9773 0.9886
No log 1.44 72 0.6936 0.6188 0.6936 0.8328
No log 1.48 74 0.7428 0.5382 0.7428 0.8618
No log 1.52 76 0.7579 0.5718 0.7579 0.8706
No log 1.56 78 0.6956 0.6097 0.6956 0.8340
No log 1.6 80 0.6202 0.6772 0.6202 0.7875
No log 1.6400 82 0.6754 0.6353 0.6754 0.8218
No log 1.6800 84 0.7392 0.6243 0.7392 0.8598
No log 1.72 86 0.6631 0.6512 0.6631 0.8143
No log 1.76 88 0.6688 0.6226 0.6688 0.8178
No log 1.8 90 0.6741 0.6374 0.6741 0.8211
No log 1.8400 92 0.6414 0.6404 0.6414 0.8009
No log 1.88 94 0.6719 0.6162 0.6719 0.8197
No log 1.92 96 0.6354 0.6708 0.6354 0.7971
No log 1.96 98 0.5975 0.6750 0.5975 0.7730
No log 2.0 100 0.6006 0.6317 0.6006 0.7750
No log 2.04 102 0.6081 0.6282 0.6081 0.7798
No log 2.08 104 0.6442 0.6032 0.6442 0.8026
No log 2.12 106 0.7079 0.6026 0.7079 0.8414
No log 2.16 108 0.8349 0.6289 0.8349 0.9137
No log 2.2 110 0.8069 0.6165 0.8069 0.8983
No log 2.24 112 0.6564 0.6365 0.6564 0.8102
No log 2.2800 114 0.6332 0.7049 0.6332 0.7958
No log 2.32 116 0.6464 0.6263 0.6464 0.8040
No log 2.36 118 0.7079 0.6528 0.7079 0.8414
No log 2.4 120 0.7446 0.6078 0.7446 0.8629
No log 2.44 122 0.7408 0.5795 0.7408 0.8607
No log 2.48 124 0.7327 0.6556 0.7327 0.8560
No log 2.52 126 0.7325 0.6659 0.7325 0.8558
No log 2.56 128 0.7626 0.6471 0.7626 0.8733
No log 2.6 130 0.7582 0.6639 0.7582 0.8708
No log 2.64 132 0.7359 0.6341 0.7359 0.8579
No log 2.68 134 0.6489 0.6339 0.6489 0.8056
No log 2.7200 136 0.7038 0.6575 0.7038 0.8389
No log 2.76 138 0.8868 0.5087 0.8868 0.9417
No log 2.8 140 0.8515 0.5634 0.8515 0.9228
No log 2.84 142 0.7006 0.6595 0.7006 0.8370
No log 2.88 144 0.6681 0.6044 0.6681 0.8174
No log 2.92 146 0.6953 0.6044 0.6953 0.8338
No log 2.96 148 0.7248 0.6609 0.7248 0.8514
No log 3.0 150 0.7187 0.6630 0.7187 0.8478
No log 3.04 152 0.7155 0.6651 0.7155 0.8459
No log 3.08 154 0.6611 0.6231 0.6611 0.8131
No log 3.12 156 0.6107 0.6038 0.6107 0.7814
No log 3.16 158 0.5901 0.5863 0.5901 0.7682
No log 3.2 160 0.6456 0.6507 0.6456 0.8035
No log 3.24 162 0.6662 0.6490 0.6662 0.8162
No log 3.2800 164 0.5750 0.6275 0.5750 0.7583
No log 3.32 166 0.6385 0.6892 0.6385 0.7991
No log 3.36 168 0.6862 0.6593 0.6862 0.8284
No log 3.4 170 0.6418 0.6669 0.6418 0.8011
No log 3.44 172 0.6395 0.6437 0.6395 0.7997
No log 3.48 174 0.6663 0.6442 0.6663 0.8163
No log 3.52 176 0.6555 0.6125 0.6555 0.8097
No log 3.56 178 0.6556 0.6358 0.6556 0.8097
No log 3.6 180 0.7009 0.6434 0.7009 0.8372
No log 3.64 182 0.7122 0.6592 0.7122 0.8439
No log 3.68 184 0.6361 0.6328 0.6361 0.7976
No log 3.7200 186 0.6153 0.6328 0.6153 0.7844
No log 3.76 188 0.6123 0.6397 0.6123 0.7825
No log 3.8 190 0.6430 0.6249 0.6430 0.8019
No log 3.84 192 0.6801 0.6414 0.6801 0.8247
No log 3.88 194 0.6803 0.6487 0.6803 0.8248
No log 3.92 196 0.6981 0.5865 0.6981 0.8355
No log 3.96 198 0.6643 0.6305 0.6643 0.8151
No log 4.0 200 0.6499 0.5857 0.6499 0.8062
No log 4.04 202 0.6387 0.6073 0.6387 0.7992
No log 4.08 204 0.6109 0.6446 0.6109 0.7816
No log 4.12 206 0.6028 0.6374 0.6028 0.7764
No log 4.16 208 0.6261 0.6479 0.6261 0.7912
No log 4.2 210 0.6104 0.6491 0.6104 0.7813
No log 4.24 212 0.6553 0.6345 0.6553 0.8095
No log 4.28 214 0.7591 0.6320 0.7591 0.8713
No log 4.32 216 0.7817 0.5961 0.7817 0.8841
No log 4.36 218 0.7578 0.6071 0.7578 0.8705
No log 4.4 220 0.7709 0.6619 0.7709 0.8780
No log 4.44 222 0.7374 0.5719 0.7374 0.8587
No log 4.48 224 0.6449 0.5989 0.6449 0.8030
No log 4.52 226 0.6383 0.6623 0.6383 0.7989
No log 4.5600 228 0.7669 0.5647 0.7669 0.8757
No log 4.6 230 0.7468 0.6230 0.7468 0.8642
No log 4.64 232 0.6573 0.6188 0.6573 0.8107
No log 4.68 234 0.5945 0.6122 0.5945 0.7711
No log 4.72 236 0.6009 0.6500 0.6009 0.7752
No log 4.76 238 0.6874 0.6263 0.6874 0.8291
No log 4.8 240 0.6637 0.6363 0.6637 0.8147
No log 4.84 242 0.6145 0.6879 0.6145 0.7839
No log 4.88 244 0.5757 0.6224 0.5757 0.7587
No log 4.92 246 0.5804 0.6401 0.5804 0.7618
No log 4.96 248 0.5783 0.5975 0.5783 0.7604
No log 5.0 250 0.5979 0.6361 0.5979 0.7733
No log 5.04 252 0.6322 0.6319 0.6322 0.7951
No log 5.08 254 0.6485 0.6493 0.6485 0.8053
No log 5.12 256 0.6433 0.6404 0.6433 0.8021
No log 5.16 258 0.6580 0.6669 0.6580 0.8112
No log 5.2 260 0.6623 0.6830 0.6623 0.8138
No log 5.24 262 0.6789 0.6830 0.6789 0.8240
No log 5.28 264 0.6778 0.6402 0.6778 0.8233
No log 5.32 266 0.6551 0.6502 0.6551 0.8094
No log 5.36 268 0.5994 0.6165 0.5994 0.7742
No log 5.4 270 0.6014 0.5618 0.6014 0.7755
No log 5.44 272 0.6121 0.6042 0.6121 0.7824
No log 5.48 274 0.6124 0.6902 0.6124 0.7826
No log 5.52 276 0.6322 0.6609 0.6322 0.7951
No log 5.5600 278 0.6668 0.5527 0.6668 0.8166
No log 5.6 280 0.6750 0.5808 0.6750 0.8216
No log 5.64 282 0.6382 0.6397 0.6382 0.7989
No log 5.68 284 0.6335 0.6022 0.6335 0.7959
No log 5.72 286 0.6644 0.5688 0.6644 0.8151
No log 5.76 288 0.7441 0.6336 0.7441 0.8626
No log 5.8 290 0.7468 0.6816 0.7468 0.8642
No log 5.84 292 0.7685 0.6419 0.7685 0.8767
No log 5.88 294 0.7513 0.6291 0.7513 0.8668
No log 5.92 296 0.7133 0.5899 0.7133 0.8446
No log 5.96 298 0.7462 0.6223 0.7462 0.8638
No log 6.0 300 0.7804 0.5724 0.7804 0.8834
No log 6.04 302 0.7585 0.5607 0.7585 0.8709
No log 6.08 304 0.7085 0.5416 0.7085 0.8417
No log 6.12 306 0.6901 0.5573 0.6901 0.8307
No log 6.16 308 0.7129 0.5864 0.7129 0.8443
No log 6.2 310 0.7170 0.6291 0.7170 0.8468
No log 6.24 312 0.6842 0.6230 0.6842 0.8272
No log 6.28 314 0.6538 0.6460 0.6538 0.8085
No log 6.32 316 0.6178 0.6167 0.6178 0.7860
No log 6.36 318 0.6006 0.5785 0.6006 0.7750
No log 6.4 320 0.5991 0.5879 0.5991 0.7740
No log 6.44 322 0.6304 0.6099 0.6304 0.7940
No log 6.48 324 0.6043 0.6144 0.6043 0.7773
No log 6.52 326 0.6114 0.6341 0.6114 0.7819
No log 6.5600 328 0.6743 0.6432 0.6743 0.8212
No log 6.6 330 0.6751 0.6590 0.6751 0.8216
No log 6.64 332 0.6234 0.6751 0.6234 0.7895
No log 6.68 334 0.6205 0.6242 0.6205 0.7877
No log 6.72 336 0.6100 0.6087 0.6100 0.7810
No log 6.76 338 0.6055 0.5703 0.6055 0.7782
No log 6.8 340 0.6149 0.5808 0.6149 0.7841
No log 6.84 342 0.6730 0.6002 0.6730 0.8204
No log 6.88 344 0.7497 0.6333 0.7497 0.8658
No log 6.92 346 0.7353 0.6189 0.7353 0.8575
No log 6.96 348 0.6846 0.5370 0.6846 0.8274
No log 7.0 350 0.6710 0.6196 0.6710 0.8192
No log 7.04 352 0.6846 0.5972 0.6846 0.8274
No log 7.08 354 0.6834 0.5972 0.6834 0.8267
No log 7.12 356 0.6739 0.6094 0.6739 0.8209
No log 7.16 358 0.7129 0.5850 0.7129 0.8443
No log 7.2 360 0.8072 0.5222 0.8072 0.8984
No log 7.24 362 0.8476 0.5114 0.8476 0.9206
No log 7.28 364 0.7814 0.5707 0.7814 0.8840
No log 7.32 366 0.7059 0.5811 0.7059 0.8402
No log 7.36 368 0.6707 0.5046 0.6707 0.8190
No log 7.4 370 0.6716 0.5174 0.6716 0.8195
No log 7.44 372 0.6669 0.5274 0.6669 0.8166
No log 7.48 374 0.6806 0.6324 0.6806 0.8250
No log 7.52 376 0.7033 0.5673 0.7033 0.8386
No log 7.5600 378 0.6993 0.5673 0.6993 0.8362
No log 7.6 380 0.6963 0.5673 0.6963 0.8345
No log 7.64 382 0.6830 0.5892 0.6830 0.8264
No log 7.68 384 0.6611 0.6278 0.6611 0.8131
No log 7.72 386 0.6554 0.6123 0.6554 0.8096
No log 7.76 388 0.6496 0.5483 0.6496 0.8060
No log 7.8 390 0.6569 0.5838 0.6569 0.8105
No log 7.84 392 0.6923 0.5911 0.6923 0.8321
No log 7.88 394 0.7517 0.5938 0.7517 0.8670
No log 7.92 396 0.7550 0.6420 0.7550 0.8689
No log 7.96 398 0.7588 0.6259 0.7588 0.8711
No log 8.0 400 0.7463 0.6342 0.7463 0.8639
No log 8.04 402 0.6994 0.5684 0.6994 0.8363
No log 8.08 404 0.6658 0.5275 0.6658 0.8160
No log 8.12 406 0.6576 0.5030 0.6576 0.8109
No log 8.16 408 0.6627 0.5169 0.6627 0.8140
No log 8.2 410 0.6905 0.5737 0.6905 0.8309
No log 8.24 412 0.6861 0.5726 0.6861 0.8283
No log 8.28 414 0.6761 0.5478 0.6761 0.8223
No log 8.32 416 0.6836 0.5964 0.6836 0.8268
No log 8.36 418 0.6999 0.6337 0.6999 0.8366
No log 8.4 420 0.6844 0.6157 0.6844 0.8273
No log 8.44 422 0.6467 0.5602 0.6467 0.8041
No log 8.48 424 0.6391 0.5492 0.6391 0.7994
No log 8.52 426 0.6387 0.5487 0.6387 0.7992
No log 8.56 428 0.6431 0.6177 0.6431 0.8019
No log 8.6 430 0.6513 0.6365 0.6513 0.8070
No log 8.64 432 0.6677 0.6406 0.6677 0.8172
No log 8.68 434 0.6726 0.6414 0.6726 0.8201
No log 8.72 436 0.6821 0.6103 0.6821 0.8259
No log 8.76 438 0.6929 0.6077 0.6929 0.8324
No log 8.8 440 0.6900 0.6377 0.6900 0.8307
No log 8.84 442 0.6683 0.6414 0.6683 0.8175
No log 8.88 444 0.6441 0.5726 0.6441 0.8025
No log 8.92 446 0.6347 0.5626 0.6347 0.7967
No log 8.96 448 0.6327 0.5505 0.6327 0.7954
No log 9.0 450 0.6490 0.6066 0.6490 0.8056
No log 9.04 452 0.6452 0.5990 0.6452 0.8032
No log 9.08 454 0.6299 0.5626 0.6299 0.7936
No log 9.12 456 0.6328 0.6042 0.6328 0.7955
No log 9.16 458 0.6348 0.6042 0.6348 0.7967
No log 9.2 460 0.6332 0.6042 0.6332 0.7958
No log 9.24 462 0.6395 0.5747 0.6395 0.7997
No log 9.28 464 0.6509 0.5002 0.6509 0.8068
No log 9.32 466 0.6647 0.5585 0.6647 0.8153
No log 9.36 468 0.6931 0.5872 0.6931 0.8325
No log 9.4 470 0.6748 0.5811 0.6748 0.8214
No log 9.44 472 0.6463 0.5017 0.6463 0.8039
No log 9.48 474 0.6302 0.5747 0.6302 0.7939
No log 9.52 476 0.6300 0.6060 0.6300 0.7937
No log 9.56 478 0.6460 0.6244 0.6460 0.8038
No log 9.6 480 0.6294 0.6812 0.6294 0.7933
No log 9.64 482 0.6339 0.6226 0.6339 0.7962
No log 9.68 484 0.6806 0.5376 0.6806 0.8250
No log 9.72 486 0.6760 0.6147 0.6760 0.8222
No log 9.76 488 0.6568 0.5467 0.6568 0.8104
No log 9.8 490 0.6488 0.5176 0.6488 0.8055
No log 9.84 492 0.6714 0.5327 0.6714 0.8194
No log 9.88 494 0.6837 0.5566 0.6837 0.8268
No log 9.92 496 0.6560 0.5852 0.6560 0.8099
No log 9.96 498 0.6501 0.6476 0.6501 0.8063
0.2706 10.0 500 0.6742 0.6335 0.6742 0.8211
0.2706 10.04 502 0.6718 0.5808 0.6718 0.8196
0.2706 10.08 504 0.6575 0.5866 0.6575 0.8109
0.2706 10.12 506 0.6514 0.5480 0.6514 0.8071
0.2706 10.16 508 0.6467 0.5510 0.6467 0.8042
0.2706 10.2 510 0.6430 0.5510 0.6430 0.8019

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k10_task5_organization

Finetuned
(4019)
this model