ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k20_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6364
  • Qwk: 0.6176
  • Mse: 0.6364
  • Rmse: 0.7978

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.02 2 3.7592 -0.0414 3.7592 1.9389
No log 0.04 4 2.0383 0.1105 2.0383 1.4277
No log 0.06 6 1.1693 0.1525 1.1693 1.0814
No log 0.08 8 0.9788 0.2740 0.9788 0.9893
No log 0.1 10 1.0223 0.1601 1.0223 1.0111
No log 0.12 12 1.1127 0.2794 1.1127 1.0548
No log 0.14 14 1.0985 0.2667 1.0985 1.0481
No log 0.16 16 1.1212 0.2511 1.1212 1.0589
No log 0.18 18 1.1448 0.3584 1.1448 1.0700
No log 0.2 20 0.9500 0.4102 0.9500 0.9747
No log 0.22 22 1.4619 0.2208 1.4619 1.2091
No log 0.24 24 1.4506 0.2243 1.4506 1.2044
No log 0.26 26 0.9591 0.5565 0.9591 0.9793
No log 0.28 28 0.7635 0.5016 0.7635 0.8738
No log 0.3 30 0.7427 0.4778 0.7427 0.8618
No log 0.32 32 0.7196 0.5054 0.7196 0.8483
No log 0.34 34 0.7014 0.5060 0.7014 0.8375
No log 0.36 36 0.6921 0.5843 0.6921 0.8319
No log 0.38 38 0.7046 0.5215 0.7046 0.8394
No log 0.4 40 0.7463 0.4671 0.7463 0.8639
No log 0.42 42 0.7281 0.6062 0.7281 0.8533
No log 0.44 44 0.7394 0.6116 0.7394 0.8599
No log 0.46 46 0.7842 0.5812 0.7842 0.8856
No log 0.48 48 0.7190 0.6266 0.7190 0.8479
No log 0.5 50 0.7454 0.6253 0.7454 0.8634
No log 0.52 52 0.7504 0.6418 0.7504 0.8663
No log 0.54 54 0.7927 0.6554 0.7927 0.8904
No log 0.56 56 1.1159 0.4688 1.1159 1.0564
No log 0.58 58 1.1075 0.4272 1.1075 1.0524
No log 0.6 60 0.7810 0.5447 0.7810 0.8837
No log 0.62 62 0.6818 0.5518 0.6818 0.8257
No log 0.64 64 0.7373 0.5898 0.7373 0.8587
No log 0.66 66 0.7979 0.6160 0.7979 0.8932
No log 0.68 68 0.7128 0.6109 0.7128 0.8443
No log 0.7 70 0.6462 0.5701 0.6462 0.8038
No log 0.72 72 0.6584 0.5287 0.6584 0.8114
No log 0.74 74 0.6351 0.6398 0.6351 0.7969
No log 0.76 76 0.6446 0.6310 0.6446 0.8029
No log 0.78 78 0.6607 0.6672 0.6607 0.8128
No log 0.8 80 0.6591 0.6468 0.6591 0.8118
No log 0.82 82 0.7726 0.6243 0.7726 0.8790
No log 0.84 84 0.7121 0.6602 0.7121 0.8439
No log 0.86 86 0.6712 0.6583 0.6712 0.8193
No log 0.88 88 0.6474 0.6431 0.6474 0.8046
No log 0.9 90 0.6521 0.6707 0.6521 0.8075
No log 0.92 92 0.6595 0.6590 0.6595 0.8121
No log 0.94 94 0.6648 0.6431 0.6648 0.8153
No log 0.96 96 0.6423 0.6518 0.6423 0.8015
No log 0.98 98 0.6420 0.6518 0.6420 0.8012
No log 1.0 100 0.7254 0.5748 0.7254 0.8517
No log 1.02 102 0.7580 0.5770 0.7580 0.8706
No log 1.04 104 0.6916 0.6130 0.6916 0.8317
No log 1.06 106 0.7352 0.5604 0.7352 0.8574
No log 1.08 108 0.7575 0.5877 0.7575 0.8703
No log 1.1 110 0.7396 0.5736 0.7396 0.8600
No log 1.12 112 0.6486 0.6087 0.6486 0.8054
No log 1.1400 114 0.7498 0.5247 0.7498 0.8659
No log 1.16 116 0.7536 0.5344 0.7536 0.8681
No log 1.18 118 0.6744 0.6433 0.6744 0.8212
No log 1.2 120 0.6708 0.6712 0.6708 0.8190
No log 1.22 122 0.6809 0.6543 0.6809 0.8252
No log 1.24 124 0.6596 0.6535 0.6596 0.8122
No log 1.26 126 0.6269 0.6447 0.6269 0.7918
No log 1.28 128 0.6329 0.5854 0.6329 0.7955
No log 1.3 130 0.6473 0.6244 0.6473 0.8046
No log 1.32 132 0.6642 0.6388 0.6642 0.8150
No log 1.34 134 0.6702 0.6533 0.6702 0.8186
No log 1.3600 136 0.6844 0.6619 0.6844 0.8273
No log 1.38 138 0.6690 0.6619 0.6690 0.8179
No log 1.4 140 0.6237 0.6686 0.6237 0.7897
No log 1.42 142 0.6177 0.7131 0.6177 0.7859
No log 1.44 144 0.6210 0.6940 0.6210 0.7881
No log 1.46 146 0.6074 0.6853 0.6074 0.7794
No log 1.48 148 0.6362 0.6457 0.6362 0.7976
No log 1.5 150 0.7363 0.6283 0.7363 0.8581
No log 1.52 152 0.6555 0.6363 0.6555 0.8097
No log 1.54 154 0.6804 0.6552 0.6804 0.8249
No log 1.56 156 0.8081 0.5660 0.8081 0.8990
No log 1.58 158 0.7155 0.5639 0.7155 0.8459
No log 1.6 160 0.6486 0.5599 0.6486 0.8054
No log 1.62 162 0.6601 0.6265 0.6601 0.8125
No log 1.6400 164 0.7818 0.5332 0.7818 0.8842
No log 1.6600 166 0.7819 0.5763 0.7819 0.8843
No log 1.6800 168 0.7080 0.6445 0.7080 0.8414
No log 1.7 170 0.7436 0.6162 0.7436 0.8623
No log 1.72 172 0.8571 0.5715 0.8571 0.9258
No log 1.74 174 0.8056 0.5575 0.8056 0.8976
No log 1.76 176 0.6932 0.5740 0.6932 0.8326
No log 1.78 178 0.6543 0.5415 0.6543 0.8089
No log 1.8 180 0.7216 0.5688 0.7216 0.8495
No log 1.8200 182 0.7109 0.6082 0.7109 0.8431
No log 1.8400 184 0.6562 0.6843 0.6562 0.8101
No log 1.8600 186 0.7529 0.5840 0.7529 0.8677
No log 1.88 188 0.9000 0.5848 0.9000 0.9487
No log 1.9 190 0.9046 0.6056 0.9046 0.9511
No log 1.92 192 0.8311 0.6527 0.8311 0.9116
No log 1.94 194 0.7801 0.6342 0.7801 0.8832
No log 1.96 196 0.7370 0.6425 0.7370 0.8585
No log 1.98 198 0.6583 0.6751 0.6583 0.8114
No log 2.0 200 0.6214 0.6094 0.6214 0.7883
No log 2.02 202 0.5999 0.5713 0.5999 0.7745
No log 2.04 204 0.6655 0.5770 0.6655 0.8158
No log 2.06 206 0.7478 0.6135 0.7478 0.8647
No log 2.08 208 0.7023 0.5922 0.7023 0.8380
No log 2.1 210 0.6737 0.6361 0.6737 0.8208
No log 2.12 212 0.7064 0.6098 0.7064 0.8405
No log 2.14 214 0.6602 0.6562 0.6602 0.8125
No log 2.16 216 0.6508 0.6931 0.6508 0.8067
No log 2.18 218 0.7661 0.6184 0.7661 0.8753
No log 2.2 220 0.8442 0.6372 0.8442 0.9188
No log 2.22 222 0.7886 0.6293 0.7886 0.8881
No log 2.24 224 0.7441 0.6242 0.7441 0.8626
No log 2.26 226 0.6953 0.6434 0.6953 0.8338
No log 2.2800 228 0.6363 0.6468 0.6363 0.7977
No log 2.3 230 0.6166 0.7231 0.6166 0.7853
No log 2.32 232 0.6171 0.7078 0.6171 0.7855
No log 2.34 234 0.6238 0.6895 0.6238 0.7898
No log 2.36 236 0.6721 0.6835 0.6721 0.8198
No log 2.38 238 0.6685 0.6503 0.6685 0.8176
No log 2.4 240 0.6116 0.7083 0.6116 0.7820
No log 2.42 242 0.6602 0.6858 0.6602 0.8125
No log 2.44 244 0.7933 0.5435 0.7933 0.8906
No log 2.46 246 0.7425 0.6292 0.7425 0.8617
No log 2.48 248 0.6258 0.6550 0.6258 0.7911
No log 2.5 250 0.6493 0.7265 0.6493 0.8058
No log 2.52 252 0.7391 0.6169 0.7391 0.8597
No log 2.54 254 0.7054 0.6128 0.7054 0.8399
No log 2.56 256 0.7070 0.4947 0.7070 0.8408
No log 2.58 258 0.6992 0.5328 0.6992 0.8362
No log 2.6 260 0.6279 0.6207 0.6279 0.7924
No log 2.62 262 0.6140 0.6446 0.6140 0.7836
No log 2.64 264 0.6073 0.6269 0.6073 0.7793
No log 2.66 266 0.6193 0.6804 0.6193 0.7870
No log 2.68 268 0.6929 0.6439 0.6929 0.8324
No log 2.7 270 0.6986 0.6366 0.6986 0.8358
No log 2.7200 272 0.6895 0.6134 0.6895 0.8304
No log 2.74 274 0.7109 0.6396 0.7109 0.8431
No log 2.76 276 0.7565 0.5422 0.7565 0.8698
No log 2.7800 278 0.7038 0.5799 0.7038 0.8389
No log 2.8 280 0.6519 0.5033 0.6519 0.8074
No log 2.82 282 0.7034 0.4829 0.7034 0.8387
No log 2.84 284 0.7852 0.4522 0.7852 0.8861
No log 2.86 286 0.7544 0.5370 0.7544 0.8685
No log 2.88 288 0.6584 0.5446 0.6584 0.8114
No log 2.9 290 0.6224 0.6291 0.6224 0.7889
No log 2.92 292 0.6976 0.6266 0.6976 0.8352
No log 2.94 294 0.7233 0.5862 0.7233 0.8505
No log 2.96 296 0.6753 0.6147 0.6753 0.8218
No log 2.98 298 0.6785 0.6383 0.6785 0.8237
No log 3.0 300 0.6966 0.6068 0.6966 0.8346
No log 3.02 302 0.7041 0.6263 0.7041 0.8391
No log 3.04 304 0.7091 0.5869 0.7091 0.8421
No log 3.06 306 0.6919 0.6224 0.6919 0.8318
No log 3.08 308 0.6921 0.5949 0.6921 0.8319
No log 3.1 310 0.6632 0.6319 0.6632 0.8143
No log 3.12 312 0.6423 0.6659 0.6423 0.8014
No log 3.14 314 0.6607 0.6080 0.6607 0.8128
No log 3.16 316 0.7140 0.5135 0.7140 0.8450
No log 3.18 318 0.6991 0.5235 0.6991 0.8361
No log 3.2 320 0.6309 0.6578 0.6309 0.7943
No log 3.22 322 0.6588 0.6236 0.6588 0.8116
No log 3.24 324 0.7391 0.6235 0.7391 0.8597
No log 3.26 326 0.7233 0.6145 0.7233 0.8505
No log 3.2800 328 0.6441 0.6708 0.6441 0.8025
No log 3.3 330 0.6463 0.6448 0.6463 0.8039
No log 3.32 332 0.6861 0.5636 0.6861 0.8283
No log 3.34 334 0.6770 0.5726 0.6770 0.8228
No log 3.36 336 0.6589 0.5979 0.6589 0.8117
No log 3.38 338 0.6738 0.5844 0.6738 0.8209
No log 3.4 340 0.6905 0.6035 0.6905 0.8310
No log 3.42 342 0.7076 0.5833 0.7076 0.8412
No log 3.44 344 0.6866 0.5992 0.6866 0.8286
No log 3.46 346 0.6400 0.5220 0.6400 0.8000
No log 3.48 348 0.6099 0.6602 0.6099 0.7809
No log 3.5 350 0.6230 0.6354 0.6230 0.7893
No log 3.52 352 0.6320 0.6457 0.6320 0.7950
No log 3.54 354 0.6186 0.6448 0.6186 0.7865
No log 3.56 356 0.6042 0.7126 0.6042 0.7773
No log 3.58 358 0.6017 0.7019 0.6017 0.7757
No log 3.6 360 0.6067 0.6278 0.6067 0.7789
No log 3.62 362 0.5895 0.6207 0.5895 0.7678
No log 3.64 364 0.5981 0.6626 0.5981 0.7734
No log 3.66 366 0.6330 0.6207 0.6330 0.7956
No log 3.68 368 0.6464 0.6853 0.6464 0.8040
No log 3.7 370 0.6520 0.6693 0.6520 0.8075
No log 3.7200 372 0.6612 0.6001 0.6612 0.8132
No log 3.74 374 0.7033 0.6131 0.7033 0.8386
No log 3.76 376 0.7079 0.5999 0.7079 0.8414
No log 3.7800 378 0.7216 0.5776 0.7216 0.8495
No log 3.8 380 0.7540 0.6357 0.7540 0.8684
No log 3.82 382 0.7535 0.5885 0.7535 0.8680
No log 3.84 384 0.7223 0.5516 0.7223 0.8499
No log 3.86 386 0.6865 0.4490 0.6865 0.8285
No log 3.88 388 0.6798 0.4893 0.6798 0.8245
No log 3.9 390 0.6878 0.5438 0.6878 0.8293
No log 3.92 392 0.6914 0.5968 0.6914 0.8315
No log 3.94 394 0.6815 0.6419 0.6815 0.8255
No log 3.96 396 0.6534 0.6783 0.6534 0.8084
No log 3.98 398 0.6215 0.7020 0.6215 0.7883
No log 4.0 400 0.5993 0.7266 0.5993 0.7741
No log 4.02 402 0.6003 0.6605 0.6003 0.7748
No log 4.04 404 0.6167 0.6012 0.6167 0.7853
No log 4.06 406 0.6340 0.5522 0.6340 0.7962
No log 4.08 408 0.6527 0.5498 0.6527 0.8079
No log 4.1 410 0.6722 0.5498 0.6722 0.8199
No log 4.12 412 0.6851 0.6318 0.6851 0.8277
No log 4.14 414 0.6448 0.6972 0.6448 0.8030
No log 4.16 416 0.6201 0.6594 0.6201 0.7874
No log 4.18 418 0.6533 0.6277 0.6533 0.8083
No log 4.2 420 0.6759 0.6277 0.6759 0.8222
No log 4.22 422 0.6394 0.5933 0.6394 0.7996
No log 4.24 424 0.6371 0.6642 0.6371 0.7982
No log 4.26 426 0.6753 0.6612 0.6753 0.8218
No log 4.28 428 0.6685 0.5933 0.6685 0.8176
No log 4.3 430 0.6387 0.5810 0.6387 0.7992
No log 4.32 432 0.6352 0.5810 0.6352 0.7970
No log 4.34 434 0.6338 0.5688 0.6338 0.7961
No log 4.36 436 0.6207 0.5375 0.6207 0.7879
No log 4.38 438 0.6015 0.5386 0.6015 0.7755
No log 4.4 440 0.5928 0.5747 0.5928 0.7699
No log 4.42 442 0.5913 0.5785 0.5913 0.7690
No log 4.44 444 0.6366 0.5832 0.6366 0.7979
No log 4.46 446 0.7161 0.6071 0.7161 0.8462
No log 4.48 448 0.7361 0.5543 0.7361 0.8579
No log 4.5 450 0.6937 0.5614 0.6937 0.8329
No log 4.52 452 0.6869 0.5751 0.6869 0.8288
No log 4.54 454 0.6730 0.5648 0.6730 0.8204
No log 4.5600 456 0.6475 0.5648 0.6475 0.8047
No log 4.58 458 0.6339 0.5842 0.6339 0.7962
No log 4.6 460 0.6652 0.6328 0.6652 0.8156
No log 4.62 462 0.6509 0.5645 0.6509 0.8068
No log 4.64 464 0.6117 0.6186 0.6117 0.7821
No log 4.66 466 0.6158 0.5898 0.6158 0.7847
No log 4.68 468 0.6363 0.5910 0.6363 0.7977
No log 4.7 470 0.6270 0.6497 0.6270 0.7919
No log 4.72 472 0.6358 0.6733 0.6358 0.7974
No log 4.74 474 0.6656 0.6369 0.6656 0.8158
No log 4.76 476 0.6801 0.6369 0.6801 0.8247
No log 4.78 478 0.6791 0.6617 0.6791 0.8241
No log 4.8 480 0.6669 0.6330 0.6669 0.8166
No log 4.82 482 0.6578 0.5855 0.6578 0.8111
No log 4.84 484 0.6567 0.5678 0.6567 0.8104
No log 4.86 486 0.6336 0.5516 0.6336 0.7960
No log 4.88 488 0.6394 0.5316 0.6394 0.7996
No log 4.9 490 0.6312 0.5910 0.6312 0.7945
No log 4.92 492 0.6088 0.6584 0.6088 0.7802
No log 4.9400 494 0.6332 0.6944 0.6332 0.7957
No log 4.96 496 0.6749 0.7106 0.6749 0.8215
No log 4.98 498 0.6826 0.7124 0.6826 0.8262
0.2724 5.0 500 0.6940 0.7233 0.6940 0.8331
0.2724 5.02 502 0.6537 0.7022 0.6537 0.8085
0.2724 5.04 504 0.6377 0.7101 0.6377 0.7985
0.2724 5.06 506 0.6132 0.6286 0.6132 0.7831
0.2724 5.08 508 0.6222 0.6324 0.6222 0.7888
0.2724 5.1 510 0.6364 0.6176 0.6364 0.7978

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k20_task5_organization

Finetuned
(4019)
this model