ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k1_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6336
  • Qwk: 0.6226
  • Mse: 0.6336
  • Rmse: 0.7960

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.4 2 3.8265 0.0012 3.8265 1.9561
No log 0.8 4 2.4916 0.0700 2.4916 1.5785
No log 1.2 6 1.2261 0.0731 1.2261 1.1073
No log 1.6 8 1.1498 0.0613 1.1498 1.0723
No log 2.0 10 1.0758 0.2834 1.0758 1.0372
No log 2.4 12 1.0712 0.2243 1.0712 1.0350
No log 2.8 14 1.0568 0.2391 1.0568 1.0280
No log 3.2 16 0.9427 0.2818 0.9427 0.9709
No log 3.6 18 0.8810 0.2944 0.8810 0.9386
No log 4.0 20 0.8305 0.4035 0.8305 0.9113
No log 4.4 22 0.8123 0.4596 0.8123 0.9013
No log 4.8 24 0.8435 0.5429 0.8435 0.9184
No log 5.2 26 0.8340 0.5873 0.8340 0.9132
No log 5.6 28 0.8426 0.5458 0.8426 0.9179
No log 6.0 30 0.8510 0.5614 0.8510 0.9225
No log 6.4 32 1.1040 0.3350 1.1040 1.0507
No log 6.8 34 1.0128 0.3889 1.0128 1.0064
No log 7.2 36 0.8260 0.5984 0.8260 0.9088
No log 7.6 38 0.9052 0.5626 0.9052 0.9514
No log 8.0 40 0.8251 0.5428 0.8251 0.9083
No log 8.4 42 0.8081 0.5363 0.8081 0.8989
No log 8.8 44 0.8266 0.4720 0.8266 0.9092
No log 9.2 46 0.7457 0.5902 0.7457 0.8635
No log 9.6 48 0.8654 0.4712 0.8654 0.9303
No log 10.0 50 0.8696 0.4812 0.8696 0.9325
No log 10.4 52 0.7484 0.6390 0.7484 0.8651
No log 10.8 54 0.7600 0.5830 0.7600 0.8718
No log 11.2 56 0.7636 0.6022 0.7636 0.8738
No log 11.6 58 0.7732 0.6094 0.7732 0.8793
No log 12.0 60 0.7734 0.6094 0.7734 0.8795
No log 12.4 62 0.7716 0.6178 0.7716 0.8784
No log 12.8 64 0.7671 0.5977 0.7671 0.8759
No log 13.2 66 0.7719 0.5710 0.7719 0.8786
No log 13.6 68 0.7808 0.6303 0.7808 0.8836
No log 14.0 70 0.8536 0.6271 0.8536 0.9239
No log 14.4 72 0.7786 0.6102 0.7786 0.8824
No log 14.8 74 0.7485 0.6162 0.7485 0.8652
No log 15.2 76 0.7077 0.6197 0.7077 0.8413
No log 15.6 78 0.7118 0.5966 0.7118 0.8437
No log 16.0 80 0.7881 0.5118 0.7881 0.8877
No log 16.4 82 0.8037 0.5309 0.8037 0.8965
No log 16.8 84 0.7121 0.6284 0.7121 0.8438
No log 17.2 86 0.7806 0.5458 0.7806 0.8835
No log 17.6 88 0.7858 0.6064 0.7858 0.8865
No log 18.0 90 0.7166 0.6197 0.7166 0.8465
No log 18.4 92 0.7128 0.5391 0.7128 0.8442
No log 18.8 94 0.7647 0.5427 0.7647 0.8745
No log 19.2 96 0.7030 0.5830 0.7030 0.8384
No log 19.6 98 0.7174 0.6451 0.7174 0.8470
No log 20.0 100 0.7716 0.6004 0.7716 0.8784
No log 20.4 102 0.7163 0.6160 0.7163 0.8464
No log 20.8 104 0.6953 0.6624 0.6953 0.8339
No log 21.2 106 0.7547 0.5554 0.7547 0.8688
No log 21.6 108 0.7634 0.6064 0.7634 0.8737
No log 22.0 110 0.7120 0.6289 0.7120 0.8438
No log 22.4 112 0.7230 0.5987 0.7230 0.8503
No log 22.8 114 0.6869 0.6032 0.6869 0.8288
No log 23.2 116 0.7046 0.5622 0.7046 0.8394
No log 23.6 118 0.7582 0.5233 0.7582 0.8708
No log 24.0 120 0.6800 0.5817 0.6800 0.8246
No log 24.4 122 0.6739 0.6324 0.6739 0.8209
No log 24.8 124 0.7167 0.6284 0.7167 0.8466
No log 25.2 126 0.6910 0.6388 0.6910 0.8312
No log 25.6 128 0.6338 0.6545 0.6338 0.7961
No log 26.0 130 0.6409 0.6447 0.6409 0.8006
No log 26.4 132 0.6742 0.6094 0.6742 0.8211
No log 26.8 134 0.6915 0.6420 0.6915 0.8315
No log 27.2 136 0.6935 0.6343 0.6935 0.8327
No log 27.6 138 0.6730 0.6554 0.6730 0.8204
No log 28.0 140 0.6424 0.6102 0.6424 0.8015
No log 28.4 142 0.6303 0.6414 0.6303 0.7939
No log 28.8 144 0.6482 0.6526 0.6482 0.8051
No log 29.2 146 0.6496 0.6148 0.6496 0.8060
No log 29.6 148 0.6461 0.6138 0.6461 0.8038
No log 30.0 150 0.6639 0.6310 0.6639 0.8148
No log 30.4 152 0.6814 0.5972 0.6814 0.8255
No log 30.8 154 0.6813 0.6197 0.6813 0.8254
No log 31.2 156 0.6678 0.6244 0.6678 0.8172
No log 31.6 158 0.6807 0.5805 0.6807 0.8250
No log 32.0 160 0.6801 0.5763 0.6801 0.8247
No log 32.4 162 0.6653 0.6310 0.6653 0.8157
No log 32.8 164 0.6806 0.6406 0.6806 0.8250
No log 33.2 166 0.6998 0.6536 0.6998 0.8366
No log 33.6 168 0.6774 0.6420 0.6774 0.8231
No log 34.0 170 0.6722 0.5883 0.6722 0.8199
No log 34.4 172 0.6865 0.6098 0.6865 0.8286
No log 34.8 174 0.6520 0.6214 0.6520 0.8075
No log 35.2 176 0.6345 0.6226 0.6345 0.7966
No log 35.6 178 0.6637 0.6197 0.6637 0.8147
No log 36.0 180 0.6622 0.6365 0.6622 0.8138
No log 36.4 182 0.6517 0.6197 0.6517 0.8073
No log 36.8 184 0.6383 0.6301 0.6383 0.7989
No log 37.2 186 0.6266 0.6157 0.6266 0.7916
No log 37.6 188 0.6272 0.6157 0.6272 0.7920
No log 38.0 190 0.6210 0.6297 0.6210 0.7880
No log 38.4 192 0.6187 0.6186 0.6187 0.7866
No log 38.8 194 0.6251 0.6333 0.6251 0.7906
No log 39.2 196 0.6626 0.6226 0.6626 0.8140
No log 39.6 198 0.6947 0.6528 0.6947 0.8335
No log 40.0 200 0.6666 0.6365 0.6666 0.8165
No log 40.4 202 0.6236 0.6510 0.6236 0.7897
No log 40.8 204 0.6311 0.6302 0.6311 0.7944
No log 41.2 206 0.6518 0.6223 0.6518 0.8073
No log 41.6 208 0.6255 0.6473 0.6255 0.7909
No log 42.0 210 0.6147 0.6493 0.6147 0.7840
No log 42.4 212 0.6215 0.6319 0.6215 0.7884
No log 42.8 214 0.6308 0.5935 0.6308 0.7943
No log 43.2 216 0.6375 0.5838 0.6375 0.7985
No log 43.6 218 0.6400 0.5935 0.6400 0.8000
No log 44.0 220 0.6428 0.5838 0.6428 0.8017
No log 44.4 222 0.6515 0.5838 0.6515 0.8071
No log 44.8 224 0.6540 0.5935 0.6540 0.8087
No log 45.2 226 0.6598 0.5673 0.6598 0.8123
No log 45.6 228 0.6773 0.5561 0.6773 0.8230
No log 46.0 230 0.6793 0.6005 0.6793 0.8242
No log 46.4 232 0.6689 0.6249 0.6689 0.8178
No log 46.8 234 0.6763 0.6075 0.6763 0.8224
No log 47.2 236 0.6681 0.6397 0.6681 0.8174
No log 47.6 238 0.6466 0.6226 0.6466 0.8041
No log 48.0 240 0.6234 0.6226 0.6234 0.7896
No log 48.4 242 0.6151 0.6275 0.6151 0.7843
No log 48.8 244 0.6213 0.6383 0.6213 0.7882
No log 49.2 246 0.6232 0.6383 0.6232 0.7894
No log 49.6 248 0.6198 0.6244 0.6198 0.7873
No log 50.0 250 0.6217 0.6406 0.6217 0.7885
No log 50.4 252 0.6243 0.6226 0.6243 0.7901
No log 50.8 254 0.6241 0.6226 0.6241 0.7900
No log 51.2 256 0.6210 0.6406 0.6210 0.7881
No log 51.6 258 0.6185 0.6120 0.6185 0.7865
No log 52.0 260 0.6153 0.6120 0.6153 0.7844
No log 52.4 262 0.6110 0.6224 0.6110 0.7817
No log 52.8 264 0.6052 0.6224 0.6052 0.7780
No log 53.2 266 0.6045 0.6129 0.6045 0.7775
No log 53.6 268 0.6100 0.6391 0.6100 0.7810
No log 54.0 270 0.6204 0.6391 0.6204 0.7877
No log 54.4 272 0.6286 0.6391 0.6286 0.7928
No log 54.8 274 0.6342 0.6391 0.6342 0.7964
No log 55.2 276 0.6398 0.6239 0.6398 0.7999
No log 55.6 278 0.6373 0.6239 0.6373 0.7983
No log 56.0 280 0.6376 0.6239 0.6376 0.7985
No log 56.4 282 0.6282 0.6484 0.6282 0.7926
No log 56.8 284 0.6309 0.6226 0.6309 0.7943
No log 57.2 286 0.6459 0.6365 0.6459 0.8037
No log 57.6 288 0.6423 0.6365 0.6423 0.8014
No log 58.0 290 0.6313 0.6197 0.6313 0.7946
No log 58.4 292 0.6286 0.6365 0.6286 0.7929
No log 58.8 294 0.6254 0.6365 0.6254 0.7908
No log 59.2 296 0.6157 0.6333 0.6157 0.7847
No log 59.6 298 0.6141 0.6266 0.6141 0.7837
No log 60.0 300 0.6151 0.6447 0.6151 0.7843
No log 60.4 302 0.6158 0.6266 0.6158 0.7848
No log 60.8 304 0.6218 0.6333 0.6218 0.7885
No log 61.2 306 0.6340 0.6468 0.6340 0.7963
No log 61.6 308 0.6488 0.6335 0.6488 0.8055
No log 62.0 310 0.6634 0.6365 0.6634 0.8145
No log 62.4 312 0.6805 0.6459 0.6805 0.8249
No log 62.8 314 0.6796 0.6493 0.6796 0.8243
No log 63.2 316 0.6635 0.6493 0.6635 0.8146
No log 63.6 318 0.6499 0.6102 0.6499 0.8062
No log 64.0 320 0.6486 0.6102 0.6486 0.8054
No log 64.4 322 0.6500 0.6102 0.6500 0.8062
No log 64.8 324 0.6644 0.6365 0.6644 0.8151
No log 65.2 326 0.6795 0.6459 0.6795 0.8243
No log 65.6 328 0.6771 0.6459 0.6771 0.8228
No log 66.0 330 0.6600 0.6365 0.6600 0.8124
No log 66.4 332 0.6430 0.6226 0.6430 0.8019
No log 66.8 334 0.6351 0.6226 0.6351 0.7969
No log 67.2 336 0.6258 0.5925 0.6258 0.7911
No log 67.6 338 0.6213 0.5959 0.6213 0.7882
No log 68.0 340 0.6179 0.6068 0.6179 0.7861
No log 68.4 342 0.6143 0.6341 0.6143 0.7838
No log 68.8 344 0.6164 0.6333 0.6164 0.7851
No log 69.2 346 0.6190 0.6272 0.6190 0.7868
No log 69.6 348 0.6398 0.6169 0.6398 0.7999
No log 70.0 350 0.6556 0.6335 0.6556 0.8097
No log 70.4 352 0.6550 0.6365 0.6550 0.8093
No log 70.8 354 0.6342 0.6226 0.6342 0.7963
No log 71.2 356 0.6121 0.6406 0.6121 0.7824
No log 71.6 358 0.6086 0.6254 0.6086 0.7801
No log 72.0 360 0.6122 0.6078 0.6122 0.7824
No log 72.4 362 0.6135 0.6078 0.6135 0.7832
No log 72.8 364 0.6159 0.6120 0.6159 0.7848
No log 73.2 366 0.6213 0.6120 0.6213 0.7882
No log 73.6 368 0.6267 0.6510 0.6267 0.7917
No log 74.0 370 0.6343 0.6226 0.6343 0.7965
No log 74.4 372 0.6381 0.6226 0.6381 0.7988
No log 74.8 374 0.6366 0.6226 0.6366 0.7979
No log 75.2 376 0.6415 0.6324 0.6415 0.8010
No log 75.6 378 0.6365 0.6324 0.6365 0.7978
No log 76.0 380 0.6338 0.6324 0.6338 0.7961
No log 76.4 382 0.6293 0.6226 0.6293 0.7933
No log 76.8 384 0.6227 0.6226 0.6227 0.7891
No log 77.2 386 0.6184 0.6058 0.6184 0.7864
No log 77.6 388 0.6176 0.5959 0.6176 0.7859
No log 78.0 390 0.6185 0.5959 0.6185 0.7864
No log 78.4 392 0.6200 0.6032 0.6200 0.7874
No log 78.8 394 0.6212 0.6226 0.6212 0.7881
No log 79.2 396 0.6242 0.6226 0.6242 0.7901
No log 79.6 398 0.6257 0.6226 0.6257 0.7910
No log 80.0 400 0.6276 0.6226 0.6276 0.7922
No log 80.4 402 0.6281 0.6226 0.6281 0.7925
No log 80.8 404 0.6268 0.6226 0.6268 0.7917
No log 81.2 406 0.6271 0.6157 0.6271 0.7919
No log 81.6 408 0.6308 0.6310 0.6308 0.7942
No log 82.0 410 0.6326 0.6016 0.6326 0.7953
No log 82.4 412 0.6302 0.6310 0.6302 0.7939
No log 82.8 414 0.6283 0.6341 0.6283 0.7927
No log 83.2 416 0.6259 0.6157 0.6259 0.7911
No log 83.6 418 0.6238 0.6157 0.6238 0.7898
No log 84.0 420 0.6213 0.6333 0.6213 0.7882
No log 84.4 422 0.6208 0.6333 0.6208 0.7879
No log 84.8 424 0.6219 0.6333 0.6219 0.7886
No log 85.2 426 0.6244 0.6226 0.6244 0.7902
No log 85.6 428 0.6261 0.6226 0.6261 0.7913
No log 86.0 430 0.6262 0.6226 0.6262 0.7914
No log 86.4 432 0.6249 0.6226 0.6249 0.7905
No log 86.8 434 0.6227 0.6256 0.6227 0.7891
No log 87.2 436 0.6219 0.6256 0.6219 0.7886
No log 87.6 438 0.6208 0.6256 0.6208 0.7879
No log 88.0 440 0.6219 0.6256 0.6219 0.7886
No log 88.4 442 0.6238 0.6256 0.6238 0.7898
No log 88.8 444 0.6248 0.6226 0.6248 0.7904
No log 89.2 446 0.6242 0.6256 0.6242 0.7901
No log 89.6 448 0.6225 0.6256 0.6225 0.7890
No log 90.0 450 0.6216 0.6256 0.6216 0.7884
No log 90.4 452 0.6203 0.6256 0.6203 0.7876
No log 90.8 454 0.6207 0.6256 0.6207 0.7879
No log 91.2 456 0.6223 0.6256 0.6223 0.7888
No log 91.6 458 0.6259 0.6256 0.6259 0.7911
No log 92.0 460 0.6295 0.6324 0.6295 0.7934
No log 92.4 462 0.6336 0.6324 0.6336 0.7960
No log 92.8 464 0.6356 0.6324 0.6356 0.7973
No log 93.2 466 0.6358 0.6324 0.6358 0.7974
No log 93.6 468 0.6357 0.6324 0.6357 0.7973
No log 94.0 470 0.6373 0.6324 0.6373 0.7983
No log 94.4 472 0.6390 0.6324 0.6390 0.7994
No log 94.8 474 0.6406 0.6324 0.6406 0.8004
No log 95.2 476 0.6414 0.6324 0.6414 0.8009
No log 95.6 478 0.6418 0.6324 0.6418 0.8011
No log 96.0 480 0.6418 0.6324 0.6418 0.8011
No log 96.4 482 0.6422 0.6324 0.6422 0.8014
No log 96.8 484 0.6412 0.6324 0.6412 0.8007
No log 97.2 486 0.6397 0.6324 0.6397 0.7998
No log 97.6 488 0.6385 0.6324 0.6385 0.7991
No log 98.0 490 0.6370 0.6324 0.6370 0.7981
No log 98.4 492 0.6359 0.6324 0.6359 0.7974
No log 98.8 494 0.6349 0.6226 0.6349 0.7968
No log 99.2 496 0.6341 0.6226 0.6341 0.7963
No log 99.6 498 0.6337 0.6226 0.6337 0.7961
0.1645 100.0 500 0.6336 0.6226 0.6336 0.7960

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k1_task5_organization

Finetuned
(4019)
this model