ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k7_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9980
  • Qwk: 0.4796
  • Mse: 0.9980
  • Rmse: 0.9990

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.05 2 4.5135 0.0163 4.5135 2.1245
No log 0.1 4 2.6317 -0.0334 2.6317 1.6222
No log 0.15 6 1.6442 0.0504 1.6442 1.2823
No log 0.2 8 1.3053 0.0205 1.3053 1.1425
No log 0.25 10 1.1061 0.2097 1.1061 1.0517
No log 0.3 12 1.0522 0.1857 1.0522 1.0258
No log 0.35 14 1.0003 0.2674 1.0003 1.0002
No log 0.4 16 1.0051 0.3762 1.0051 1.0025
No log 0.45 18 1.2197 0.2252 1.2197 1.1044
No log 0.5 20 1.2587 0.2605 1.2587 1.1219
No log 0.55 22 1.0760 0.3033 1.0760 1.0373
No log 0.6 24 0.9315 0.3493 0.9315 0.9651
No log 0.65 26 1.0116 0.3141 1.0116 1.0058
No log 0.7 28 1.0780 0.3870 1.0780 1.0383
No log 0.75 30 0.9534 0.3584 0.9534 0.9764
No log 0.8 32 1.0114 0.3869 1.0114 1.0057
No log 0.85 34 1.1168 0.2788 1.1168 1.0568
No log 0.9 36 1.4125 0.2568 1.4125 1.1885
No log 0.95 38 1.3935 0.2525 1.3935 1.1804
No log 1.0 40 1.0109 0.3881 1.0109 1.0054
No log 1.05 42 0.8605 0.4671 0.8605 0.9276
No log 1.1 44 0.8279 0.4321 0.8279 0.9099
No log 1.15 46 0.8544 0.3868 0.8544 0.9244
No log 1.2 48 0.9025 0.3590 0.9025 0.9500
No log 1.25 50 0.8987 0.4261 0.8987 0.9480
No log 1.3 52 0.8974 0.4342 0.8974 0.9473
No log 1.35 54 0.8953 0.5156 0.8953 0.9462
No log 1.4 56 0.9231 0.4572 0.9231 0.9608
No log 1.45 58 0.9496 0.4650 0.9496 0.9745
No log 1.5 60 0.8830 0.4720 0.8830 0.9397
No log 1.55 62 0.7900 0.4181 0.7900 0.8888
No log 1.6 64 0.7739 0.5446 0.7739 0.8797
No log 1.65 66 0.7782 0.5062 0.7782 0.8821
No log 1.7 68 0.8386 0.4161 0.8386 0.9157
No log 1.75 70 0.9037 0.4665 0.9037 0.9506
No log 1.8 72 0.9095 0.5495 0.9095 0.9537
No log 1.85 74 0.8366 0.5578 0.8366 0.9146
No log 1.9 76 0.7904 0.4668 0.7904 0.8891
No log 1.95 78 0.7965 0.5260 0.7965 0.8925
No log 2.0 80 0.8303 0.5169 0.8303 0.9112
No log 2.05 82 0.8924 0.5814 0.8924 0.9447
No log 2.1 84 0.9846 0.6217 0.9846 0.9923
No log 2.15 86 1.0022 0.5448 1.0022 1.0011
No log 2.2 88 0.8934 0.5445 0.8934 0.9452
No log 2.25 90 0.7591 0.5352 0.7591 0.8713
No log 2.3 92 0.7563 0.4738 0.7563 0.8697
No log 2.35 94 0.8193 0.4197 0.8193 0.9051
No log 2.4 96 0.9202 0.4301 0.9202 0.9593
No log 2.45 98 0.9371 0.4449 0.9371 0.9680
No log 2.5 100 0.8676 0.5276 0.8676 0.9314
No log 2.55 102 0.7686 0.5335 0.7686 0.8767
No log 2.6 104 0.7571 0.5706 0.7571 0.8701
No log 2.65 106 0.8149 0.5391 0.8149 0.9027
No log 2.7 108 0.8190 0.6289 0.8190 0.9050
No log 2.75 110 0.7706 0.6333 0.7706 0.8778
No log 2.8 112 0.7676 0.6414 0.7676 0.8761
No log 2.85 114 0.8095 0.6489 0.8095 0.8997
No log 2.9 116 0.8182 0.5598 0.8182 0.9046
No log 2.95 118 0.8006 0.5050 0.8006 0.8947
No log 3.0 120 0.7455 0.5997 0.7455 0.8635
No log 3.05 122 0.7090 0.6346 0.7090 0.8420
No log 3.1 124 0.6940 0.6722 0.6940 0.8331
No log 3.15 126 0.7232 0.6186 0.7232 0.8504
No log 3.2 128 0.8225 0.5594 0.8225 0.9069
No log 3.25 130 0.9330 0.5768 0.9330 0.9659
No log 3.3 132 1.0429 0.5268 1.0429 1.0212
No log 3.35 134 1.0175 0.4817 1.0175 1.0087
No log 3.4 136 0.9261 0.4722 0.9261 0.9623
No log 3.45 138 0.9004 0.4539 0.9004 0.9489
No log 3.5 140 0.9271 0.4065 0.9271 0.9628
No log 3.55 142 0.9551 0.4236 0.9551 0.9773
No log 3.6 144 0.9514 0.4449 0.9514 0.9754
No log 3.65 146 0.9157 0.4628 0.9157 0.9569
No log 3.7 148 0.8584 0.4554 0.8584 0.9265
No log 3.75 150 0.8132 0.5349 0.8132 0.9018
No log 3.8 152 0.7974 0.5735 0.7974 0.8929
No log 3.85 154 0.8276 0.5189 0.8276 0.9097
No log 3.9 156 0.8816 0.5492 0.8816 0.9389
No log 3.95 158 0.9117 0.5733 0.9117 0.9548
No log 4.0 160 0.9223 0.5857 0.9223 0.9604
No log 4.05 162 0.8674 0.5079 0.8674 0.9314
No log 4.1 164 0.8181 0.4597 0.8181 0.9045
No log 4.15 166 0.8458 0.5192 0.8458 0.9197
No log 4.2 168 0.9756 0.5528 0.9756 0.9877
No log 4.25 170 1.1385 0.5315 1.1385 1.0670
No log 4.3 172 1.0891 0.5732 1.0891 1.0436
No log 4.35 174 0.9656 0.5573 0.9656 0.9826
No log 4.4 176 0.8721 0.5136 0.8721 0.9339
No log 4.45 178 0.8232 0.5179 0.8232 0.9073
No log 4.5 180 0.8042 0.4953 0.8042 0.8968
No log 4.55 182 0.8111 0.5864 0.8111 0.9006
No log 4.6 184 0.8979 0.5249 0.8979 0.9476
No log 4.65 186 1.0480 0.5095 1.0480 1.0237
No log 4.7 188 1.0936 0.5059 1.0936 1.0457
No log 4.75 190 1.0198 0.5222 1.0198 1.0098
No log 4.8 192 0.8737 0.5458 0.8737 0.9347
No log 4.85 194 0.7969 0.6064 0.7969 0.8927
No log 4.9 196 0.7675 0.6302 0.7675 0.8761
No log 4.95 198 0.7821 0.6064 0.7821 0.8844
No log 5.0 200 0.8474 0.5424 0.8474 0.9205
No log 5.05 202 0.8424 0.5172 0.8424 0.9178
No log 5.1 204 0.7927 0.5836 0.7927 0.8903
No log 5.15 206 0.8196 0.4852 0.8196 0.9053
No log 5.2 208 0.8544 0.5358 0.8544 0.9244
No log 5.25 210 0.8638 0.5370 0.8638 0.9294
No log 5.3 212 0.8601 0.5724 0.8601 0.9274
No log 5.35 214 0.8420 0.5814 0.8420 0.9176
No log 5.4 216 0.8117 0.6004 0.8117 0.9009
No log 5.45 218 0.7866 0.5304 0.7866 0.8869
No log 5.5 220 0.7801 0.5169 0.7801 0.8833
No log 5.55 222 0.7971 0.5712 0.7971 0.8928
No log 5.6 224 0.8003 0.5675 0.8003 0.8946
No log 5.65 226 0.8134 0.5855 0.8134 0.9019
No log 5.7 228 0.8578 0.6026 0.8578 0.9262
No log 5.75 230 0.9347 0.4866 0.9347 0.9668
No log 5.8 232 1.0064 0.5867 1.0064 1.0032
No log 5.85 234 0.9863 0.5665 0.9863 0.9931
No log 5.9 236 0.8972 0.6276 0.8972 0.9472
No log 5.95 238 0.8151 0.6696 0.8151 0.9028
No log 6.0 240 0.7956 0.6427 0.7956 0.8919
No log 6.05 242 0.7939 0.6109 0.7939 0.8910
No log 6.1 244 0.7913 0.6235 0.7913 0.8895
No log 6.15 246 0.8155 0.5986 0.8155 0.9030
No log 6.2 248 0.8311 0.6167 0.8311 0.9116
No log 6.25 250 0.7883 0.6179 0.7883 0.8879
No log 6.3 252 0.7684 0.6098 0.7684 0.8766
No log 6.35 254 0.7527 0.5944 0.7527 0.8676
No log 6.4 256 0.7438 0.6294 0.7438 0.8625
No log 6.45 258 0.7798 0.5611 0.7798 0.8831
No log 6.5 260 0.8794 0.5772 0.8794 0.9378
No log 6.55 262 0.8791 0.5370 0.8791 0.9376
No log 6.6 264 0.8730 0.5293 0.8730 0.9343
No log 6.65 266 0.9087 0.4832 0.9087 0.9533
No log 6.7 268 0.8773 0.5786 0.8773 0.9367
No log 6.75 270 0.8563 0.5921 0.8563 0.9254
No log 6.8 272 0.8070 0.6071 0.8070 0.8983
No log 6.85 274 0.7720 0.6598 0.7720 0.8786
No log 6.9 276 0.7752 0.6677 0.7752 0.8804
No log 6.95 278 0.8004 0.6481 0.8004 0.8947
No log 7.0 280 0.8643 0.6458 0.8643 0.9297
No log 7.05 282 0.8689 0.6468 0.8689 0.9321
No log 7.1 284 0.8486 0.5997 0.8486 0.9212
No log 7.15 286 0.8080 0.5479 0.8080 0.8989
No log 7.2 288 0.8267 0.5406 0.8267 0.9092
No log 7.25 290 0.8626 0.5091 0.8626 0.9288
No log 7.3 292 0.9083 0.5476 0.9083 0.9531
No log 7.35 294 0.8812 0.5669 0.8812 0.9387
No log 7.4 296 0.8140 0.5679 0.8140 0.9022
No log 7.45 298 0.7545 0.6308 0.7545 0.8686
No log 7.5 300 0.7792 0.5853 0.7792 0.8827
No log 7.55 302 0.8169 0.5750 0.8169 0.9038
No log 7.6 304 0.8356 0.5728 0.8356 0.9141
No log 7.65 306 0.8794 0.5399 0.8794 0.9378
No log 7.7 308 0.8861 0.5294 0.8861 0.9413
No log 7.75 310 0.9034 0.4473 0.9034 0.9505
No log 7.8 312 0.9324 0.4440 0.9324 0.9656
No log 7.85 314 1.0133 0.3838 1.0133 1.0067
No log 7.9 316 1.0283 0.4444 1.0283 1.0140
No log 7.95 318 0.9611 0.4962 0.9611 0.9803
No log 8.0 320 0.8598 0.5291 0.8598 0.9272
No log 8.05 322 0.8250 0.5781 0.8250 0.9083
No log 8.1 324 0.8552 0.5836 0.8552 0.9248
No log 8.15 326 0.9553 0.5104 0.9553 0.9774
No log 8.2 328 1.0477 0.5399 1.0477 1.0236
No log 8.25 330 0.9927 0.5551 0.9927 0.9964
No log 8.3 332 0.8601 0.5634 0.8601 0.9274
No log 8.35 334 0.7989 0.5896 0.7989 0.8938
No log 8.4 336 0.7844 0.5711 0.7844 0.8856
No log 8.45 338 0.7714 0.5760 0.7714 0.8783
No log 8.5 340 0.7598 0.5455 0.7598 0.8717
No log 8.55 342 0.7511 0.6032 0.7511 0.8667
No log 8.6 344 0.7721 0.6098 0.7721 0.8787
No log 8.65 346 0.8762 0.5554 0.8762 0.9361
No log 8.7 348 0.9317 0.4884 0.9317 0.9653
No log 8.75 350 0.8685 0.4841 0.8685 0.9319
No log 8.8 352 0.8793 0.4924 0.8793 0.9377
No log 8.85 354 0.9818 0.4866 0.9818 0.9908
No log 8.9 356 1.0960 0.5049 1.0960 1.0469
No log 8.95 358 1.0533 0.4911 1.0533 1.0263
No log 9.0 360 0.9088 0.4751 0.9088 0.9533
No log 9.05 362 0.8320 0.5105 0.8320 0.9121
No log 9.1 364 0.8079 0.5053 0.8079 0.8988
No log 9.15 366 0.7882 0.5079 0.7882 0.8878
No log 9.2 368 0.7714 0.5276 0.7714 0.8783
No log 9.25 370 0.8026 0.5338 0.8026 0.8959
No log 9.3 372 0.8345 0.5733 0.8345 0.9135
No log 9.35 374 0.8613 0.5810 0.8613 0.9280
No log 9.4 376 0.8806 0.5925 0.8806 0.9384
No log 9.45 378 0.8372 0.6333 0.8372 0.9150
No log 9.5 380 0.8123 0.6334 0.8123 0.9013
No log 9.55 382 0.7868 0.6151 0.7868 0.8870
No log 9.6 384 0.7566 0.5545 0.7566 0.8698
No log 9.65 386 0.7628 0.5725 0.7628 0.8734
No log 9.7 388 0.8138 0.5556 0.8138 0.9021
No log 9.75 390 0.8348 0.5893 0.8348 0.9137
No log 9.8 392 0.7845 0.6017 0.7845 0.8857
No log 9.85 394 0.7582 0.5415 0.7582 0.8707
No log 9.9 396 0.7500 0.5965 0.7500 0.8660
No log 9.95 398 0.7481 0.6128 0.7481 0.8649
No log 10.0 400 0.7568 0.6179 0.7568 0.8699
No log 10.05 402 0.7657 0.6408 0.7657 0.8750
No log 10.1 404 0.7972 0.5700 0.7972 0.8928
No log 10.15 406 0.8368 0.5339 0.8368 0.9148
No log 10.2 408 0.8217 0.5326 0.8217 0.9065
No log 10.25 410 0.7890 0.5541 0.7890 0.8882
No log 10.3 412 0.7628 0.5779 0.7628 0.8734
No log 10.35 414 0.7587 0.6309 0.7587 0.8710
No log 10.4 416 0.7609 0.6193 0.7609 0.8723
No log 10.45 418 0.7746 0.6237 0.7746 0.8801
No log 10.5 420 0.7894 0.6144 0.7894 0.8885
No log 10.55 422 0.7692 0.5929 0.7692 0.8770
No log 10.6 424 0.7606 0.5666 0.7606 0.8721
No log 10.65 426 0.7492 0.5864 0.7492 0.8656
No log 10.7 428 0.7452 0.5859 0.7452 0.8633
No log 10.75 430 0.7624 0.6118 0.7624 0.8732
No log 10.8 432 0.7699 0.5886 0.7699 0.8774
No log 10.85 434 0.7611 0.6055 0.7611 0.8724
No log 10.9 436 0.7976 0.5557 0.7976 0.8931
No log 10.95 438 0.8389 0.5487 0.8389 0.9159
No log 11.0 440 0.8806 0.5374 0.8806 0.9384
No log 11.05 442 0.8955 0.5260 0.8955 0.9463
No log 11.1 444 0.9488 0.4918 0.9488 0.9740
No log 11.15 446 1.0367 0.4421 1.0367 1.0182
No log 11.2 448 1.0129 0.5080 1.0129 1.0064
No log 11.25 450 0.8985 0.5233 0.8985 0.9479
No log 11.3 452 0.8009 0.5401 0.8009 0.8949
No log 11.35 454 0.7840 0.5406 0.7840 0.8854
No log 11.4 456 0.8011 0.5235 0.8011 0.8950
No log 11.45 458 0.8655 0.5176 0.8655 0.9303
No log 11.5 460 0.8882 0.5328 0.8882 0.9424
No log 11.55 462 0.8293 0.5472 0.8293 0.9106
No log 11.6 464 0.7501 0.5406 0.7501 0.8661
No log 11.65 466 0.6988 0.6487 0.6988 0.8359
No log 11.7 468 0.6875 0.6487 0.6875 0.8292
No log 11.75 470 0.6850 0.6586 0.6850 0.8276
No log 11.8 472 0.6864 0.6399 0.6864 0.8285
No log 11.85 474 0.7313 0.5473 0.7313 0.8552
No log 11.9 476 0.8003 0.5385 0.8003 0.8946
No log 11.95 478 0.8647 0.5056 0.8647 0.9299
No log 12.0 480 0.9434 0.5121 0.9434 0.9713
No log 12.05 482 1.0537 0.5300 1.0537 1.0265
No log 12.1 484 1.0380 0.5234 1.0380 1.0188
No log 12.15 486 0.9595 0.5134 0.9595 0.9795
No log 12.2 488 0.9072 0.4567 0.9072 0.9525
No log 12.25 490 0.8873 0.4567 0.8873 0.9420
No log 12.3 492 0.8916 0.4567 0.8916 0.9443
No log 12.35 494 0.9224 0.5062 0.9224 0.9604
No log 12.4 496 0.8816 0.5171 0.8816 0.9389
No log 12.45 498 0.8194 0.5154 0.8194 0.9052
0.3281 12.5 500 0.7553 0.5291 0.7553 0.8691
0.3281 12.55 502 0.7266 0.5636 0.7266 0.8524
0.3281 12.6 504 0.7011 0.6266 0.7011 0.8373
0.3281 12.65 506 0.6819 0.6875 0.6819 0.8257
0.3281 12.7 508 0.7124 0.6578 0.7124 0.8440
0.3281 12.75 510 0.7623 0.6595 0.7623 0.8731
0.3281 12.8 512 0.7666 0.6149 0.7666 0.8755
0.3281 12.85 514 0.7342 0.5370 0.7342 0.8569
0.3281 12.9 516 0.7148 0.5585 0.7148 0.8455
0.3281 12.95 518 0.6702 0.6285 0.6702 0.8186
0.3281 13.0 520 0.6450 0.6548 0.6450 0.8031
0.3281 13.05 522 0.6424 0.6773 0.6424 0.8015
0.3281 13.1 524 0.6473 0.7059 0.6473 0.8046
0.3281 13.15 526 0.6461 0.7191 0.6461 0.8038
0.3281 13.2 528 0.6569 0.6611 0.6569 0.8105
0.3281 13.25 530 0.7060 0.6162 0.7060 0.8402
0.3281 13.3 532 0.7303 0.5515 0.7303 0.8546
0.3281 13.35 534 0.7667 0.5405 0.7667 0.8756
0.3281 13.4 536 0.8012 0.5719 0.8012 0.8951
0.3281 13.45 538 0.8547 0.5297 0.8547 0.9245
0.3281 13.5 540 0.8714 0.5355 0.8714 0.9335
0.3281 13.55 542 0.8747 0.4828 0.8747 0.9353
0.3281 13.6 544 0.9488 0.4400 0.9488 0.9740
0.3281 13.65 546 0.9847 0.4400 0.9847 0.9923
0.3281 13.7 548 1.0122 0.4776 1.0122 1.0061
0.3281 13.75 550 0.9980 0.4796 0.9980 0.9990

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
3
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k7_task2_organization

Finetuned
(4019)
this model