ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k2_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6078
  • Qwk: 0.5912
  • Mse: 0.6078
  • Rmse: 0.7796

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.2 2 4.0041 0.0125 4.0041 2.0010
No log 0.4 4 2.3890 0.0700 2.3890 1.5456
No log 0.6 6 2.0685 0.0529 2.0685 1.4382
No log 0.8 8 1.0276 0.2667 1.0276 1.0137
No log 1.0 10 1.0854 0.0 1.0854 1.0418
No log 1.2 12 1.1225 0.1203 1.1225 1.0595
No log 1.4 14 1.0795 0.1854 1.0795 1.0390
No log 1.6 16 1.0091 0.1516 1.0091 1.0045
No log 1.8 18 1.0169 0.3414 1.0169 1.0084
No log 2.0 20 1.1362 0.1498 1.1362 1.0659
No log 2.2 22 1.0793 0.3171 1.0793 1.0389
No log 2.4 24 0.9631 0.2991 0.9631 0.9814
No log 2.6 26 0.9329 0.2161 0.9329 0.9658
No log 2.8 28 0.8941 0.3243 0.8941 0.9456
No log 3.0 30 0.9716 0.3396 0.9716 0.9857
No log 3.2 32 1.0564 0.2903 1.0564 1.0278
No log 3.4 34 0.8501 0.3896 0.8501 0.9220
No log 3.6 36 0.8907 0.3902 0.8907 0.9438
No log 3.8 38 1.0249 0.2864 1.0249 1.0124
No log 4.0 40 0.9467 0.3236 0.9467 0.9730
No log 4.2 42 0.7391 0.3611 0.7391 0.8597
No log 4.4 44 0.6400 0.5302 0.6400 0.8000
No log 4.6 46 0.6539 0.4990 0.6539 0.8086
No log 4.8 48 0.5883 0.5529 0.5883 0.7670
No log 5.0 50 0.7191 0.5443 0.7191 0.8480
No log 5.2 52 1.0071 0.5240 1.0071 1.0035
No log 5.4 54 0.9462 0.5161 0.9462 0.9727
No log 5.6 56 0.8079 0.5604 0.8079 0.8988
No log 5.8 58 0.7307 0.5275 0.7307 0.8548
No log 6.0 60 0.7232 0.5403 0.7232 0.8504
No log 6.2 62 0.7065 0.5798 0.7065 0.8405
No log 6.4 64 0.7682 0.5789 0.7682 0.8765
No log 6.6 66 0.7939 0.5484 0.7939 0.8910
No log 6.8 68 0.9516 0.5240 0.9516 0.9755
No log 7.0 70 0.8435 0.5391 0.8435 0.9184
No log 7.2 72 0.5973 0.6438 0.5973 0.7728
No log 7.4 74 0.5808 0.6940 0.5808 0.7621
No log 7.6 76 0.6124 0.6940 0.6124 0.7826
No log 7.8 78 0.7086 0.5915 0.7086 0.8418
No log 8.0 80 0.6065 0.6642 0.6065 0.7788
No log 8.2 82 0.6488 0.5841 0.6488 0.8055
No log 8.4 84 0.6465 0.5610 0.6465 0.8041
No log 8.6 86 0.7015 0.5963 0.7015 0.8376
No log 8.8 88 1.0109 0.5119 1.0109 1.0054
No log 9.0 90 0.8778 0.5213 0.8778 0.9369
No log 9.2 92 0.6122 0.7057 0.6122 0.7824
No log 9.4 94 0.5614 0.7266 0.5614 0.7493
No log 9.6 96 0.6060 0.6482 0.6060 0.7785
No log 9.8 98 0.5511 0.6697 0.5511 0.7424
No log 10.0 100 0.6164 0.6493 0.6164 0.7851
No log 10.2 102 0.6386 0.5877 0.6386 0.7991
No log 10.4 104 0.5990 0.6567 0.5990 0.7740
No log 10.6 106 0.5637 0.6814 0.5637 0.7508
No log 10.8 108 0.5571 0.6814 0.5571 0.7464
No log 11.0 110 0.5632 0.6813 0.5632 0.7505
No log 11.2 112 0.5935 0.7015 0.5935 0.7704
No log 11.4 114 0.6798 0.6431 0.6798 0.8245
No log 11.6 116 0.6510 0.6409 0.6510 0.8069
No log 11.8 118 0.6527 0.6562 0.6527 0.8079
No log 12.0 120 0.8198 0.5438 0.8198 0.9054
No log 12.2 122 0.7968 0.5658 0.7968 0.8926
No log 12.4 124 0.6130 0.6377 0.6130 0.7829
No log 12.6 126 0.5649 0.5917 0.5649 0.7516
No log 12.8 128 0.5726 0.6021 0.5726 0.7567
No log 13.0 130 0.5325 0.5820 0.5325 0.7297
No log 13.2 132 0.6022 0.6157 0.6022 0.7760
No log 13.4 134 0.6432 0.6079 0.6432 0.8020
No log 13.6 136 0.5760 0.6687 0.5760 0.7589
No log 13.8 138 0.5556 0.6557 0.5556 0.7454
No log 14.0 140 0.5644 0.6307 0.5644 0.7512
No log 14.2 142 0.5775 0.6360 0.5775 0.7599
No log 14.4 144 0.6162 0.6262 0.6162 0.7850
No log 14.6 146 0.6544 0.5978 0.6544 0.8090
No log 14.8 148 0.6939 0.5788 0.6939 0.8330
No log 15.0 150 0.6318 0.5700 0.6318 0.7949
No log 15.2 152 0.6468 0.6099 0.6468 0.8042
No log 15.4 154 0.6110 0.6032 0.6110 0.7817
No log 15.6 156 0.6108 0.6822 0.6108 0.7815
No log 15.8 158 0.8116 0.5759 0.8116 0.9009
No log 16.0 160 0.9348 0.5427 0.9348 0.9669
No log 16.2 162 0.8022 0.5465 0.8022 0.8957
No log 16.4 164 0.6058 0.6567 0.6058 0.7783
No log 16.6 166 0.5846 0.6916 0.5846 0.7646
No log 16.8 168 0.6259 0.6240 0.6259 0.7911
No log 17.0 170 0.7298 0.5877 0.7298 0.8543
No log 17.2 172 0.7314 0.6061 0.7314 0.8552
No log 17.4 174 0.6482 0.6109 0.6482 0.8051
No log 17.6 176 0.5703 0.6969 0.5703 0.7552
No log 17.8 178 0.5686 0.6882 0.5686 0.7540
No log 18.0 180 0.5864 0.7094 0.5864 0.7658
No log 18.2 182 0.6566 0.6091 0.6566 0.8103
No log 18.4 184 0.6403 0.6091 0.6403 0.8002
No log 18.6 186 0.5701 0.6617 0.5701 0.7550
No log 18.8 188 0.5475 0.6944 0.5475 0.7399
No log 19.0 190 0.5645 0.6438 0.5645 0.7513
No log 19.2 192 0.5900 0.6429 0.5900 0.7681
No log 19.4 194 0.5959 0.6429 0.5959 0.7720
No log 19.6 196 0.5782 0.6289 0.5782 0.7604
No log 19.8 198 0.5257 0.8047 0.5257 0.7251
No log 20.0 200 0.5245 0.7814 0.5245 0.7242
No log 20.2 202 0.5282 0.7665 0.5282 0.7268
No log 20.4 204 0.5354 0.7175 0.5354 0.7317
No log 20.6 206 0.5313 0.7348 0.5313 0.7289
No log 20.8 208 0.5731 0.6669 0.5731 0.7571
No log 21.0 210 0.6612 0.6047 0.6612 0.8132
No log 21.2 212 0.6451 0.6047 0.6451 0.8032
No log 21.4 214 0.5790 0.6464 0.5790 0.7609
No log 21.6 216 0.5333 0.6962 0.5333 0.7302
No log 21.8 218 0.5355 0.7136 0.5355 0.7318
No log 22.0 220 0.5819 0.6272 0.5819 0.7628
No log 22.2 222 0.7460 0.5725 0.7460 0.8637
No log 22.4 224 0.8346 0.5484 0.8346 0.9136
No log 22.6 226 0.7409 0.5663 0.7409 0.8608
No log 22.8 228 0.5878 0.6272 0.5878 0.7667
No log 23.0 230 0.5454 0.7404 0.5454 0.7385
No log 23.2 232 0.5508 0.7354 0.5508 0.7422
No log 23.4 234 0.6206 0.6257 0.6206 0.7878
No log 23.6 236 0.7161 0.6111 0.7161 0.8463
No log 23.8 238 0.7524 0.5625 0.7524 0.8674
No log 24.0 240 0.6631 0.5811 0.6631 0.8143
No log 24.2 242 0.5844 0.6139 0.5844 0.7645
No log 24.4 244 0.5897 0.5931 0.5897 0.7679
No log 24.6 246 0.5863 0.6139 0.5863 0.7657
No log 24.8 248 0.6193 0.6491 0.6193 0.7870
No log 25.0 250 0.6369 0.5686 0.6369 0.7980
No log 25.2 252 0.6450 0.5888 0.6450 0.8031
No log 25.4 254 0.6009 0.6395 0.6009 0.7752
No log 25.6 256 0.5955 0.6502 0.5955 0.7717
No log 25.8 258 0.6372 0.5777 0.6372 0.7982
No log 26.0 260 0.6061 0.6042 0.6061 0.7785
No log 26.2 262 0.5499 0.6566 0.5499 0.7415
No log 26.4 264 0.5636 0.6536 0.5636 0.7507
No log 26.6 266 0.5486 0.6259 0.5486 0.7407
No log 26.8 268 0.5999 0.6573 0.5999 0.7746
No log 27.0 270 0.6768 0.6341 0.6768 0.8227
No log 27.2 272 0.7152 0.6120 0.7152 0.8457
No log 27.4 274 0.6532 0.6362 0.6532 0.8082
No log 27.6 276 0.5791 0.6672 0.5791 0.7610
No log 27.8 278 0.5883 0.6302 0.5883 0.7670
No log 28.0 280 0.6060 0.5010 0.6060 0.7785
No log 28.2 282 0.6576 0.5728 0.6576 0.8109
No log 28.4 284 0.7557 0.5463 0.7557 0.8693
No log 28.6 286 0.7674 0.5543 0.7674 0.8760
No log 28.8 288 0.6534 0.5927 0.6534 0.8084
No log 29.0 290 0.5598 0.6335 0.5598 0.7482
No log 29.2 292 0.5523 0.5955 0.5523 0.7432
No log 29.4 294 0.5592 0.5972 0.5592 0.7478
No log 29.6 296 0.6021 0.6305 0.6021 0.7759
No log 29.8 298 0.6899 0.5707 0.6899 0.8306
No log 30.0 300 0.7363 0.5777 0.7363 0.8581
No log 30.2 302 0.7274 0.5756 0.7274 0.8529
No log 30.4 304 0.6619 0.5927 0.6619 0.8136
No log 30.6 306 0.5784 0.6464 0.5784 0.7605
No log 30.8 308 0.5627 0.6946 0.5627 0.7501
No log 31.0 310 0.5588 0.6806 0.5588 0.7475
No log 31.2 312 0.5826 0.6352 0.5826 0.7633
No log 31.4 314 0.6331 0.5912 0.6331 0.7957
No log 31.6 316 0.6147 0.6157 0.6147 0.7841
No log 31.8 318 0.5781 0.6500 0.5781 0.7603
No log 32.0 320 0.5834 0.6352 0.5834 0.7638
No log 32.2 322 0.5878 0.6352 0.5878 0.7667
No log 32.4 324 0.6068 0.6109 0.6068 0.7790
No log 32.6 326 0.6566 0.6099 0.6566 0.8103
No log 32.8 328 0.6411 0.6287 0.6411 0.8007
No log 33.0 330 0.5806 0.6464 0.5806 0.7620
No log 33.2 332 0.5668 0.6838 0.5668 0.7529
No log 33.4 334 0.5889 0.6352 0.5889 0.7674
No log 33.6 336 0.5996 0.6386 0.5996 0.7743
No log 33.8 338 0.6025 0.6386 0.6025 0.7762
No log 34.0 340 0.5779 0.6386 0.5779 0.7602
No log 34.2 342 0.5768 0.6386 0.5768 0.7595
No log 34.4 344 0.5920 0.6386 0.5920 0.7694
No log 34.6 346 0.5815 0.6352 0.5815 0.7626
No log 34.8 348 0.6065 0.6091 0.6065 0.7788
No log 35.0 350 0.5949 0.6380 0.5949 0.7713
No log 35.2 352 0.5570 0.7324 0.5570 0.7463
No log 35.4 354 0.5704 0.7191 0.5704 0.7553
No log 35.6 356 0.5655 0.7327 0.5655 0.7520
No log 35.8 358 0.5553 0.6880 0.5553 0.7452
No log 36.0 360 0.6199 0.6054 0.6199 0.7873
No log 36.2 362 0.7056 0.6071 0.7056 0.8400
No log 36.4 364 0.6904 0.6071 0.6904 0.8309
No log 36.6 366 0.6015 0.6678 0.6015 0.7755
No log 36.8 368 0.5659 0.6332 0.5659 0.7523
No log 37.0 370 0.5720 0.6337 0.5720 0.7563
No log 37.2 372 0.5573 0.6227 0.5573 0.7465
No log 37.4 374 0.5555 0.6482 0.5555 0.7453
No log 37.6 376 0.6483 0.6145 0.6483 0.8052
No log 37.8 378 0.7576 0.6421 0.7576 0.8704
No log 38.0 380 0.8364 0.5988 0.8364 0.9145
No log 38.2 382 0.7818 0.6288 0.7818 0.8842
No log 38.4 384 0.6972 0.6247 0.6972 0.8350
No log 38.6 386 0.6637 0.6071 0.6637 0.8147
No log 38.8 388 0.6102 0.6025 0.6102 0.7812
No log 39.0 390 0.5761 0.6325 0.5761 0.7590
No log 39.2 392 0.5847 0.6252 0.5847 0.7647
No log 39.4 394 0.5881 0.5822 0.5881 0.7669
No log 39.6 396 0.5897 0.5529 0.5897 0.7679
No log 39.8 398 0.5892 0.5631 0.5892 0.7676
No log 40.0 400 0.5884 0.5862 0.5884 0.7671
No log 40.2 402 0.6100 0.6157 0.6100 0.7810
No log 40.4 404 0.6578 0.5666 0.6578 0.8110
No log 40.6 406 0.7073 0.5810 0.7073 0.8410
No log 40.8 408 0.6776 0.5938 0.6776 0.8232
No log 41.0 410 0.6468 0.6257 0.6468 0.8042
No log 41.2 412 0.6679 0.5938 0.6679 0.8172
No log 41.4 414 0.6733 0.5756 0.6733 0.8205
No log 41.6 416 0.6762 0.5756 0.6762 0.8223
No log 41.8 418 0.6971 0.5443 0.6971 0.8349
No log 42.0 420 0.6605 0.5902 0.6605 0.8127
No log 42.2 422 0.6582 0.5902 0.6582 0.8113
No log 42.4 424 0.6609 0.5902 0.6609 0.8129
No log 42.6 426 0.6661 0.5902 0.6661 0.8161
No log 42.8 428 0.6697 0.5788 0.6697 0.8184
No log 43.0 430 0.6488 0.5927 0.6488 0.8055
No log 43.2 432 0.6470 0.5927 0.6470 0.8043
No log 43.4 434 0.6538 0.5811 0.6538 0.8086
No log 43.6 436 0.6437 0.5927 0.6437 0.8023
No log 43.8 438 0.5825 0.6464 0.5825 0.7632
No log 44.0 440 0.5521 0.6673 0.5521 0.7430
No log 44.2 442 0.5479 0.6970 0.5479 0.7402
No log 44.4 444 0.5646 0.6756 0.5646 0.7514
No log 44.6 446 0.6272 0.6071 0.6272 0.7919
No log 44.8 448 0.6719 0.5877 0.6719 0.8197
No log 45.0 450 0.6665 0.5877 0.6665 0.8164
No log 45.2 452 0.6556 0.5686 0.6556 0.8097
No log 45.4 454 0.6249 0.5686 0.6249 0.7905
No log 45.6 456 0.6020 0.6547 0.6020 0.7759
No log 45.8 458 0.6102 0.6547 0.6102 0.7811
No log 46.0 460 0.6394 0.6305 0.6394 0.7996
No log 46.2 462 0.6547 0.5998 0.6547 0.8091
No log 46.4 464 0.6916 0.5766 0.6916 0.8316
No log 46.6 466 0.6872 0.5572 0.6872 0.8290
No log 46.8 468 0.6505 0.5799 0.6505 0.8065
No log 47.0 470 0.6347 0.5912 0.6347 0.7967
No log 47.2 472 0.6019 0.6361 0.6019 0.7758
No log 47.4 474 0.5808 0.6798 0.5808 0.7621
No log 47.6 476 0.5861 0.6687 0.5861 0.7656
No log 47.8 478 0.5975 0.6328 0.5975 0.7730
No log 48.0 480 0.6390 0.5888 0.6390 0.7994
No log 48.2 482 0.6536 0.6071 0.6536 0.8084
No log 48.4 484 0.6594 0.6071 0.6594 0.8120
No log 48.6 486 0.6266 0.5998 0.6266 0.7916
No log 48.8 488 0.5847 0.6328 0.5847 0.7647
No log 49.0 490 0.5658 0.7027 0.5658 0.7522
No log 49.2 492 0.5620 0.7417 0.5620 0.7497
No log 49.4 494 0.5565 0.7567 0.5565 0.7460
No log 49.6 496 0.5492 0.7417 0.5492 0.7411
No log 49.8 498 0.5629 0.6685 0.5629 0.7503
0.2222 50.0 500 0.5774 0.6476 0.5774 0.7599
0.2222 50.2 502 0.5697 0.6511 0.5697 0.7548
0.2222 50.4 504 0.5857 0.6291 0.5857 0.7653
0.2222 50.6 506 0.6062 0.5642 0.6062 0.7786
0.2222 50.8 508 0.6102 0.5642 0.6102 0.7812
0.2222 51.0 510 0.6149 0.5938 0.6149 0.7842
0.2222 51.2 512 0.6078 0.5912 0.6078 0.7796

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k2_task5_organization

Finetuned
(4019)
this model