ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k7_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8755
  • Qwk: 0.6383
  • Mse: 0.8755
  • Rmse: 0.9357

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0625 2 7.0123 0.0179 7.0123 2.6481
No log 0.125 4 4.7038 0.1004 4.7038 2.1688
No log 0.1875 6 3.6990 -0.0106 3.6990 1.9233
No log 0.25 8 3.3079 0.0112 3.3079 1.8188
No log 0.3125 10 2.1580 0.1440 2.1580 1.4690
No log 0.375 12 2.1646 0.1333 2.1646 1.4712
No log 0.4375 14 2.2284 -0.0161 2.2284 1.4928
No log 0.5 16 2.1081 0.1071 2.1081 1.4519
No log 0.5625 18 2.2974 0.0310 2.2974 1.5157
No log 0.625 20 2.2264 0.0465 2.2264 1.4921
No log 0.6875 22 2.1492 0.1818 2.1492 1.4660
No log 0.75 24 1.9889 0.2105 1.9889 1.4103
No log 0.8125 26 2.2608 0.1867 2.2608 1.5036
No log 0.875 28 3.1326 0.1809 3.1326 1.7699
No log 0.9375 30 2.7872 0.2000 2.7872 1.6695
No log 1.0 32 2.0519 0.3205 2.0519 1.4324
No log 1.0625 34 1.6584 0.4143 1.6584 1.2878
No log 1.125 36 1.9036 0.3553 1.9036 1.3797
No log 1.1875 38 2.0013 0.3165 2.0013 1.4147
No log 1.25 40 1.8514 0.3067 1.8514 1.3607
No log 1.3125 42 1.9600 0.2763 1.9600 1.4000
No log 1.375 44 1.7105 0.4 1.7105 1.3079
No log 1.4375 46 1.3155 0.5147 1.3155 1.1469
No log 1.5 48 1.1506 0.5113 1.1506 1.0727
No log 1.5625 50 1.1783 0.4923 1.1783 1.0855
No log 1.625 52 1.0953 0.5606 1.0953 1.0466
No log 1.6875 54 1.0519 0.5693 1.0519 1.0256
No log 1.75 56 1.3216 0.5063 1.3216 1.1496
No log 1.8125 58 1.7801 0.4634 1.7801 1.3342
No log 1.875 60 1.7645 0.4762 1.7645 1.3284
No log 1.9375 62 1.3030 0.4755 1.3030 1.1415
No log 2.0 64 1.0892 0.5224 1.0892 1.0437
No log 2.0625 66 1.1473 0.4697 1.1473 1.0711
No log 2.125 68 1.4841 0.4658 1.4841 1.2182
No log 2.1875 70 1.8211 0.3974 1.8211 1.3495
No log 2.25 72 1.8143 0.4151 1.8143 1.3470
No log 2.3125 74 1.4165 0.5753 1.4165 1.1902
No log 2.375 76 1.1572 0.5507 1.1572 1.0757
No log 2.4375 78 1.2783 0.5616 1.2783 1.1306
No log 2.5 80 1.4077 0.5733 1.4077 1.1865
No log 2.5625 82 1.5114 0.5767 1.5114 1.2294
No log 2.625 84 1.9012 0.4536 1.9012 1.3789
No log 2.6875 86 1.6527 0.5492 1.6527 1.2856
No log 2.75 88 1.4181 0.6421 1.4181 1.1908
No log 2.8125 90 1.5535 0.5771 1.5535 1.2464
No log 2.875 92 1.2551 0.6492 1.2551 1.1203
No log 2.9375 94 1.0830 0.6441 1.0830 1.0407
No log 3.0 96 1.0523 0.6471 1.0523 1.0258
No log 3.0625 98 1.2906 0.5829 1.2906 1.1360
No log 3.125 100 1.7718 0.5025 1.7718 1.3311
No log 3.1875 102 1.6615 0.5417 1.6615 1.2890
No log 3.25 104 1.2306 0.6541 1.2306 1.1093
No log 3.3125 106 1.3251 0.4706 1.3251 1.1511
No log 3.375 108 1.2029 0.5039 1.2029 1.0968
No log 3.4375 110 0.8763 0.6377 0.8763 0.9361
No log 3.5 112 0.9084 0.7125 0.9084 0.9531
No log 3.5625 114 1.1971 0.6145 1.1971 1.0941
No log 3.625 116 1.0396 0.6581 1.0396 1.0196
No log 3.6875 118 0.8148 0.6389 0.8148 0.9027
No log 3.75 120 1.2089 0.496 1.2089 1.0995
No log 3.8125 122 1.3312 0.4553 1.3312 1.1538
No log 3.875 124 1.3064 0.4615 1.3064 1.1430
No log 3.9375 126 1.0610 0.5143 1.0610 1.0300
No log 4.0 128 1.1029 0.6225 1.1029 1.0502
No log 4.0625 130 1.1738 0.6380 1.1738 1.0834
No log 4.125 132 1.1668 0.6543 1.1668 1.0802
No log 4.1875 134 0.9592 0.6842 0.9592 0.9794
No log 4.25 136 0.7209 0.72 0.7209 0.8491
No log 4.3125 138 0.7042 0.72 0.7042 0.8392
No log 4.375 140 0.6279 0.7260 0.6279 0.7924
No log 4.4375 142 0.6065 0.7483 0.6065 0.7788
No log 4.5 144 0.7718 0.7027 0.7718 0.8785
No log 4.5625 146 0.8938 0.6883 0.8938 0.9454
No log 4.625 148 0.8463 0.6887 0.8463 0.9199
No log 4.6875 150 0.7318 0.7152 0.7318 0.8554
No log 4.75 152 0.5972 0.8129 0.5972 0.7728
No log 4.8125 154 0.6712 0.8105 0.6712 0.8193
No log 4.875 156 0.6834 0.7792 0.6834 0.8267
No log 4.9375 158 0.8517 0.6892 0.8517 0.9229
No log 5.0 160 1.2024 0.6272 1.2024 1.0965
No log 5.0625 162 1.2384 0.6272 1.2384 1.1128
No log 5.125 164 1.0433 0.6259 1.0433 1.0214
No log 5.1875 166 0.8852 0.6667 0.8852 0.9409
No log 5.25 168 0.8827 0.5873 0.8827 0.9395
No log 5.3125 170 0.8341 0.6308 0.8341 0.9133
No log 5.375 172 0.8318 0.6712 0.8318 0.9120
No log 5.4375 174 0.9443 0.6531 0.9443 0.9717
No log 5.5 176 0.9229 0.6711 0.9229 0.9607
No log 5.5625 178 0.7529 0.7205 0.7529 0.8677
No log 5.625 180 0.5926 0.7771 0.5926 0.7698
No log 5.6875 182 0.6178 0.7730 0.6178 0.7860
No log 5.75 184 0.6756 0.7904 0.6756 0.8219
No log 5.8125 186 0.8206 0.7066 0.8206 0.9058
No log 5.875 188 0.8712 0.6581 0.8712 0.9334
No log 5.9375 190 0.8798 0.6395 0.8798 0.9380
No log 6.0 192 0.8993 0.625 0.8993 0.9483
No log 6.0625 194 0.9065 0.625 0.9065 0.9521
No log 6.125 196 0.8927 0.625 0.8927 0.9448
No log 6.1875 198 0.8316 0.6528 0.8316 0.9119
No log 6.25 200 0.7784 0.6939 0.7784 0.8823
No log 6.3125 202 0.7027 0.7162 0.7027 0.8383
No log 6.375 204 0.6479 0.7248 0.6479 0.8049
No log 6.4375 206 0.6550 0.7578 0.6550 0.8093
No log 6.5 208 0.7414 0.7439 0.7414 0.8611
No log 6.5625 210 0.6903 0.75 0.6903 0.8308
No log 6.625 212 0.6517 0.7467 0.6517 0.8073
No log 6.6875 214 0.6776 0.7451 0.6776 0.8232
No log 6.75 216 0.7832 0.7089 0.7832 0.8850
No log 6.8125 218 0.7550 0.6892 0.7550 0.8689
No log 6.875 220 0.6687 0.7162 0.6687 0.8178
No log 6.9375 222 0.6307 0.7397 0.6307 0.7942
No log 7.0 224 0.5742 0.8 0.5742 0.7578
No log 7.0625 226 0.5313 0.8258 0.5313 0.7289
No log 7.125 228 0.6033 0.75 0.6033 0.7767
No log 7.1875 230 0.7011 0.7485 0.7011 0.8373
No log 7.25 232 0.6464 0.72 0.6464 0.8040
No log 7.3125 234 0.6473 0.7586 0.6473 0.8046
No log 7.375 236 0.6827 0.7832 0.6827 0.8263
No log 7.4375 238 0.6698 0.7273 0.6698 0.8184
No log 7.5 240 0.7751 0.6846 0.7751 0.8804
No log 7.5625 242 0.9159 0.6486 0.9159 0.9570
No log 7.625 244 0.8113 0.6486 0.8113 0.9007
No log 7.6875 246 0.6310 0.7347 0.6310 0.7944
No log 7.75 248 0.6164 0.7703 0.6164 0.7851
No log 7.8125 250 0.6485 0.7632 0.6485 0.8053
No log 7.875 252 0.8251 0.6622 0.8251 0.9084
No log 7.9375 254 0.8983 0.6826 0.8983 0.9478
No log 8.0 256 0.7980 0.7229 0.7980 0.8933
No log 8.0625 258 0.7337 0.7442 0.7337 0.8566
No log 8.125 260 0.5722 0.8101 0.5722 0.7564
No log 8.1875 262 0.5608 0.8205 0.5608 0.7489
No log 8.25 264 0.5414 0.8228 0.5414 0.7358
No log 8.3125 266 0.6423 0.7285 0.6423 0.8014
No log 8.375 268 0.7541 0.6933 0.7541 0.8684
No log 8.4375 270 0.6955 0.6939 0.6955 0.8339
No log 8.5 272 0.6576 0.72 0.6576 0.8109
No log 8.5625 274 0.6672 0.7114 0.6672 0.8168
No log 8.625 276 0.7182 0.7027 0.7182 0.8475
No log 8.6875 278 0.8125 0.6533 0.8125 0.9014
No log 8.75 280 0.7704 0.6962 0.7704 0.8777
No log 8.8125 282 0.6160 0.75 0.6160 0.7849
No log 8.875 284 0.5851 0.76 0.5851 0.7649
No log 8.9375 286 0.6288 0.7324 0.6288 0.7929
No log 9.0 288 0.7056 0.7391 0.7056 0.8400
No log 9.0625 290 0.7485 0.6822 0.7485 0.8651
No log 9.125 292 0.7634 0.6512 0.7634 0.8737
No log 9.1875 294 0.7859 0.6619 0.7859 0.8865
No log 9.25 296 0.8113 0.6331 0.8113 0.9007
No log 9.3125 298 0.7611 0.6619 0.7611 0.8724
No log 9.375 300 0.7277 0.6870 0.7277 0.8531
No log 9.4375 302 0.7134 0.6818 0.7134 0.8446
No log 9.5 304 0.7698 0.6573 0.7698 0.8774
No log 9.5625 306 0.9381 0.6438 0.9381 0.9686
No log 9.625 308 0.9713 0.6577 0.9713 0.9855
No log 9.6875 310 0.8422 0.6099 0.8422 0.9177
No log 9.75 312 0.6984 0.6944 0.6984 0.8357
No log 9.8125 314 0.5941 0.7397 0.5941 0.7708
No log 9.875 316 0.5176 0.8 0.5176 0.7195
No log 9.9375 318 0.5041 0.8153 0.5041 0.7100
No log 10.0 320 0.5641 0.7953 0.5641 0.7510
No log 10.0625 322 0.5556 0.8023 0.5556 0.7454
No log 10.125 324 0.5286 0.8554 0.5286 0.7270
No log 10.1875 326 0.5229 0.8375 0.5229 0.7231
No log 10.25 328 0.5227 0.8375 0.5227 0.7230
No log 10.3125 330 0.5153 0.8101 0.5153 0.7178
No log 10.375 332 0.5143 0.8153 0.5143 0.7172
No log 10.4375 334 0.5385 0.7733 0.5385 0.7338
No log 10.5 336 0.5641 0.7838 0.5641 0.7511
No log 10.5625 338 0.5841 0.7891 0.5841 0.7642
No log 10.625 340 0.5446 0.7895 0.5446 0.7379
No log 10.6875 342 0.6189 0.7632 0.6189 0.7867
No log 10.75 344 0.7411 0.6531 0.7411 0.8609
No log 10.8125 346 0.6793 0.7162 0.6793 0.8242
No log 10.875 348 0.5851 0.7619 0.5851 0.7649
No log 10.9375 350 0.5870 0.7651 0.5870 0.7661
No log 11.0 352 0.6655 0.75 0.6655 0.8158
No log 11.0625 354 0.6537 0.7692 0.6537 0.8085
No log 11.125 356 0.6393 0.7722 0.6393 0.7996
No log 11.1875 358 0.6278 0.7722 0.6278 0.7923
No log 11.25 360 0.5916 0.7898 0.5916 0.7692
No log 11.3125 362 0.5800 0.7947 0.5800 0.7616
No log 11.375 364 0.5769 0.8182 0.5769 0.7596
No log 11.4375 366 0.5685 0.7974 0.5685 0.7540
No log 11.5 368 0.5503 0.7947 0.5503 0.7418
No log 11.5625 370 0.5571 0.7949 0.5571 0.7464
No log 11.625 372 0.5931 0.7879 0.5931 0.7701
No log 11.6875 374 0.5719 0.7901 0.5719 0.7563
No log 11.75 376 0.5606 0.7949 0.5606 0.7487
No log 11.8125 378 0.6158 0.7517 0.6158 0.7847
No log 11.875 380 0.6796 0.7310 0.6796 0.8244
No log 11.9375 382 0.7104 0.7183 0.7104 0.8429
No log 12.0 384 0.7482 0.6849 0.7482 0.8650
No log 12.0625 386 0.7496 0.6849 0.7496 0.8658
No log 12.125 388 0.6984 0.7383 0.6984 0.8357
No log 12.1875 390 0.6392 0.7347 0.6392 0.7995
No log 12.25 392 0.6091 0.7838 0.6091 0.7804
No log 12.3125 394 0.5896 0.8133 0.5896 0.7679
No log 12.375 396 0.5728 0.7867 0.5728 0.7568
No log 12.4375 398 0.6036 0.7651 0.6036 0.7769
No log 12.5 400 0.6410 0.7651 0.6410 0.8006
No log 12.5625 402 0.7169 0.7114 0.7169 0.8467
No log 12.625 404 0.8424 0.6351 0.8424 0.9178
No log 12.6875 406 0.8531 0.625 0.8531 0.9237
No log 12.75 408 0.8188 0.6475 0.8188 0.9049
No log 12.8125 410 0.7860 0.6715 0.7860 0.8866
No log 12.875 412 0.7431 0.7050 0.7431 0.8620
No log 12.9375 414 0.7105 0.6906 0.7105 0.8429
No log 13.0 416 0.7771 0.6667 0.7771 0.8815
No log 13.0625 418 0.9043 0.6538 0.9043 0.9509
No log 13.125 420 0.8511 0.6438 0.8511 0.9225
No log 13.1875 422 0.6864 0.7310 0.6864 0.8285
No log 13.25 424 0.6726 0.7612 0.6726 0.8201
No log 13.3125 426 0.7659 0.6818 0.7659 0.8751
No log 13.375 428 0.7739 0.6970 0.7739 0.8797
No log 13.4375 430 0.7353 0.7023 0.7353 0.8575
No log 13.5 432 0.8148 0.6434 0.8148 0.9026
No log 13.5625 434 0.9525 0.6622 0.9525 0.9760
No log 13.625 436 0.9603 0.6622 0.9603 0.9800
No log 13.6875 438 0.8158 0.6483 0.8158 0.9032
No log 13.75 440 0.6627 0.7347 0.6627 0.8141
No log 13.8125 442 0.5871 0.7483 0.5871 0.7662
No log 13.875 444 0.5893 0.8054 0.5893 0.7676
No log 13.9375 446 0.5916 0.7973 0.5916 0.7692
No log 14.0 448 0.6403 0.7260 0.6403 0.8002
No log 14.0625 450 0.7580 0.7020 0.7580 0.8706
No log 14.125 452 0.8580 0.6622 0.8580 0.9263
No log 14.1875 454 0.8282 0.6483 0.8282 0.9101
No log 14.25 456 0.7274 0.6525 0.7274 0.8529
No log 14.3125 458 0.6860 0.7133 0.6860 0.8282
No log 14.375 460 0.6779 0.7172 0.6779 0.8233
No log 14.4375 462 0.7374 0.6849 0.7374 0.8587
No log 14.5 464 0.7474 0.6759 0.7474 0.8645
No log 14.5625 466 0.7507 0.6383 0.7507 0.8664
No log 14.625 468 0.7331 0.6475 0.7331 0.8562
No log 14.6875 470 0.7361 0.6571 0.7361 0.8580
No log 14.75 472 0.7826 0.6232 0.7826 0.8847
No log 14.8125 474 0.8004 0.5985 0.8004 0.8946
No log 14.875 476 0.8072 0.5942 0.8072 0.8984
No log 14.9375 478 0.8391 0.6043 0.8391 0.9160
No log 15.0 480 0.9040 0.6187 0.9040 0.9508
No log 15.0625 482 0.8931 0.6338 0.8931 0.9450
No log 15.125 484 0.8384 0.6294 0.8384 0.9156
No log 15.1875 486 0.9065 0.6623 0.9065 0.9521
No log 15.25 488 0.9993 0.6957 0.9993 0.9997
No log 15.3125 490 0.9697 0.6883 0.9697 0.9847
No log 15.375 492 0.8811 0.6294 0.8811 0.9387
No log 15.4375 494 0.7444 0.6806 0.7444 0.8628
No log 15.5 496 0.7063 0.6912 0.7063 0.8404
No log 15.5625 498 0.7462 0.6615 0.7462 0.8638
0.3625 15.625 500 0.7510 0.6667 0.7510 0.8666
0.3625 15.6875 502 0.7727 0.6761 0.7727 0.8790
0.3625 15.75 504 0.8265 0.6479 0.8265 0.9091
0.3625 15.8125 506 0.7900 0.6479 0.7900 0.8888
0.3625 15.875 508 0.7070 0.6944 0.7070 0.8409
0.3625 15.9375 510 0.6880 0.6901 0.6880 0.8295
0.3625 16.0 512 0.7204 0.6950 0.7204 0.8488
0.3625 16.0625 514 0.7746 0.6620 0.7746 0.8801
0.3625 16.125 516 0.7944 0.6525 0.7944 0.8913
0.3625 16.1875 518 0.7595 0.6571 0.7595 0.8715
0.3625 16.25 520 0.7365 0.6713 0.7365 0.8582
0.3625 16.3125 522 0.7355 0.6713 0.7355 0.8576
0.3625 16.375 524 0.7035 0.7083 0.7035 0.8387
0.3625 16.4375 526 0.7163 0.7083 0.7163 0.8464
0.3625 16.5 528 0.7997 0.6620 0.7997 0.8943
0.3625 16.5625 530 0.8677 0.6383 0.8677 0.9315
0.3625 16.625 532 0.8538 0.6383 0.8538 0.9240
0.3625 16.6875 534 0.8818 0.6383 0.8818 0.9390
0.3625 16.75 536 0.8755 0.6383 0.8755 0.9357

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k7_task1_organization

Finetuned
(4019)
this model