ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k6_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8249
  • Qwk: 0.6716
  • Mse: 0.8249
  • Rmse: 0.9082

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0667 2 6.8161 0.0116 6.8161 2.6108
No log 0.1333 4 4.5232 0.0601 4.5232 2.1268
No log 0.2 6 3.2746 0.0 3.2746 1.8096
No log 0.2667 8 2.2695 0.0730 2.2695 1.5065
No log 0.3333 10 1.7956 0.2222 1.7956 1.3400
No log 0.4 12 1.8336 0.1538 1.8336 1.3541
No log 0.4667 14 1.8485 0.1121 1.8485 1.3596
No log 0.5333 16 1.6345 0.1538 1.6345 1.2785
No log 0.6 18 1.5394 0.1538 1.5394 1.2407
No log 0.6667 20 1.6688 0.1869 1.6688 1.2918
No log 0.7333 22 1.8799 0.2712 1.8799 1.3711
No log 0.8 24 1.9968 0.2080 1.9968 1.4131
No log 0.8667 26 2.0434 0.1575 2.0434 1.4295
No log 0.9333 28 2.3672 0.1449 2.3672 1.5386
No log 1.0 30 2.6414 0.0526 2.6414 1.6252
No log 1.0667 32 2.3479 0.1727 2.3479 1.5323
No log 1.1333 34 1.7115 0.2906 1.7115 1.3082
No log 1.2 36 1.6325 0.3471 1.6325 1.2777
No log 1.2667 38 1.5247 0.3036 1.5247 1.2348
No log 1.3333 40 1.4282 0.2264 1.4282 1.1951
No log 1.4 42 1.3815 0.2430 1.3815 1.1754
No log 1.4667 44 1.3644 0.2430 1.3644 1.1681
No log 1.5333 46 1.3643 0.2430 1.3643 1.1680
No log 1.6 48 1.3851 0.3009 1.3851 1.1769
No log 1.6667 50 1.3217 0.3009 1.3217 1.1497
No log 1.7333 52 1.2859 0.3009 1.2859 1.1340
No log 1.8 54 1.4737 0.3529 1.4737 1.2140
No log 1.8667 56 1.6967 0.4032 1.6967 1.3026
No log 1.9333 58 1.6579 0.3937 1.6579 1.2876
No log 2.0 60 1.6799 0.4341 1.6799 1.2961
No log 2.0667 62 1.5684 0.4341 1.5684 1.2524
No log 2.1333 64 1.3640 0.4194 1.3640 1.1679
No log 2.2 66 1.0670 0.4706 1.0670 1.0330
No log 2.2667 68 1.0682 0.4444 1.0682 1.0335
No log 2.3333 70 1.1741 0.3932 1.1741 1.0835
No log 2.4 72 1.5276 0.3968 1.5276 1.2360
No log 2.4667 74 1.3030 0.4480 1.3030 1.1415
No log 2.5333 76 0.9514 0.5920 0.9514 0.9754
No log 2.6 78 0.8935 0.624 0.8935 0.9453
No log 2.6667 80 0.9474 0.5968 0.9474 0.9734
No log 2.7333 82 0.9326 0.5645 0.9326 0.9657
No log 2.8 84 0.8667 0.6299 0.8667 0.9310
No log 2.8667 86 0.8058 0.6769 0.8058 0.8976
No log 2.9333 88 0.7982 0.6615 0.7982 0.8934
No log 3.0 90 0.7933 0.6462 0.7933 0.8907
No log 3.0667 92 0.8294 0.6462 0.8294 0.9107
No log 3.1333 94 0.9599 0.6462 0.9599 0.9798
No log 3.2 96 1.0515 0.6061 1.0515 1.0254
No log 3.2667 98 1.0130 0.6512 1.0130 1.0065
No log 3.3333 100 0.9290 0.5323 0.9290 0.9638
No log 3.4 102 0.8275 0.6202 0.8275 0.9097
No log 3.4667 104 0.7236 0.6466 0.7236 0.8506
No log 3.5333 106 0.6700 0.6963 0.6700 0.8185
No log 3.6 108 0.6529 0.7206 0.6529 0.8080
No log 3.6667 110 0.6440 0.7234 0.6440 0.8025
No log 3.7333 112 0.6161 0.7361 0.6161 0.7849
No log 3.8 114 0.6298 0.75 0.6298 0.7936
No log 3.8667 116 0.7562 0.6950 0.7562 0.8696
No log 3.9333 118 0.9171 0.6809 0.9171 0.9577
No log 4.0 120 1.1986 0.4853 1.1986 1.0948
No log 4.0667 122 1.1544 0.5441 1.1544 1.0744
No log 4.1333 124 0.9661 0.6087 0.9661 0.9829
No log 4.2 126 0.8824 0.6377 0.8824 0.9394
No log 4.2667 128 0.7573 0.6619 0.7573 0.8702
No log 4.3333 130 0.7671 0.6619 0.7671 0.8758
No log 4.4 132 0.8530 0.6119 0.8530 0.9236
No log 4.4667 134 0.9326 0.6232 0.9326 0.9657
No log 4.5333 136 0.9831 0.5926 0.9831 0.9915
No log 4.6 138 0.8717 0.6471 0.8717 0.9337
No log 4.6667 140 0.8452 0.6 0.8452 0.9193
No log 4.7333 142 0.9067 0.6047 0.9067 0.9522
No log 4.8 144 0.7855 0.6316 0.7855 0.8863
No log 4.8667 146 0.7248 0.7133 0.7248 0.8513
No log 4.9333 148 0.8095 0.6621 0.8095 0.8997
No log 5.0 150 0.7940 0.6338 0.7940 0.8910
No log 5.0667 152 0.7834 0.6761 0.7834 0.8851
No log 5.1333 154 0.6961 0.7368 0.6961 0.8344
No log 5.2 156 0.5913 0.7925 0.5913 0.7689
No log 5.2667 158 0.6243 0.7639 0.6243 0.7901
No log 5.3333 160 0.6651 0.7310 0.6651 0.8155
No log 5.4 162 0.7657 0.6619 0.7657 0.8750
No log 5.4667 164 0.8394 0.6143 0.8394 0.9162
No log 5.5333 166 0.8015 0.6331 0.8015 0.8953
No log 5.6 168 0.7218 0.6812 0.7218 0.8496
No log 5.6667 170 0.6840 0.7083 0.6840 0.8271
No log 5.7333 172 0.6924 0.7114 0.6924 0.8321
No log 5.8 174 0.8027 0.6944 0.8027 0.8959
No log 5.8667 176 1.0911 0.6225 1.0911 1.0446
No log 5.9333 178 1.0431 0.6622 1.0431 1.0213
No log 6.0 180 0.8274 0.6519 0.8274 0.9096
No log 6.0667 182 0.7403 0.7042 0.7403 0.8604
No log 6.1333 184 0.8157 0.6714 0.8157 0.9031
No log 6.2 186 0.8094 0.6131 0.8094 0.8997
No log 6.2667 188 0.8613 0.6519 0.8613 0.9281
No log 6.3333 190 1.0329 0.6131 1.0329 1.0163
No log 6.4 192 1.0119 0.5755 1.0119 1.0060
No log 6.4667 194 0.8696 0.6241 0.8696 0.9325
No log 6.5333 196 0.8161 0.6765 0.8161 0.9034
No log 6.6 198 0.8283 0.6667 0.8283 0.9101
No log 6.6667 200 0.8618 0.6370 0.8618 0.9283
No log 6.7333 202 0.9105 0.6519 0.9105 0.9542
No log 6.8 204 0.8818 0.6471 0.8818 0.9391
No log 6.8667 206 0.8416 0.6569 0.8416 0.9174
No log 6.9333 208 0.7870 0.6667 0.7870 0.8871
No log 7.0 210 0.7685 0.6812 0.7685 0.8767
No log 7.0667 212 0.7712 0.6812 0.7712 0.8782
No log 7.1333 214 0.7766 0.6715 0.7766 0.8812
No log 7.2 216 0.7151 0.6866 0.7151 0.8457
No log 7.2667 218 0.6675 0.6963 0.6675 0.8170
No log 7.3333 220 0.6128 0.7838 0.6128 0.7828
No log 7.4 222 0.5952 0.7550 0.5952 0.7715
No log 7.4667 224 0.5754 0.7949 0.5754 0.7586
No log 7.5333 226 0.5426 0.8176 0.5426 0.7366
No log 7.6 228 0.5700 0.7891 0.5700 0.7550
No log 7.6667 230 0.7248 0.7067 0.7248 0.8513
No log 7.7333 232 0.7874 0.6573 0.7874 0.8873
No log 7.8 234 0.6906 0.6957 0.6906 0.8310
No log 7.8667 236 0.6955 0.7376 0.6955 0.8340
No log 7.9333 238 0.7586 0.7059 0.7586 0.8710
No log 8.0 240 0.9016 0.6519 0.9016 0.9495
No log 8.0667 242 1.0248 0.5926 1.0248 1.0123
No log 8.1333 244 0.9908 0.6277 0.9908 0.9954
No log 8.2 246 0.8439 0.6567 0.8439 0.9187
No log 8.2667 248 0.7232 0.6957 0.7232 0.8504
No log 8.3333 250 0.6964 0.7286 0.6964 0.8345
No log 8.4 252 0.7778 0.6667 0.7778 0.8819
No log 8.4667 254 0.9804 0.625 0.9804 0.9901
No log 8.5333 256 1.1821 0.56 1.1821 1.0873
No log 8.6 258 1.2606 0.5430 1.2606 1.1228
No log 8.6667 260 1.0952 0.5797 1.0952 1.0465
No log 8.7333 262 0.8507 0.6316 0.8507 0.9223
No log 8.8 264 0.6981 0.7445 0.6981 0.8355
No log 8.8667 266 0.6862 0.7353 0.6862 0.8284
No log 8.9333 268 0.7399 0.7153 0.7399 0.8602
No log 9.0 270 0.9161 0.6269 0.9161 0.9571
No log 9.0667 272 0.9615 0.5970 0.9615 0.9806
No log 9.1333 274 0.8634 0.6269 0.8634 0.9292
No log 9.2 276 0.7840 0.7007 0.7839 0.8854
No log 9.2667 278 0.7836 0.7007 0.7836 0.8852
No log 9.3333 280 0.8497 0.6667 0.8497 0.9218
No log 9.4 282 1.0414 0.6087 1.0414 1.0205
No log 9.4667 284 1.1292 0.5390 1.1292 1.0626
No log 9.5333 286 1.0359 0.5882 1.0359 1.0178
No log 9.6 288 0.8219 0.6765 0.8219 0.9066
No log 9.6667 290 0.7212 0.7429 0.7212 0.8493
No log 9.7333 292 0.7119 0.7273 0.7119 0.8438
No log 9.8 294 0.7170 0.7518 0.7170 0.8467
No log 9.8667 296 0.7826 0.7059 0.7826 0.8846
No log 9.9333 298 0.8547 0.6567 0.8547 0.9245
No log 10.0 300 0.8773 0.6567 0.8773 0.9367
No log 10.0667 302 0.8727 0.6567 0.8727 0.9342
No log 10.1333 304 0.8188 0.6667 0.8188 0.9049
No log 10.2 306 0.7895 0.7007 0.7895 0.8885
No log 10.2667 308 0.8422 0.6567 0.8422 0.9177
No log 10.3333 310 0.9030 0.6567 0.9030 0.9502
No log 10.4 312 0.8857 0.6412 0.8857 0.9411
No log 10.4667 314 0.8763 0.5891 0.8763 0.9361
No log 10.5333 316 0.8714 0.5891 0.8714 0.9335
No log 10.6 318 0.8996 0.6519 0.8996 0.9485
No log 10.6667 320 1.0064 0.6479 1.0064 1.0032
No log 10.7333 322 1.0034 0.6434 1.0034 1.0017
No log 10.8 324 0.9519 0.6286 0.9519 0.9756
No log 10.8667 326 0.9201 0.6522 0.9201 0.9592
No log 10.9333 328 0.9460 0.6475 0.9460 0.9726
No log 11.0 330 0.9739 0.6475 0.9739 0.9869
No log 11.0667 332 1.0086 0.6429 1.0086 1.0043
No log 11.1333 334 0.9969 0.6324 0.9969 0.9984
No log 11.2 336 0.9237 0.6370 0.9237 0.9611
No log 11.2667 338 0.9034 0.6475 0.9034 0.9505
No log 11.3333 340 0.9070 0.6620 0.9070 0.9524
No log 11.4 342 0.9279 0.6475 0.9279 0.9633
No log 11.4667 344 0.8386 0.6471 0.8386 0.9157
No log 11.5333 346 0.7217 0.7101 0.7217 0.8495
No log 11.6 348 0.7364 0.6571 0.7364 0.8581
No log 11.6667 350 0.7722 0.6324 0.7722 0.8788
No log 11.7333 352 0.7687 0.6471 0.7687 0.8768
No log 11.8 354 0.8138 0.6667 0.8138 0.9021
No log 11.8667 356 0.9742 0.6014 0.9742 0.9870
No log 11.9333 358 0.9858 0.6014 0.9858 0.9929
No log 12.0 360 0.8739 0.6761 0.8739 0.9348
No log 12.0667 362 0.7712 0.7162 0.7712 0.8782
No log 12.1333 364 0.7495 0.7027 0.7495 0.8657
No log 12.2 366 0.7519 0.6897 0.7519 0.8671
No log 12.2667 368 0.7777 0.7347 0.7777 0.8819
No log 12.3333 370 0.9227 0.6667 0.9227 0.9606
No log 12.4 372 1.0062 0.6316 1.0062 1.0031
No log 12.4667 374 0.9206 0.6714 0.9206 0.9595
No log 12.5333 376 0.8523 0.6901 0.8523 0.9232
No log 12.6 378 0.7999 0.6861 0.7999 0.8944
No log 12.6667 380 0.7982 0.6418 0.7982 0.8934
No log 12.7333 382 0.8291 0.6866 0.8291 0.9105
No log 12.8 384 0.9095 0.6667 0.9095 0.9537
No log 12.8667 386 1.1042 0.5946 1.1042 1.0508
No log 12.9333 388 1.1580 0.5442 1.1580 1.0761
No log 13.0 390 1.0747 0.5775 1.0747 1.0367
No log 13.0667 392 0.9379 0.6043 0.9379 0.9685
No log 13.1333 394 0.8331 0.6763 0.8331 0.9127
No log 13.2 396 0.7746 0.7568 0.7746 0.8801
No log 13.2667 398 0.7987 0.7123 0.7987 0.8937
No log 13.3333 400 0.7994 0.7483 0.7994 0.8941
No log 13.4 402 0.8082 0.7397 0.8082 0.8990
No log 13.4667 404 0.8444 0.6763 0.8444 0.9189
No log 13.5333 406 0.8469 0.6861 0.8469 0.9203
No log 13.6 408 0.8353 0.6667 0.8353 0.9139
No log 13.6667 410 0.8563 0.6571 0.8563 0.9254
No log 13.7333 412 0.9215 0.6389 0.9215 0.9600
No log 13.8 414 0.9710 0.6207 0.9710 0.9854
No log 13.8667 416 0.9377 0.6438 0.9377 0.9684
No log 13.9333 418 0.9205 0.6438 0.9205 0.9594
No log 14.0 420 0.8892 0.6621 0.8892 0.9430
No log 14.0667 422 0.8281 0.6763 0.8281 0.9100
No log 14.1333 424 0.7833 0.6861 0.7833 0.8850
No log 14.2 426 0.7574 0.7050 0.7574 0.8703
No log 14.2667 428 0.7862 0.6957 0.7862 0.8867
No log 14.3333 430 0.8601 0.6812 0.8601 0.9274
No log 14.4 432 0.8577 0.6944 0.8577 0.9261
No log 14.4667 434 0.7748 0.7083 0.7748 0.8803
No log 14.5333 436 0.7308 0.7083 0.7308 0.8549
No log 14.6 438 0.7787 0.7211 0.7787 0.8824
No log 14.6667 440 0.8954 0.7089 0.8954 0.9462
No log 14.7333 442 0.9547 0.6380 0.9547 0.9771
No log 14.8 444 0.9857 0.6627 0.9857 0.9928
No log 14.8667 446 1.0566 0.6 1.0566 1.0279
No log 14.9333 448 0.9714 0.6232 0.9714 0.9856
No log 15.0 450 0.9033 0.6619 0.9033 0.9504
No log 15.0667 452 0.8492 0.6620 0.8492 0.9215
No log 15.1333 454 0.8302 0.6667 0.8302 0.9112
No log 15.2 456 0.9069 0.6345 0.9069 0.9523
No log 15.2667 458 0.9225 0.6294 0.9225 0.9605
No log 15.3333 460 0.9004 0.6901 0.9004 0.9489
No log 15.4 462 0.9469 0.6429 0.9469 0.9731
No log 15.4667 464 0.9263 0.6993 0.9263 0.9624
No log 15.5333 466 0.9623 0.6475 0.9623 0.9810
No log 15.6 468 1.0423 0.6087 1.0423 1.0209
No log 15.6667 470 1.0304 0.6286 1.0304 1.0151
No log 15.7333 472 1.0380 0.6294 1.0380 1.0188
No log 15.8 474 1.0401 0.6573 1.0401 1.0199
No log 15.8667 476 1.0584 0.6528 1.0584 1.0288
No log 15.9333 478 1.0328 0.6429 1.0328 1.0163
No log 16.0 480 1.0546 0.5857 1.0546 1.0270
No log 16.0667 482 0.9367 0.6569 0.9367 0.9678
No log 16.1333 484 0.8556 0.6809 0.8556 0.9250
No log 16.2 486 0.7790 0.6950 0.7790 0.8826
No log 16.2667 488 0.7406 0.7067 0.7406 0.8606
No log 16.3333 490 0.7535 0.7152 0.7535 0.8681
No log 16.4 492 0.7950 0.7239 0.7950 0.8917
No log 16.4667 494 0.7505 0.7152 0.7505 0.8663
No log 16.5333 496 0.7882 0.7162 0.7882 0.8878
No log 16.6 498 0.7679 0.6950 0.7679 0.8763
0.4294 16.6667 500 0.7739 0.7092 0.7739 0.8797
0.4294 16.7333 502 0.8086 0.7211 0.8086 0.8992
0.4294 16.8 504 0.8198 0.7320 0.8198 0.9055
0.4294 16.8667 506 0.7220 0.7211 0.7220 0.8497
0.4294 16.9333 508 0.6467 0.6993 0.6467 0.8042
0.4294 17.0 510 0.6416 0.7260 0.6416 0.8010
0.4294 17.0667 512 0.6567 0.6993 0.6567 0.8104
0.4294 17.1333 514 0.7371 0.7092 0.7371 0.8586
0.4294 17.2 516 0.8437 0.6714 0.8437 0.9186
0.4294 17.2667 518 0.8167 0.6857 0.8167 0.9037
0.4294 17.3333 520 0.7419 0.7429 0.7419 0.8614
0.4294 17.4 522 0.7112 0.7376 0.7112 0.8433
0.4294 17.4667 524 0.6915 0.7465 0.6915 0.8316
0.4294 17.5333 526 0.7046 0.7413 0.7046 0.8394
0.4294 17.6 528 0.7537 0.7013 0.7537 0.8682
0.4294 17.6667 530 0.7943 0.7168 0.7943 0.8912
0.4294 17.7333 532 0.7788 0.7229 0.7788 0.8825
0.4294 17.8 534 0.6975 0.7397 0.6975 0.8352
0.4294 17.8667 536 0.6556 0.7671 0.6556 0.8097
0.4294 17.9333 538 0.6665 0.7671 0.6665 0.8164
0.4294 18.0 540 0.6862 0.7552 0.6862 0.8284
0.4294 18.0667 542 0.7533 0.7338 0.7533 0.8679
0.4294 18.1333 544 0.8203 0.6715 0.8203 0.9057
0.4294 18.2 546 0.8031 0.6715 0.8031 0.8962
0.4294 18.2667 548 0.7971 0.6715 0.7971 0.8928
0.4294 18.3333 550 0.7312 0.7482 0.7312 0.8551
0.4294 18.4 552 0.6905 0.7482 0.6905 0.8310
0.4294 18.4667 554 0.6886 0.7429 0.6886 0.8298
0.4294 18.5333 556 0.7507 0.7260 0.7507 0.8664
0.4294 18.6 558 0.8793 0.6452 0.8793 0.9377
0.4294 18.6667 560 1.0092 0.6824 1.0092 1.0046
0.4294 18.7333 562 0.9699 0.6164 0.9699 0.9848
0.4294 18.8 564 0.8808 0.6714 0.8808 0.9385
0.4294 18.8667 566 0.8249 0.6716 0.8249 0.9082

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k6_task1_organization

Finetuned
(4023)
this model