ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k4_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7385
  • Qwk: 0.7568
  • Mse: 0.7385
  • Rmse: 0.8594

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0952 2 6.7695 0.0308 6.7695 2.6018
No log 0.1905 4 4.4600 0.0402 4.4600 2.1119
No log 0.2857 6 2.6965 0.0497 2.6965 1.6421
No log 0.3810 8 1.9946 0.1849 1.9946 1.4123
No log 0.4762 10 3.1843 0.0377 3.1843 1.7845
No log 0.5714 12 3.2548 0.0838 3.2548 1.8041
No log 0.6667 14 1.8642 0.2281 1.8642 1.3653
No log 0.7619 16 1.7095 0.0784 1.7095 1.3075
No log 0.8571 18 1.9469 0.0962 1.9469 1.3953
No log 0.9524 20 2.7546 -0.0405 2.7546 1.6597
No log 1.0476 22 4.0823 -0.0417 4.0823 2.0205
No log 1.1429 24 4.3971 -0.0731 4.3971 2.0969
No log 1.2381 26 3.8000 0.0460 3.8000 1.9494
No log 1.3333 28 2.9829 0.0261 2.9829 1.7271
No log 1.4286 30 2.4070 -0.0134 2.4070 1.5515
No log 1.5238 32 2.0422 0.3387 2.0422 1.4291
No log 1.6190 34 1.9670 0.3276 1.9670 1.4025
No log 1.7143 36 1.8597 0.2037 1.8597 1.3637
No log 1.8095 38 1.7444 0.1538 1.7444 1.3208
No log 1.9048 40 1.7899 0.0784 1.7899 1.3379
No log 2.0 42 1.9260 0.0962 1.9260 1.3878
No log 2.0952 44 2.1543 0.1818 2.1543 1.4678
No log 2.1905 46 2.2190 0.1000 2.2190 1.4896
No log 2.2857 48 2.2490 0.0559 2.2490 1.4997
No log 2.3810 50 2.2646 0.0274 2.2646 1.5049
No log 2.4762 52 2.1397 0.2029 2.1397 1.4628
No log 2.5714 54 2.0395 0.2727 2.0395 1.4281
No log 2.6667 56 1.8241 0.2857 1.8241 1.3506
No log 2.7619 58 1.7069 0.1682 1.7069 1.3065
No log 2.8571 60 1.5533 0.1346 1.5533 1.2463
No log 2.9524 62 1.4705 0.1714 1.4705 1.2126
No log 3.0476 64 1.4180 0.2075 1.4180 1.1908
No log 3.1429 66 1.3424 0.3063 1.3425 1.1586
No log 3.2381 68 1.2577 0.3363 1.2577 1.1215
No log 3.3333 70 1.2127 0.4202 1.2127 1.1012
No log 3.4286 72 1.0874 0.5000 1.0874 1.0428
No log 3.5238 74 0.9981 0.4918 0.9981 0.9990
No log 3.6190 76 1.0286 0.5669 1.0286 1.0142
No log 3.7143 78 0.9567 0.5827 0.9567 0.9781
No log 3.8095 80 1.0409 0.5954 1.0409 1.0202
No log 3.9048 82 1.5905 0.4088 1.5905 1.2611
No log 4.0 84 1.7395 0.3768 1.7395 1.3189
No log 4.0952 86 1.2423 0.4818 1.2423 1.1146
No log 4.1905 88 0.8376 0.6619 0.8376 0.9152
No log 4.2857 90 1.0714 0.6061 1.0714 1.0351
No log 4.3810 92 1.2522 0.4885 1.2522 1.1190
No log 4.4762 94 1.0317 0.5625 1.0317 1.0157
No log 4.5714 96 0.7674 0.7164 0.7674 0.8760
No log 4.6667 98 0.7052 0.7376 0.7052 0.8397
No log 4.7619 100 0.6993 0.7273 0.6993 0.8362
No log 4.8571 102 0.6698 0.75 0.6698 0.8184
No log 4.9524 104 0.7612 0.6667 0.7612 0.8724
No log 5.0476 106 1.0882 0.5714 1.0882 1.0432
No log 5.1429 108 1.2334 0.5324 1.2334 1.1106
No log 5.2381 110 1.0945 0.5612 1.0945 1.0462
No log 5.3333 112 0.8377 0.6423 0.8377 0.9153
No log 5.4286 114 0.7159 0.7042 0.7159 0.8461
No log 5.5238 116 0.6957 0.7534 0.6957 0.8341
No log 5.6190 118 0.6781 0.7552 0.6781 0.8235
No log 5.7143 120 0.6781 0.6906 0.6781 0.8234
No log 5.8095 122 0.7723 0.6906 0.7723 0.8788
No log 5.9048 124 0.8694 0.6294 0.8694 0.9324
No log 6.0 126 0.8196 0.6331 0.8196 0.9053
No log 6.0952 128 0.8622 0.6331 0.8622 0.9286
No log 6.1905 130 0.8344 0.6525 0.8344 0.9135
No log 6.2857 132 0.8850 0.6479 0.8850 0.9407
No log 6.3810 134 0.8815 0.6525 0.8815 0.9389
No log 6.4762 136 0.8027 0.6806 0.8027 0.8959
No log 6.5714 138 0.7600 0.7682 0.7600 0.8718
No log 6.6667 140 0.7528 0.7483 0.7528 0.8676
No log 6.7619 142 0.6903 0.7448 0.6903 0.8308
No log 6.8571 144 0.6471 0.7211 0.6471 0.8045
No log 6.9524 146 0.6113 0.7517 0.6113 0.7819
No log 7.0476 148 0.5927 0.7532 0.5927 0.7699
No log 7.1429 150 0.6046 0.7805 0.6046 0.7775
No log 7.2381 152 0.6243 0.7711 0.6243 0.7902
No log 7.3333 154 0.6305 0.7619 0.6305 0.7940
No log 7.4286 156 0.7045 0.7848 0.7045 0.8393
No log 7.5238 158 0.7381 0.76 0.7381 0.8591
No log 7.6190 160 0.7060 0.7448 0.7060 0.8402
No log 7.7143 162 0.7053 0.7143 0.7053 0.8398
No log 7.8095 164 0.7350 0.6950 0.7350 0.8573
No log 7.9048 166 0.6516 0.7183 0.6516 0.8072
No log 8.0 168 0.6001 0.7843 0.6001 0.7747
No log 8.0952 170 0.6021 0.7895 0.6021 0.7760
No log 8.1905 172 0.5758 0.7949 0.5758 0.7588
No log 8.2857 174 0.5738 0.7582 0.5738 0.7575
No log 8.3810 176 0.5571 0.7582 0.5571 0.7464
No log 8.4762 178 0.5807 0.7673 0.5807 0.7620
No log 8.5714 180 0.6031 0.7625 0.6031 0.7766
No log 8.6667 182 0.5892 0.7673 0.5892 0.7676
No log 8.7619 184 0.6271 0.7532 0.6271 0.7919
No log 8.8571 186 0.5815 0.7613 0.5815 0.7626
No log 8.9524 188 0.5470 0.7871 0.5470 0.7396
No log 9.0476 190 0.5757 0.8052 0.5757 0.7588
No log 9.1429 192 0.5668 0.7792 0.5668 0.7529
No log 9.2381 194 0.6234 0.7413 0.6234 0.7895
No log 9.3333 196 0.6098 0.7413 0.6098 0.7809
No log 9.4286 198 0.5770 0.7606 0.5770 0.7596
No log 9.5238 200 0.5651 0.7552 0.5651 0.7517
No log 9.6190 202 0.5519 0.7671 0.5519 0.7429
No log 9.7143 204 0.5367 0.7895 0.5367 0.7326
No log 9.8095 206 0.5539 0.7895 0.5539 0.7442
No log 9.9048 208 0.5777 0.8075 0.5777 0.7601
No log 10.0 210 0.6370 0.7975 0.6370 0.7981
No log 10.0952 212 0.7095 0.7468 0.7095 0.8423
No log 10.1905 214 0.8624 0.7051 0.8624 0.9286
No log 10.2857 216 0.8563 0.6950 0.8563 0.9254
No log 10.3810 218 0.7331 0.7143 0.7331 0.8562
No log 10.4762 220 0.6560 0.7891 0.6560 0.8099
No log 10.5714 222 0.6230 0.7712 0.6230 0.7893
No log 10.6667 224 0.5922 0.7792 0.5922 0.7695
No log 10.7619 226 0.5724 0.7742 0.5724 0.7566
No log 10.8571 228 0.5630 0.8129 0.5630 0.7503
No log 10.9524 230 0.5370 0.7974 0.5370 0.7328
No log 11.0476 232 0.5392 0.8079 0.5392 0.7343
No log 11.1429 234 0.5398 0.7673 0.5398 0.7347
No log 11.2381 236 0.5494 0.775 0.5494 0.7412
No log 11.3333 238 0.5498 0.7848 0.5498 0.7415
No log 11.4286 240 0.5859 0.7451 0.5859 0.7654
No log 11.5238 242 0.6035 0.7483 0.6035 0.7768
No log 11.6190 244 0.6177 0.7483 0.6177 0.7860
No log 11.7143 246 0.5653 0.7871 0.5653 0.7519
No log 11.8095 248 0.5683 0.8302 0.5683 0.7538
No log 11.9048 250 0.5507 0.8025 0.5507 0.7421
No log 12.0 252 0.5450 0.7785 0.5450 0.7382
No log 12.0952 254 0.5972 0.7643 0.5972 0.7728
No log 12.1905 256 0.6325 0.7550 0.6325 0.7953
No log 12.2857 258 0.6057 0.7771 0.6057 0.7783
No log 12.3810 260 0.5333 0.8079 0.5333 0.7303
No log 12.4762 262 0.5452 0.8052 0.5452 0.7384
No log 12.5714 264 0.5937 0.8077 0.5937 0.7705
No log 12.6667 266 0.5825 0.8153 0.5825 0.7632
No log 12.7619 268 0.5218 0.7975 0.5218 0.7224
No log 12.8571 270 0.5077 0.8171 0.5077 0.7125
No log 12.9524 272 0.6256 0.7692 0.6256 0.7910
No log 13.0476 274 0.6905 0.7485 0.6905 0.8310
No log 13.1429 276 0.6701 0.7564 0.6701 0.8186
No log 13.2381 278 0.6278 0.7898 0.6278 0.7923
No log 13.3333 280 0.6381 0.7898 0.6381 0.7988
No log 13.4286 282 0.6597 0.7821 0.6597 0.8122
No log 13.5238 284 0.6414 0.7815 0.6414 0.8009
No log 13.6190 286 0.6556 0.7815 0.6556 0.8097
No log 13.7143 288 0.6647 0.7397 0.6647 0.8153
No log 13.8095 290 0.6898 0.7733 0.6898 0.8305
No log 13.9048 292 0.6639 0.7568 0.6639 0.8148
No log 14.0 294 0.6335 0.7843 0.6335 0.7960
No log 14.0952 296 0.6424 0.7843 0.6424 0.8015
No log 14.1905 298 0.6225 0.8025 0.6225 0.7890
No log 14.2857 300 0.6221 0.7821 0.6221 0.7888
No log 14.3810 302 0.6095 0.7929 0.6095 0.7807
No log 14.4762 304 0.5859 0.7765 0.5859 0.7655
No log 14.5714 306 0.5542 0.8293 0.5542 0.7445
No log 14.6667 308 0.5517 0.8129 0.5517 0.7428
No log 14.7619 310 0.5557 0.8129 0.5557 0.7455
No log 14.8571 312 0.5825 0.7875 0.5825 0.7632
No log 14.9524 314 0.6121 0.76 0.6121 0.7824
No log 15.0476 316 0.6199 0.7483 0.6199 0.7873
No log 15.1429 318 0.6096 0.76 0.6096 0.7808
No log 15.2381 320 0.6312 0.7673 0.6312 0.7945
No log 15.3333 322 0.5858 0.7778 0.5858 0.7654
No log 15.4286 324 0.5977 0.7673 0.5977 0.7731
No log 15.5238 326 0.5815 0.7927 0.5815 0.7626
No log 15.6190 328 0.5978 0.7949 0.5978 0.7732
No log 15.7143 330 0.5891 0.7949 0.5891 0.7675
No log 15.8095 332 0.5869 0.8101 0.5869 0.7661
No log 15.9048 334 0.5735 0.7871 0.5735 0.7573
No log 16.0 336 0.6088 0.7799 0.6088 0.7803
No log 16.0952 338 0.7193 0.7882 0.7193 0.8481
No log 16.1905 340 0.6931 0.7929 0.6931 0.8325
No log 16.2857 342 0.5784 0.7898 0.5784 0.7605
No log 16.3810 344 0.5497 0.7974 0.5497 0.7414
No log 16.4762 346 0.5580 0.8105 0.5580 0.7470
No log 16.5714 348 0.5613 0.8105 0.5613 0.7492
No log 16.6667 350 0.5445 0.8129 0.5445 0.7379
No log 16.7619 352 0.5326 0.8025 0.5326 0.7298
No log 16.8571 354 0.5240 0.8391 0.5240 0.7239
No log 16.9524 356 0.5208 0.8444 0.5208 0.7216
No log 17.0476 358 0.5071 0.8508 0.5071 0.7121
No log 17.1429 360 0.4968 0.8144 0.4968 0.7048
No log 17.2381 362 0.5128 0.8171 0.5128 0.7161
No log 17.3333 364 0.5276 0.8171 0.5276 0.7263
No log 17.4286 366 0.5485 0.8171 0.5485 0.7406
No log 17.5238 368 0.5786 0.8098 0.5786 0.7606
No log 17.6190 370 0.6026 0.8 0.6026 0.7763
No log 17.7143 372 0.6323 0.8 0.6323 0.7952
No log 17.8095 374 0.6955 0.7654 0.6955 0.8339
No log 17.9048 376 0.7623 0.7123 0.7623 0.8731
No log 18.0 378 0.7965 0.7123 0.7965 0.8925
No log 18.0952 380 0.7915 0.7123 0.7915 0.8897
No log 18.1905 382 0.7827 0.7123 0.7827 0.8847
No log 18.2857 384 0.7412 0.7310 0.7412 0.8609
No log 18.3810 386 0.7094 0.7550 0.7094 0.8422
No log 18.4762 388 0.6901 0.7613 0.6901 0.8307
No log 18.5714 390 0.6838 0.7468 0.6838 0.8269
No log 18.6667 392 0.6820 0.7468 0.6820 0.8259
No log 18.7619 394 0.6412 0.7712 0.6412 0.8008
No log 18.8571 396 0.6206 0.76 0.6206 0.7878
No log 18.9524 398 0.6183 0.7703 0.6183 0.7863
No log 19.0476 400 0.6032 0.7651 0.6032 0.7767
No log 19.1429 402 0.5920 0.7867 0.5920 0.7694
No log 19.2381 404 0.5810 0.8101 0.5810 0.7622
No log 19.3333 406 0.5802 0.8148 0.5802 0.7617
No log 19.4286 408 0.6095 0.7952 0.6095 0.7807
No log 19.5238 410 0.6731 0.7784 0.6731 0.8204
No log 19.6190 412 0.6850 0.7673 0.6850 0.8277
No log 19.7143 414 0.6482 0.7517 0.6482 0.8051
No log 19.8095 416 0.6330 0.7483 0.6330 0.7956
No log 19.9048 418 0.6387 0.7586 0.6387 0.7992
No log 20.0 420 0.6704 0.7286 0.6704 0.8188
No log 20.0952 422 0.7239 0.7286 0.7239 0.8508
No log 20.1905 424 0.7623 0.7211 0.7623 0.8731
No log 20.2857 426 0.7811 0.7355 0.7811 0.8838
No log 20.3810 428 0.7287 0.7590 0.7287 0.8536
No log 20.4762 430 0.6649 0.7765 0.6649 0.8154
No log 20.5714 432 0.6689 0.7746 0.6689 0.8179
No log 20.6667 434 0.7043 0.7746 0.7043 0.8392
No log 20.7619 436 0.7895 0.7386 0.7895 0.8885
No log 20.8571 438 0.8366 0.6957 0.8366 0.9147
No log 20.9524 440 0.7966 0.7211 0.7966 0.8925
No log 21.0476 442 0.7375 0.7432 0.7375 0.8588
No log 21.1429 444 0.7041 0.7632 0.7041 0.8391
No log 21.2381 446 0.6927 0.75 0.6927 0.8323
No log 21.3333 448 0.6967 0.75 0.6967 0.8347
No log 21.4286 450 0.6963 0.7532 0.6963 0.8344
No log 21.5238 452 0.7258 0.7771 0.7258 0.8519
No log 21.6190 454 0.7546 0.7595 0.7546 0.8687
No log 21.7143 456 0.7634 0.7654 0.7634 0.8737
No log 21.8095 458 0.7192 0.7517 0.7192 0.8480
No log 21.9048 460 0.6949 0.76 0.6949 0.8336
No log 22.0 462 0.6836 0.7785 0.6836 0.8268
No log 22.0952 464 0.6747 0.7568 0.6747 0.8214
No log 22.1905 466 0.6690 0.7568 0.6690 0.8179
No log 22.2857 468 0.6583 0.7703 0.6583 0.8114
No log 22.3810 470 0.6653 0.7534 0.6653 0.8156
No log 22.4762 472 0.7020 0.7534 0.7020 0.8378
No log 22.5714 474 0.7725 0.7333 0.7725 0.8789
No log 22.6667 476 0.7546 0.7417 0.7546 0.8687
No log 22.7619 478 0.7242 0.7483 0.7242 0.8510
No log 22.8571 480 0.6884 0.7413 0.6884 0.8297
No log 22.9524 482 0.7072 0.7413 0.7072 0.8409
No log 23.0476 484 0.7474 0.7483 0.7474 0.8645
No log 23.1429 486 0.7717 0.7143 0.7717 0.8785
No log 23.2381 488 0.7489 0.7143 0.7489 0.8654
No log 23.3333 490 0.7434 0.7324 0.7434 0.8622
No log 23.4286 492 0.7098 0.7286 0.7098 0.8425
No log 23.5238 494 0.7023 0.7286 0.7023 0.8380
No log 23.6190 496 0.6869 0.7413 0.6869 0.8288
No log 23.7143 498 0.6647 0.7534 0.6647 0.8153
0.4437 23.8095 500 0.6774 0.7682 0.6774 0.8230
0.4437 23.9048 502 0.6840 0.7643 0.6840 0.8270
0.4437 24.0 504 0.6640 0.7643 0.6640 0.8149
0.4437 24.0952 506 0.6700 0.7683 0.6700 0.8185
0.4437 24.1905 508 0.6404 0.7831 0.6404 0.8002
0.4437 24.2857 510 0.6516 0.7792 0.6516 0.8072
0.4437 24.3810 512 0.6672 0.7792 0.6672 0.8168
0.4437 24.4762 514 0.6787 0.7792 0.6787 0.8238
0.4437 24.5714 516 0.6769 0.7651 0.6769 0.8227
0.4437 24.6667 518 0.6956 0.7534 0.6956 0.8340
0.4437 24.7619 520 0.6988 0.7534 0.6988 0.8360
0.4437 24.8571 522 0.6705 0.7733 0.6705 0.8189
0.4437 24.9524 524 0.6618 0.8077 0.6618 0.8135
0.4437 25.0476 526 0.6642 0.7682 0.6642 0.8150
0.4437 25.1429 528 0.6586 0.7871 0.6586 0.8116
0.4437 25.2381 530 0.6740 0.7662 0.6740 0.8209
0.4437 25.3333 532 0.6868 0.76 0.6868 0.8287
0.4437 25.4286 534 0.7101 0.7568 0.7101 0.8427
0.4437 25.5238 536 0.7234 0.7568 0.7234 0.8505
0.4437 25.6190 538 0.7385 0.7568 0.7385 0.8594

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k4_task1_organization

Finetuned
(4023)
this model