ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k20_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9695
  • Qwk: -0.0862
  • Mse: 0.9695
  • Rmse: 0.9846

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0377 2 3.5098 -0.0047 3.5098 1.8734
No log 0.0755 4 1.9733 0.0672 1.9733 1.4047
No log 0.1132 6 2.0754 0.0500 2.0754 1.4406
No log 0.1509 8 1.5131 -0.0190 1.5131 1.2301
No log 0.1887 10 1.1337 -0.0253 1.1337 1.0648
No log 0.2264 12 0.8506 0.0549 0.8506 0.9223
No log 0.2642 14 0.7762 0.0375 0.7762 0.8810
No log 0.3019 16 1.2061 -0.0049 1.2061 1.0982
No log 0.3396 18 1.2731 0.0176 1.2731 1.1283
No log 0.3774 20 0.9313 0.0089 0.9313 0.9651
No log 0.4151 22 0.8317 0.1150 0.8317 0.9120
No log 0.4528 24 0.8421 0.0207 0.8421 0.9177
No log 0.4906 26 0.7490 -0.0188 0.7490 0.8655
No log 0.5283 28 0.7218 -0.0035 0.7218 0.8496
No log 0.5660 30 0.7002 0.0 0.7002 0.8368
No log 0.6038 32 0.8668 0.0486 0.8668 0.9310
No log 0.6415 34 0.9471 0.0378 0.9471 0.9732
No log 0.6792 36 1.0633 0.0045 1.0633 1.0312
No log 0.7170 38 0.8965 -0.0513 0.8965 0.9468
No log 0.7547 40 0.7568 -0.0774 0.7568 0.8700
No log 0.7925 42 0.7810 0.0229 0.7810 0.8837
No log 0.8302 44 0.8796 0.0609 0.8796 0.9379
No log 0.8679 46 1.0421 0.0305 1.0421 1.0209
No log 0.9057 48 0.7131 0.0914 0.7131 0.8444
No log 0.9434 50 0.8819 -0.0477 0.8819 0.9391
No log 0.9811 52 0.8102 -0.2154 0.8102 0.9001
No log 1.0189 54 0.8183 0.0016 0.8183 0.9046
No log 1.0566 56 1.5303 0.0064 1.5303 1.2371
No log 1.0943 58 1.5198 0.0065 1.5198 1.2328
No log 1.1321 60 0.9645 0.0067 0.9645 0.9821
No log 1.1698 62 0.7428 -0.0551 0.7428 0.8619
No log 1.2075 64 0.8293 -0.0774 0.8293 0.9107
No log 1.2453 66 0.7594 -0.0451 0.7594 0.8714
No log 1.2830 68 0.8365 -0.1191 0.8365 0.9146
No log 1.3208 70 1.0799 -0.0435 1.0799 1.0392
No log 1.3585 72 1.0396 -0.0411 1.0396 1.0196
No log 1.3962 74 0.7765 0.0159 0.7765 0.8812
No log 1.4340 76 0.7097 0.0555 0.7097 0.8424
No log 1.4717 78 0.7198 0.0033 0.7198 0.8484
No log 1.5094 80 0.8199 -0.0477 0.8199 0.9055
No log 1.5472 82 0.8698 0.0319 0.8698 0.9327
No log 1.5849 84 1.1274 -0.0492 1.1274 1.0618
No log 1.6226 86 1.7019 -0.0215 1.7019 1.3046
No log 1.6604 88 1.2443 -0.0572 1.2443 1.1155
No log 1.6981 90 0.8972 0.0178 0.8972 0.9472
No log 1.7358 92 0.8752 0.0139 0.8752 0.9355
No log 1.7736 94 0.9386 0.0062 0.9386 0.9688
No log 1.8113 96 0.9148 0.0113 0.9148 0.9564
No log 1.8491 98 0.8185 0.0053 0.8185 0.9047
No log 1.8868 100 0.8270 0.1094 0.8270 0.9094
No log 1.9245 102 0.9739 -0.0076 0.9739 0.9869
No log 1.9623 104 0.7763 0.0874 0.7763 0.8811
No log 2.0 106 0.7508 0.0141 0.7508 0.8665
No log 2.0377 108 0.7087 0.0768 0.7087 0.8419
No log 2.0755 110 0.9311 0.0576 0.9311 0.9650
No log 2.1132 112 1.0446 0.0666 1.0446 1.0221
No log 2.1509 114 0.7486 0.0588 0.7486 0.8652
No log 2.1887 116 0.7704 0.0637 0.7704 0.8777
No log 2.2264 118 0.9007 0.0169 0.9007 0.9491
No log 2.2642 120 0.7945 0.1379 0.7945 0.8914
No log 2.3019 122 1.3259 0.0561 1.3259 1.1515
No log 2.3396 124 2.0497 0.0335 2.0497 1.4317
No log 2.3774 126 1.5154 0.0082 1.5154 1.2310
No log 2.4151 128 0.8040 0.0611 0.8040 0.8966
No log 2.4528 130 0.7932 -0.0849 0.7932 0.8906
No log 2.4906 132 0.8319 -0.0116 0.8319 0.9121
No log 2.5283 134 0.7340 0.0918 0.7340 0.8567
No log 2.5660 136 0.8069 -0.0274 0.8069 0.8983
No log 2.6038 138 0.9471 0.1493 0.9471 0.9732
No log 2.6415 140 0.8083 0.0525 0.8083 0.8991
No log 2.6792 142 0.7230 0.0922 0.7230 0.8503
No log 2.7170 144 0.7842 -0.0385 0.7842 0.8856
No log 2.7547 146 0.7950 -0.0230 0.7950 0.8916
No log 2.7925 148 0.8331 0.0167 0.8331 0.9127
No log 2.8302 150 0.8197 0.1315 0.8197 0.9054
No log 2.8679 152 0.8328 -0.0187 0.8328 0.9126
No log 2.9057 154 0.8542 -0.0458 0.8542 0.9243
No log 2.9434 156 0.7849 0.0503 0.7849 0.8859
No log 2.9811 158 0.8415 0.0175 0.8415 0.9173
No log 3.0189 160 0.8297 0.0690 0.8297 0.9109
No log 3.0566 162 0.7573 -0.0030 0.7573 0.8702
No log 3.0943 164 0.7499 0.0033 0.7499 0.8660
No log 3.1321 166 0.7384 0.0479 0.7384 0.8593
No log 3.1698 168 0.7529 -0.0520 0.7529 0.8677
No log 3.2075 170 0.7700 -0.0204 0.7700 0.8775
No log 3.2453 172 1.0266 0.0293 1.0266 1.0132
No log 3.2830 174 0.9160 0.0538 0.9160 0.9571
No log 3.3208 176 0.8118 0.0240 0.8118 0.9010
No log 3.3585 178 0.8832 -0.0289 0.8832 0.9398
No log 3.3962 180 0.8583 0.0406 0.8583 0.9264
No log 3.4340 182 0.7667 0.1081 0.7667 0.8756
No log 3.4717 184 0.9201 -0.0008 0.9201 0.9592
No log 3.5094 186 1.1957 0.0296 1.1957 1.0935
No log 3.5472 188 0.9329 -0.0504 0.9329 0.9659
No log 3.5849 190 0.7405 0.0863 0.7405 0.8605
No log 3.6226 192 0.7445 0.0454 0.7445 0.8629
No log 3.6604 194 0.8215 -0.0303 0.8215 0.9064
No log 3.6981 196 0.8896 0.0091 0.8896 0.9432
No log 3.7358 198 0.8810 0.0091 0.8810 0.9386
No log 3.7736 200 0.8838 -0.0008 0.8838 0.9401
No log 3.8113 202 0.7480 0.0205 0.7480 0.8649
No log 3.8491 204 0.7728 0.0279 0.7728 0.8791
No log 3.8868 206 0.7732 0.0664 0.7732 0.8793
No log 3.9245 208 0.7414 0.1181 0.7414 0.8610
No log 3.9623 210 0.9582 0.0175 0.9582 0.9789
No log 4.0 212 0.9222 0.0277 0.9222 0.9603
No log 4.0377 214 0.8597 0.0392 0.8597 0.9272
No log 4.0755 216 0.7894 0.0884 0.7894 0.8885
No log 4.1132 218 0.7992 0.0157 0.7992 0.8940
No log 4.1509 220 0.7726 0.1951 0.7726 0.8790
No log 4.1887 222 0.7944 0.1001 0.7944 0.8913
No log 4.2264 224 0.9072 -0.0471 0.9072 0.9525
No log 4.2642 226 0.8680 0.0016 0.8680 0.9316
No log 4.3019 228 0.7952 0.1687 0.7952 0.8918
No log 4.3396 230 0.8016 0.0647 0.8016 0.8953
No log 4.3774 232 0.7934 0.1007 0.7934 0.8908
No log 4.4151 234 0.8155 0.0956 0.8155 0.9031
No log 4.4528 236 0.7733 0.0989 0.7733 0.8794
No log 4.4906 238 0.7753 0.0220 0.7753 0.8805
No log 4.5283 240 0.7595 0.0481 0.7595 0.8715
No log 4.5660 242 0.7556 0.1433 0.7556 0.8693
No log 4.6038 244 0.7793 0.1047 0.7793 0.8828
No log 4.6415 246 0.7973 0.1047 0.7973 0.8929
No log 4.6792 248 0.7812 0.1199 0.7812 0.8839
No log 4.7170 250 0.7931 0.1807 0.7931 0.8905
No log 4.7547 252 0.8065 0.0680 0.8065 0.8981
No log 4.7925 254 0.8084 0.0639 0.8084 0.8991
No log 4.8302 256 0.7858 -0.1331 0.7858 0.8864
No log 4.8679 258 0.9187 -0.1145 0.9187 0.9585
No log 4.9057 260 0.8257 0.0155 0.8257 0.9087
No log 4.9434 262 0.7143 0.0 0.7143 0.8451
No log 4.9811 264 0.7084 0.0436 0.7084 0.8417
No log 5.0189 266 0.7078 0.0 0.7078 0.8413
No log 5.0566 268 0.7234 0.0 0.7234 0.8505
No log 5.0943 270 0.7301 0.1023 0.7301 0.8545
No log 5.1321 272 0.7476 0.0922 0.7476 0.8646
No log 5.1698 274 0.8015 0.0118 0.8015 0.8952
No log 5.2075 276 0.8081 0.0118 0.8081 0.8989
No log 5.2453 278 0.7607 0.0357 0.7607 0.8722
No log 5.2830 280 0.7298 0.1371 0.7298 0.8543
No log 5.3208 282 0.7100 0.1023 0.7100 0.8426
No log 5.3585 284 0.7274 0.0528 0.7274 0.8529
No log 5.3962 286 0.8611 -0.0761 0.8611 0.9279
No log 5.4340 288 1.0039 0.0226 1.0039 1.0019
No log 5.4717 290 0.8577 -0.0801 0.8577 0.9261
No log 5.5094 292 0.7609 0.1096 0.7609 0.8723
No log 5.5472 294 0.9751 -0.0151 0.9751 0.9875
No log 5.5849 296 0.9218 0.0293 0.9218 0.9601
No log 5.6226 298 0.7597 0.0338 0.7597 0.8716
No log 5.6604 300 0.7960 -0.0144 0.7960 0.8922
No log 5.6981 302 0.8501 0.0050 0.8501 0.9220
No log 5.7358 304 0.7069 -0.0520 0.7069 0.8408
No log 5.7736 306 0.6820 0.1512 0.6820 0.8259
No log 5.8113 308 0.7254 0.1965 0.7254 0.8517
No log 5.8491 310 0.6750 0.1512 0.6750 0.8216
No log 5.8868 312 0.7031 -0.0473 0.7031 0.8385
No log 5.9245 314 0.7199 -0.0473 0.7199 0.8485
No log 5.9623 316 0.7049 0.0033 0.7049 0.8396
No log 6.0 318 0.7745 0.1701 0.7745 0.8801
No log 6.0377 320 0.9065 0.1065 0.9065 0.9521
No log 6.0755 322 0.7934 0.1395 0.7934 0.8907
No log 6.1132 324 0.7454 0.0973 0.7454 0.8634
No log 6.1509 326 0.7548 0.0094 0.7548 0.8688
No log 6.1887 328 0.7339 0.1675 0.7339 0.8567
No log 6.2264 330 0.8326 0.0287 0.8326 0.9125
No log 6.2642 332 0.9613 0.0404 0.9613 0.9805
No log 6.3019 334 0.8385 0.0287 0.8385 0.9157
No log 6.3396 336 0.7336 0.1318 0.7336 0.8565
No log 6.3774 338 0.7738 -0.0334 0.7738 0.8797
No log 6.4151 340 0.7933 -0.0391 0.7933 0.8907
No log 6.4528 342 0.8391 0.0600 0.8391 0.9160
No log 6.4906 344 0.8598 0.0409 0.8598 0.9272
No log 6.5283 346 0.7715 0.0236 0.7715 0.8783
No log 6.5660 348 0.7690 -0.0849 0.7690 0.8769
No log 6.6038 350 0.7647 -0.1470 0.7647 0.8745
No log 6.6415 352 0.7250 0.1023 0.7250 0.8515
No log 6.6792 354 0.7760 0.1097 0.7760 0.8809
No log 6.7170 356 0.8489 0.0017 0.8489 0.9214
No log 6.7547 358 0.8676 -0.0008 0.8676 0.9314
No log 6.7925 360 0.7698 0.1148 0.7698 0.8774
No log 6.8302 362 0.7657 -0.0113 0.7657 0.8750
No log 6.8679 364 0.7582 0.0814 0.7582 0.8708
No log 6.9057 366 0.7325 0.0814 0.7325 0.8559
No log 6.9434 368 0.7287 0.1318 0.7287 0.8536
No log 6.9811 370 0.8618 -0.0408 0.8618 0.9283
No log 7.0189 372 0.9365 -0.0490 0.9365 0.9677
No log 7.0566 374 0.8104 0.0071 0.8104 0.9002
No log 7.0943 376 0.7226 0.0814 0.7226 0.8501
No log 7.1321 378 0.7451 0.0970 0.7451 0.8632
No log 7.1698 380 0.7518 0.0471 0.7518 0.8671
No log 7.2075 382 0.8058 0.0826 0.8058 0.8976
No log 7.2453 384 0.8947 -0.0079 0.8947 0.9459
No log 7.2830 386 0.8782 -0.0079 0.8782 0.9371
No log 7.3208 388 0.7546 0.1097 0.7546 0.8687
No log 7.3585 390 0.7140 0.1318 0.7140 0.8450
No log 7.3962 392 0.7118 0.1081 0.7118 0.8437
No log 7.4340 394 0.7310 0.1318 0.7310 0.8550
No log 7.4717 396 0.8218 0.0442 0.8218 0.9065
No log 7.5094 398 0.8629 -0.0459 0.8629 0.9289
No log 7.5472 400 0.7910 0.1395 0.7910 0.8894
No log 7.5849 402 0.7406 0.1379 0.7406 0.8606
No log 7.6226 404 0.7431 0.1202 0.7431 0.8621
No log 7.6604 406 0.8516 -0.0504 0.8516 0.9228
No log 7.6981 408 0.9170 -0.0617 0.9170 0.9576
No log 7.7358 410 0.9396 -0.0955 0.9396 0.9694
No log 7.7736 412 0.8376 -0.1257 0.8376 0.9152
No log 7.8113 414 0.7693 0.1148 0.7693 0.8771
No log 7.8491 416 0.8011 0.0639 0.8011 0.8950
No log 7.8868 418 0.8484 -0.1257 0.8484 0.9211
No log 7.9245 420 0.8933 -0.0870 0.8933 0.9451
No log 7.9623 422 1.0306 -0.0316 1.0306 1.0152
No log 8.0 424 0.9808 0.0046 0.9808 0.9903
No log 8.0377 426 0.8441 -0.1261 0.8441 0.9187
No log 8.0755 428 0.7867 0.0821 0.7867 0.8870
No log 8.1132 430 0.7814 0.0821 0.7814 0.8840
No log 8.1509 432 0.7955 0.0183 0.7955 0.8919
No log 8.1887 434 0.8141 0.0152 0.8141 0.9023
No log 8.2264 436 0.8498 0.0068 0.8498 0.9219
No log 8.2642 438 0.8504 0.0068 0.8504 0.9222
No log 8.3019 440 0.8287 0.0814 0.8287 0.9103
No log 8.3396 442 0.8049 0.1347 0.8049 0.8971
No log 8.3774 444 0.7894 0.1362 0.7894 0.8885
No log 8.4151 446 0.8695 -0.0390 0.8695 0.9325
No log 8.4528 448 1.0157 -0.0606 1.0157 1.0078
No log 8.4906 450 0.9563 -0.1273 0.9563 0.9779
No log 8.5283 452 0.7981 0.0159 0.7981 0.8934
No log 8.5660 454 0.7475 0.0541 0.7475 0.8646
No log 8.6038 456 0.8068 0.0148 0.8068 0.8982
No log 8.6415 458 0.7774 0.0978 0.7774 0.8817
No log 8.6792 460 0.8304 0.0043 0.8304 0.9113
No log 8.7170 462 1.0242 -0.0306 1.0242 1.0120
No log 8.7547 464 1.0410 -0.0030 1.0410 1.0203
No log 8.7925 466 0.8462 -0.0033 0.8462 0.9199
No log 8.8302 468 0.7293 0.0479 0.7293 0.8540
No log 8.8679 470 0.7268 0.0479 0.7268 0.8525
No log 8.9057 472 0.7388 0.1902 0.7388 0.8595
No log 8.9434 474 0.9124 -0.0532 0.9124 0.9552
No log 8.9811 476 1.0057 -0.1285 1.0057 1.0028
No log 9.0189 478 0.9841 -0.1285 0.9841 0.9920
No log 9.0566 480 0.8461 -0.0056 0.8461 0.9198
No log 9.0943 482 0.8126 0.1449 0.8126 0.9014
No log 9.1321 484 0.8347 0.0909 0.8347 0.9136
No log 9.1698 486 0.8542 0.0409 0.8542 0.9242
No log 9.2075 488 0.9064 -0.0894 0.9064 0.9521
No log 9.2453 490 0.9365 -0.0909 0.9365 0.9677
No log 9.2830 492 0.9296 -0.1271 0.9296 0.9642
No log 9.3208 494 0.8693 -0.0033 0.8693 0.9324
No log 9.3585 496 0.8374 0.0600 0.8374 0.9151
No log 9.3962 498 0.8177 0.0680 0.8177 0.9043
0.3057 9.4340 500 0.8105 0.0639 0.8105 0.9003
0.3057 9.4717 502 0.8886 -0.0504 0.8886 0.9427
0.3057 9.5094 504 1.0266 0.0006 1.0266 1.0132
0.3057 9.5472 506 1.0099 0.0046 1.0099 1.0049
0.3057 9.5849 508 0.8332 0.0909 0.8332 0.9128
0.3057 9.6226 510 0.7177 0.1902 0.7177 0.8471
0.3057 9.6604 512 0.7369 0.1902 0.7369 0.8584
0.3057 9.6981 514 0.7791 0.1902 0.7791 0.8826
0.3057 9.7358 516 0.8624 0.0512 0.8624 0.9286
0.3057 9.7736 518 0.8943 -0.0788 0.8943 0.9457
0.3057 9.8113 520 0.9536 -0.0837 0.9536 0.9765
0.3057 9.8491 522 1.0140 -0.0583 1.0140 1.0070
0.3057 9.8868 524 0.9695 -0.0862 0.9695 0.9846

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k20_task3_organization

Finetuned
(4023)
this model