ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k12_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7787
  • Qwk: 0.0394
  • Mse: 0.7787
  • Rmse: 0.8824

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0606 2 3.8789 0.0200 3.8789 1.9695
No log 0.1212 4 1.8833 0.0531 1.8833 1.3723
No log 0.1818 6 1.3819 0.0323 1.3819 1.1756
No log 0.2424 8 0.7212 0.1318 0.7212 0.8493
No log 0.3030 10 0.6952 -0.0101 0.6952 0.8338
No log 0.3636 12 0.9494 -0.0658 0.9494 0.9744
No log 0.4242 14 1.0794 -0.0178 1.0794 1.0390
No log 0.4848 16 0.8391 0.0984 0.8391 0.9160
No log 0.5455 18 0.8363 0.0909 0.8363 0.9145
No log 0.6061 20 1.0356 0.0119 1.0356 1.0176
No log 0.6667 22 1.0620 -0.0149 1.0620 1.0305
No log 0.7273 24 1.2419 -0.0221 1.2419 1.1144
No log 0.7879 26 1.1764 -0.0468 1.1764 1.0846
No log 0.8485 28 0.8880 -0.0269 0.8880 0.9423
No log 0.9091 30 0.6987 -0.0035 0.6987 0.8359
No log 0.9697 32 0.7013 -0.0069 0.7013 0.8374
No log 1.0303 34 0.7279 -0.1765 0.7279 0.8532
No log 1.0909 36 0.8575 -0.1688 0.8575 0.9260
No log 1.1515 38 1.2382 0.0 1.2382 1.1127
No log 1.2121 40 1.4248 0.0 1.4248 1.1936
No log 1.2727 42 1.4831 0.0 1.4831 1.2178
No log 1.3333 44 1.3667 0.0 1.3667 1.1691
No log 1.3939 46 1.2339 0.0 1.2339 1.1108
No log 1.4545 48 1.0530 -0.0221 1.0530 1.0262
No log 1.5152 50 0.8688 -0.1660 0.8688 0.9321
No log 1.5758 52 0.7691 0.0460 0.7691 0.8770
No log 1.6364 54 0.7938 0.0857 0.7938 0.8909
No log 1.6970 56 0.8232 -0.0309 0.8232 0.9073
No log 1.7576 58 0.9762 -0.0638 0.9762 0.9880
No log 1.8182 60 1.2169 -0.0736 1.2169 1.1031
No log 1.8788 62 1.3636 -0.0736 1.3636 1.1677
No log 1.9394 64 1.2378 0.0156 1.2378 1.1126
No log 2.0 66 0.9144 -0.0518 0.9144 0.9562
No log 2.0606 68 0.8071 0.0129 0.8071 0.8984
No log 2.1212 70 0.7701 0.0714 0.7701 0.8776
No log 2.1818 72 0.6908 0.0506 0.6908 0.8312
No log 2.2424 74 0.7735 -0.0351 0.7735 0.8795
No log 2.3030 76 1.1033 0.0446 1.1033 1.0504
No log 2.3636 78 1.4305 -0.0193 1.4305 1.1960
No log 2.4242 80 1.2206 -0.0178 1.2206 1.1048
No log 2.4848 82 0.9148 0.1499 0.9148 0.9565
No log 2.5455 84 0.8554 0.1316 0.8554 0.9249
No log 2.6061 86 0.7348 -0.0264 0.7348 0.8572
No log 2.6667 88 0.8339 0.0409 0.8339 0.9132
No log 2.7273 90 0.8517 0.0442 0.8517 0.9229
No log 2.7879 92 1.1151 0.0282 1.1151 1.0560
No log 2.8485 94 1.1207 -0.0334 1.1207 1.0586
No log 2.9091 96 0.9966 -0.0122 0.9966 0.9983
No log 2.9697 98 1.0005 -0.0122 1.0005 1.0002
No log 3.0303 100 1.1924 0.0111 1.1924 1.0920
No log 3.0909 102 0.7999 -0.0252 0.7999 0.8943
No log 3.1515 104 0.7564 -0.0609 0.7564 0.8697
No log 3.2121 106 0.8826 0.0346 0.8826 0.9395
No log 3.2727 108 0.9427 0.0711 0.9427 0.9710
No log 3.3333 110 0.8551 0.0871 0.8551 0.9247
No log 3.3939 112 1.1278 -0.0218 1.1278 1.0620
No log 3.4545 114 1.0547 -0.0218 1.0547 1.0270
No log 3.5152 116 0.9779 0.0676 0.9779 0.9889
No log 3.5758 118 0.7986 0.0247 0.7986 0.8937
No log 3.6364 120 0.8874 0.1775 0.8874 0.9420
No log 3.6970 122 1.3448 -0.0303 1.3448 1.1597
No log 3.7576 124 0.9689 0.1193 0.9689 0.9843
No log 3.8182 126 0.8290 0.0159 0.8290 0.9105
No log 3.8788 128 0.8209 0.0767 0.8209 0.9061
No log 3.9394 130 0.8405 0.1277 0.8405 0.9168
No log 4.0 132 0.9979 0.0484 0.9979 0.9989
No log 4.0606 134 1.0835 0.0445 1.0835 1.0409
No log 4.1212 136 0.8131 0.1196 0.8131 0.9017
No log 4.1818 138 0.7613 0.0741 0.7613 0.8725
No log 4.2424 140 0.7286 -0.0406 0.7286 0.8536
No log 4.3030 142 0.8332 0.0188 0.8332 0.9128
No log 4.3636 144 0.9183 0.0091 0.9183 0.9583
No log 4.4242 146 0.8044 0.0749 0.8044 0.8969
No log 4.4848 148 0.8176 0.0279 0.8176 0.9042
No log 4.5455 150 0.8560 0.0392 0.8560 0.9252
No log 4.6061 152 0.8010 -0.0238 0.8010 0.8950
No log 4.6667 154 0.9175 -0.0735 0.9175 0.9578
No log 4.7273 156 0.9697 -0.1152 0.9697 0.9847
No log 4.7879 158 0.9116 -0.1098 0.9116 0.9548
No log 4.8485 160 0.8313 0.0995 0.8313 0.9118
No log 4.9091 162 0.8496 0.0319 0.8496 0.9217
No log 4.9697 164 0.8010 -0.0583 0.8010 0.8950
No log 5.0303 166 0.9949 -0.0425 0.9949 0.9974
No log 5.0909 168 0.9263 -0.0788 0.9263 0.9625
No log 5.1515 170 0.8103 -0.0086 0.8103 0.9002
No log 5.2121 172 0.8368 -0.1033 0.8368 0.9147
No log 5.2727 174 0.8324 -0.1737 0.8324 0.9123
No log 5.3333 176 0.9354 0.0146 0.9354 0.9672
No log 5.3939 178 1.0933 -0.0409 1.0933 1.0456
No log 5.4545 180 0.9291 -0.0686 0.9291 0.9639
No log 5.5152 182 0.8345 -0.0939 0.8345 0.9135
No log 5.5758 184 0.8289 -0.0581 0.8289 0.9105
No log 5.6364 186 0.8085 -0.0059 0.8085 0.8992
No log 5.6970 188 0.8527 -0.0672 0.8527 0.9234
No log 5.7576 190 0.9368 -0.0033 0.9368 0.9679
No log 5.8182 192 0.8558 -0.0672 0.8558 0.9251
No log 5.8788 194 0.8241 -0.0992 0.8241 0.9078
No log 5.9394 196 0.8918 -0.1512 0.8918 0.9444
No log 6.0 198 0.9000 -0.1512 0.9000 0.9487
No log 6.0606 200 0.8555 -0.1394 0.8555 0.9249
No log 6.1212 202 0.8281 -0.0599 0.8281 0.9100
No log 6.1818 204 0.9878 -0.0030 0.9878 0.9939
No log 6.2424 206 1.0584 -0.0175 1.0584 1.0288
No log 6.3030 208 0.8316 -0.1187 0.8316 0.9119
No log 6.3636 210 0.7853 -0.0451 0.7853 0.8862
No log 6.4242 212 0.8068 -0.1263 0.8068 0.8982
No log 6.4848 214 0.7905 -0.0578 0.7905 0.8891
No log 6.5455 216 0.9071 -0.0336 0.9071 0.9524
No log 6.6061 218 1.0784 -0.0513 1.0784 1.0385
No log 6.6667 220 0.8516 -0.1191 0.8516 0.9228
No log 6.7273 222 0.8082 -0.1529 0.8082 0.8990
No log 6.7879 224 0.7930 0.0471 0.7930 0.8905
No log 6.8485 226 0.9460 -0.0054 0.9460 0.9726
No log 6.9091 228 1.0251 0.0193 1.0251 1.0124
No log 6.9697 230 0.7747 -0.0274 0.7747 0.8801
No log 7.0303 232 0.7445 0.0094 0.7445 0.8629
No log 7.0909 234 0.7310 0.0094 0.7310 0.8550
No log 7.1515 236 0.7277 0.1318 0.7277 0.8531
No log 7.2121 238 0.7947 0.0867 0.7947 0.8914
No log 7.2727 240 0.7448 0.0863 0.7448 0.8630
No log 7.3333 242 0.7468 0.0863 0.7468 0.8642
No log 7.3939 244 0.8231 0.0068 0.8231 0.9072
No log 7.4545 246 0.9344 0.0250 0.9344 0.9666
No log 7.5152 248 0.9314 0.0277 0.9314 0.9651
No log 7.5758 250 0.8521 0.0966 0.8521 0.9231
No log 7.6364 252 0.9315 0.0016 0.9315 0.9652
No log 7.6970 254 1.4084 0.0188 1.4084 1.1868
No log 7.7576 256 1.9951 0.0112 1.9951 1.4125
No log 7.8182 258 1.7596 0.0389 1.7596 1.3265
No log 7.8788 260 1.0971 0.0129 1.0971 1.0474
No log 7.9394 262 0.7919 0.0444 0.7919 0.8899
No log 8.0 264 0.8352 -0.0116 0.8352 0.9139
No log 8.0606 266 0.7582 -0.0532 0.7582 0.8708
No log 8.1212 268 0.8664 -0.0056 0.8664 0.9308
No log 8.1818 270 1.1329 -0.0013 1.1329 1.0644
No log 8.2424 272 0.9984 0.0157 0.9984 0.9992
No log 8.3030 274 0.7350 0.0334 0.7350 0.8573
No log 8.3636 276 0.7445 -0.0499 0.7445 0.8628
No log 8.4242 278 0.7537 -0.0499 0.7537 0.8682
No log 8.4848 280 0.7274 0.0909 0.7274 0.8529
No log 8.5455 282 0.9392 0.0711 0.9392 0.9691
No log 8.6061 284 1.0525 0.0157 1.0525 1.0259
No log 8.6667 286 0.8203 0.0409 0.8203 0.9057
No log 8.7273 288 0.7253 0.1371 0.7253 0.8517
No log 8.7879 290 0.7349 0.1030 0.7349 0.8573
No log 8.8485 292 0.7195 0.2195 0.7195 0.8483
No log 8.9091 294 0.7778 0.0095 0.7778 0.8819
No log 8.9697 296 0.8942 -0.0425 0.8942 0.9456
No log 9.0303 298 1.0277 -0.0138 1.0277 1.0138
No log 9.0909 300 1.0507 0.0526 1.0507 1.0251
No log 9.1515 302 0.8639 -0.0809 0.8639 0.9295
No log 9.2121 304 0.7789 0.0123 0.7789 0.8825
No log 9.2727 306 0.8420 -0.0425 0.8420 0.9176
No log 9.3333 308 1.0964 0.1007 1.0964 1.0471
No log 9.3939 310 1.2446 -0.0101 1.2446 1.1156
No log 9.4545 312 1.0648 0.1042 1.0648 1.0319
No log 9.5152 314 0.8477 -0.0425 0.8477 0.9207
No log 9.5758 316 0.7863 -0.0390 0.7863 0.8868
No log 9.6364 318 0.8214 -0.0408 0.8214 0.9063
No log 9.6970 320 0.9092 -0.0033 0.9092 0.9535
No log 9.7576 322 0.8871 0.0867 0.8871 0.9418
No log 9.8182 324 0.8326 0.0622 0.8326 0.9125
No log 9.8788 326 0.8330 0.0660 0.8330 0.9127
No log 9.9394 328 0.8772 0.0525 0.8772 0.9366
No log 10.0 330 0.9257 0.0392 0.9257 0.9621
No log 10.0606 332 0.8802 0.0456 0.8802 0.9382
No log 10.1212 334 0.7981 0.0639 0.7981 0.8933
No log 10.1818 336 0.7617 0.0323 0.7617 0.8728
No log 10.2424 338 0.7501 0.0863 0.7501 0.8661
No log 10.3030 340 0.8060 -0.0371 0.8060 0.8978
No log 10.3636 342 0.8659 -0.0490 0.8659 0.9306
No log 10.4242 344 0.9371 0.0157 0.9371 0.9680
No log 10.4848 346 0.8592 -0.0459 0.8592 0.9269
No log 10.5455 348 0.8113 -0.0331 0.8113 0.9007
No log 10.6061 350 0.7857 0.0247 0.7857 0.8864
No log 10.6667 352 0.8080 0.0680 0.8080 0.8989
No log 10.7273 354 0.9198 -0.0054 0.9198 0.9591
No log 10.7879 356 0.8642 -0.0008 0.8642 0.9296
No log 10.8485 358 0.7705 0.0247 0.7705 0.8778
No log 10.9091 360 0.7492 0.0918 0.7492 0.8655
No log 10.9697 362 0.7564 0.0869 0.7564 0.8697
No log 11.0303 364 0.7665 0.0282 0.7665 0.8755
No log 11.0909 366 0.7921 -0.0295 0.7921 0.8900
No log 11.1515 368 0.7914 0.0068 0.7914 0.8896
No log 11.2121 370 0.7773 -0.0295 0.7773 0.8817
No log 11.2727 372 0.7980 0.0099 0.7980 0.8933
No log 11.3333 374 0.7812 -0.0295 0.7812 0.8839
No log 11.3939 376 0.7586 -0.0228 0.7586 0.8710
No log 11.4545 378 0.7688 -0.0228 0.7688 0.8768
No log 11.5152 380 0.7933 0.0600 0.7933 0.8907
No log 11.5758 382 0.8995 -0.0054 0.8995 0.9484
No log 11.6364 384 0.9564 -0.0054 0.9564 0.9780
No log 11.6970 386 0.8000 0.0690 0.8000 0.8944
No log 11.7576 388 0.7608 0.0821 0.7608 0.8722
No log 11.8182 390 0.7355 0.1371 0.7355 0.8576
No log 11.8788 392 0.8305 -0.0033 0.8305 0.9113
No log 11.9394 394 1.0072 -0.0236 1.0072 1.0036
No log 12.0 396 0.9435 -0.0218 0.9435 0.9713
No log 12.0606 398 0.7934 0.0099 0.7934 0.8907
No log 12.1212 400 0.7457 0.0914 0.7457 0.8636
No log 12.1818 402 0.7687 0.0914 0.7687 0.8768
No log 12.2424 404 0.7941 -0.0228 0.7941 0.8911
No log 12.3030 406 0.8789 -0.0425 0.8789 0.9375
No log 12.3636 408 0.9392 -0.0894 0.9392 0.9691
No log 12.4242 410 0.9181 -0.0870 0.9181 0.9582
No log 12.4848 412 0.8684 -0.0425 0.8684 0.9319
No log 12.5455 414 0.8585 -0.0373 0.8585 0.9265
No log 12.6061 416 0.8327 -0.0316 0.8327 0.9125
No log 12.6667 418 0.8442 -0.0767 0.8442 0.9188
No log 12.7273 420 0.8633 -0.0799 0.8633 0.9291
No log 12.7879 422 0.8280 -0.0295 0.8280 0.9099
No log 12.8485 424 0.8011 -0.0628 0.8011 0.8950
No log 12.9091 426 0.7987 0.0432 0.7987 0.8937
No log 12.9697 428 0.7957 -0.0658 0.7957 0.8920
No log 13.0303 430 0.8265 -0.0316 0.8265 0.9091
No log 13.0909 432 0.8755 0.0711 0.8755 0.9357
No log 13.1515 434 0.8134 0.0512 0.8134 0.9019
No log 13.2121 436 0.7436 0.1202 0.7436 0.8624
No log 13.2727 438 0.7449 -0.0062 0.7449 0.8631
No log 13.3333 440 0.7647 0.0303 0.7647 0.8745
No log 13.3939 442 0.8632 0.0091 0.8632 0.9291
No log 13.4545 444 0.9534 0.0362 0.9534 0.9764
No log 13.5152 446 1.0639 -0.0118 1.0639 1.0314
No log 13.5758 448 0.9410 -0.0008 0.9410 0.9700
No log 13.6364 450 0.8628 0.0961 0.8628 0.9288
No log 13.6970 452 0.8237 -0.0459 0.8237 0.9076
No log 13.7576 454 0.8140 -0.0076 0.8140 0.9022
No log 13.8182 456 0.8335 -0.0262 0.8335 0.9130
No log 13.8788 458 0.8420 -0.0336 0.8420 0.9176
No log 13.9394 460 0.8314 0.0574 0.8314 0.9118
No log 14.0 462 0.8231 0.0700 0.8231 0.9072
No log 14.0606 464 0.8318 0.0205 0.8318 0.9120
No log 14.1212 466 0.8219 0.0205 0.8219 0.9066
No log 14.1818 468 0.8038 0.0 0.8038 0.8965
No log 14.2424 470 0.8170 0.0157 0.8170 0.9039
No log 14.3030 472 0.7984 -0.0389 0.7984 0.8935
No log 14.3636 474 0.7841 0.0303 0.7841 0.8855
No log 14.4242 476 0.7991 0.0247 0.7991 0.8939
No log 14.4848 478 0.8096 0.1095 0.8096 0.8998
No log 14.5455 480 0.7974 0.1722 0.7974 0.8930
No log 14.6061 482 0.7789 0.1298 0.7789 0.8825
No log 14.6667 484 0.7896 0.0509 0.7896 0.8886
No log 14.7273 486 0.7793 0.0 0.7793 0.8828
No log 14.7879 488 0.8043 -0.0274 0.8043 0.8968
No log 14.8485 490 0.8431 -0.0373 0.8431 0.9182
No log 14.9091 492 0.8383 -0.0355 0.8383 0.9156
No log 14.9697 494 0.8359 0.0118 0.8358 0.9142
No log 15.0303 496 0.8386 0.0504 0.8386 0.9158
No log 15.0909 498 0.8184 0.0741 0.8184 0.9046
0.3339 15.1515 500 0.8148 0.1282 0.8148 0.9027
0.3339 15.2121 502 0.8044 0.1769 0.8044 0.8969
0.3339 15.2727 504 0.7893 0.1705 0.7893 0.8884
0.3339 15.3333 506 0.7732 0.1751 0.7732 0.8793
0.3339 15.3939 508 0.7686 0.0247 0.7686 0.8767
0.3339 15.4545 510 0.7605 0.0814 0.7605 0.8721
0.3339 15.5152 512 0.7527 0.0814 0.7527 0.8676
0.3339 15.5758 514 0.7693 0.1318 0.7693 0.8771
0.3339 15.6364 516 0.8555 0.0316 0.8555 0.9249
0.3339 15.6970 518 0.9081 0.0260 0.9081 0.9529
0.3339 15.7576 520 0.8713 -0.0056 0.8713 0.9335
0.3339 15.8182 522 0.7906 0.0814 0.7906 0.8892
0.3339 15.8788 524 0.7787 0.0394 0.7787 0.8824

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k12_task3_organization

Finetuned
(4023)
this model