ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k14_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7267
  • Qwk: 0.7222
  • Mse: 0.7267
  • Rmse: 0.8525

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0312 2 7.0324 -0.0057 7.0324 2.6519
No log 0.0625 4 4.8556 0.0449 4.8556 2.2035
No log 0.0938 6 3.3568 -0.0333 3.3568 1.8322
No log 0.125 8 3.8141 0.0102 3.8141 1.9530
No log 0.1562 10 2.4583 0.1471 2.4583 1.5679
No log 0.1875 12 1.9216 0.1947 1.9216 1.3862
No log 0.2188 14 2.0737 0.1552 2.0737 1.4400
No log 0.25 16 2.6899 0.0523 2.6899 1.6401
No log 0.2812 18 2.8492 0.0488 2.8492 1.6879
No log 0.3125 20 2.1966 0.1986 2.1966 1.4821
No log 0.3438 22 1.6396 0.2632 1.6396 1.2805
No log 0.375 24 1.5032 0.2569 1.5032 1.2260
No log 0.4062 26 1.4669 0.2075 1.4669 1.2112
No log 0.4375 28 1.6304 0.1714 1.6304 1.2769
No log 0.4688 30 1.7132 0.2364 1.7132 1.3089
No log 0.5 32 1.8697 0.2623 1.8697 1.3674
No log 0.5312 34 2.1820 0.0719 2.1820 1.4772
No log 0.5625 36 2.5414 0.0405 2.5414 1.5942
No log 0.5938 38 2.3569 0.0680 2.3569 1.5352
No log 0.625 40 1.7800 0.2439 1.7800 1.3342
No log 0.6562 42 1.3258 0.1869 1.3258 1.1514
No log 0.6875 44 1.5396 0.2456 1.5396 1.2408
No log 0.7188 46 1.7682 0.1273 1.7682 1.3298
No log 0.75 48 2.0295 0.0769 2.0295 1.4246
No log 0.7812 50 1.8323 0.1714 1.8323 1.3536
No log 0.8125 52 1.5512 0.2037 1.5512 1.2455
No log 0.8438 54 1.5842 0.3390 1.5842 1.2586
No log 0.875 56 1.8386 0.2901 1.8386 1.3560
No log 0.9062 58 1.9545 0.3111 1.9545 1.3980
No log 0.9375 60 2.1631 0.2621 2.1631 1.4708
No log 0.9688 62 2.2053 0.2585 2.2053 1.4850
No log 1.0 64 1.7933 0.3262 1.7933 1.3391
No log 1.0312 66 1.4371 0.35 1.4371 1.1988
No log 1.0625 68 1.3164 0.1869 1.3164 1.1474
No log 1.0938 70 1.3403 0.2569 1.3403 1.1577
No log 1.125 72 1.3101 0.2883 1.3101 1.1446
No log 1.1562 74 1.3141 0.3214 1.3141 1.1463
No log 1.1875 76 1.2096 0.3036 1.2096 1.0998
No log 1.2188 78 1.1688 0.4370 1.1688 1.0811
No log 1.25 80 1.1702 0.4590 1.1702 1.0817
No log 1.2812 82 1.1144 0.4923 1.1144 1.0557
No log 1.3125 84 1.0284 0.5821 1.0284 1.0141
No log 1.3438 86 1.0261 0.5926 1.0261 1.0129
No log 1.375 88 1.0674 0.5915 1.0674 1.0332
No log 1.4062 90 0.9452 0.6202 0.9452 0.9722
No log 1.4375 92 1.0353 0.6316 1.0353 1.0175
No log 1.4688 94 1.2560 0.4545 1.2560 1.1207
No log 1.5 96 1.4095 0.4427 1.4095 1.1872
No log 1.5312 98 1.4072 0.4427 1.4072 1.1863
No log 1.5625 100 1.3105 0.4531 1.3105 1.1448
No log 1.5938 102 1.0833 0.5736 1.0833 1.0408
No log 1.625 104 1.0083 0.5581 1.0083 1.0041
No log 1.6562 106 1.0198 0.5692 1.0198 1.0099
No log 1.6875 108 1.1311 0.6029 1.1311 1.0635
No log 1.7188 110 1.2766 0.5850 1.2766 1.1298
No log 1.75 112 1.0012 0.6667 1.0012 1.0006
No log 1.7812 114 0.8489 0.6562 0.8489 0.9213
No log 1.8125 116 0.9975 0.6094 0.9975 0.9988
No log 1.8438 118 1.1415 0.5156 1.1415 1.0684
No log 1.875 120 1.2298 0.5294 1.2298 1.1089
No log 1.9062 122 1.2597 0.4615 1.2597 1.1224
No log 1.9375 124 1.1984 0.5156 1.1984 1.0947
No log 1.9688 126 1.1373 0.5344 1.1373 1.0664
No log 2.0 128 1.0335 0.5802 1.0335 1.0166
No log 2.0312 130 0.8992 0.6667 0.8992 0.9483
No log 2.0625 132 0.8307 0.6667 0.8307 0.9114
No log 2.0938 134 0.8065 0.6512 0.8065 0.8980
No log 2.125 136 0.7778 0.6866 0.7778 0.8819
No log 2.1562 138 0.7608 0.7 0.7608 0.8723
No log 2.1875 140 0.8790 0.6906 0.8790 0.9375
No log 2.2188 142 1.0086 0.6131 1.0086 1.0043
No log 2.25 144 0.9095 0.6074 0.9095 0.9537
No log 2.2812 146 0.7895 0.6861 0.7895 0.8885
No log 2.3125 148 0.8224 0.6716 0.8224 0.9069
No log 2.3438 150 0.8784 0.6617 0.8784 0.9372
No log 2.375 152 0.8794 0.6617 0.8794 0.9377
No log 2.4062 154 0.7841 0.6715 0.7841 0.8855
No log 2.4375 156 0.7572 0.7 0.7572 0.8702
No log 2.4688 158 0.7841 0.7092 0.7841 0.8855
No log 2.5 160 0.7949 0.6857 0.7949 0.8916
No log 2.5312 162 0.7295 0.7534 0.7295 0.8541
No log 2.5625 164 0.7113 0.7448 0.7113 0.8434
No log 2.5938 166 0.7013 0.7361 0.7013 0.8374
No log 2.625 168 0.7532 0.6853 0.7532 0.8679
No log 2.6562 170 0.9391 0.6883 0.9391 0.9691
No log 2.6875 172 1.0388 0.6579 1.0388 1.0192
No log 2.7188 174 0.8880 0.6579 0.8880 0.9423
No log 2.75 176 0.7212 0.7722 0.7212 0.8492
No log 2.7812 178 0.6891 0.7582 0.6891 0.8301
No log 2.8125 180 0.8596 0.6986 0.8596 0.9272
No log 2.8438 182 0.9191 0.6944 0.9191 0.9587
No log 2.875 184 0.9080 0.7042 0.9080 0.9529
No log 2.9062 186 0.8876 0.7260 0.8876 0.9421
No log 2.9375 188 0.8310 0.7310 0.8310 0.9116
No log 2.9688 190 0.8217 0.7643 0.8217 0.9065
No log 3.0 192 0.8325 0.7643 0.8325 0.9124
No log 3.0312 194 0.8458 0.7273 0.8458 0.9197
No log 3.0625 196 0.8772 0.7310 0.8772 0.9366
No log 3.0938 198 0.8621 0.7432 0.8621 0.9285
No log 3.125 200 0.7744 0.7843 0.7744 0.8800
No log 3.1562 202 0.7280 0.7436 0.7280 0.8532
No log 3.1875 204 0.7256 0.7722 0.7256 0.8518
No log 3.2188 206 0.7614 0.7898 0.7614 0.8726
No log 3.25 208 0.7860 0.7742 0.7860 0.8866
No log 3.2812 210 0.7738 0.7771 0.7738 0.8797
No log 3.3125 212 0.7734 0.7722 0.7734 0.8794
No log 3.3438 214 0.7793 0.7451 0.7793 0.8828
No log 3.375 216 0.7965 0.7368 0.7965 0.8925
No log 3.4062 218 0.8237 0.7190 0.8237 0.9076
No log 3.4375 220 0.8447 0.6575 0.8447 0.9191
No log 3.4688 222 0.8064 0.7067 0.8064 0.8980
No log 3.5 224 0.7904 0.7114 0.7904 0.8891
No log 3.5312 226 0.8027 0.6842 0.8027 0.8960
No log 3.5625 228 0.7642 0.7355 0.7642 0.8742
No log 3.5938 230 0.7390 0.7436 0.7390 0.8596
No log 3.625 232 0.7277 0.7949 0.7277 0.8531
No log 3.6562 234 0.7620 0.7733 0.7620 0.8729
No log 3.6875 236 0.8156 0.7123 0.8156 0.9031
No log 3.7188 238 0.8937 0.6621 0.8937 0.9454
No log 3.75 240 0.9318 0.6621 0.9318 0.9653
No log 3.7812 242 0.9071 0.6483 0.9071 0.9524
No log 3.8125 244 0.8627 0.6986 0.8627 0.9288
No log 3.8438 246 0.9460 0.6806 0.9460 0.9726
No log 3.875 248 1.0023 0.6667 1.0023 1.0011
No log 3.9062 250 0.9798 0.6861 0.9798 0.9899
No log 3.9375 252 0.9419 0.6331 0.9419 0.9705
No log 3.9688 254 0.9054 0.6087 0.9054 0.9515
No log 4.0 256 0.7805 0.7682 0.7805 0.8834
No log 4.0312 258 0.7435 0.7692 0.7435 0.8623
No log 4.0625 260 0.7546 0.7152 0.7546 0.8687
No log 4.0938 262 0.7441 0.7368 0.7441 0.8626
No log 4.125 264 0.7348 0.7733 0.7348 0.8572
No log 4.1562 266 0.7826 0.7397 0.7826 0.8846
No log 4.1875 268 0.7934 0.7310 0.7934 0.8907
No log 4.2188 270 0.8607 0.7092 0.8607 0.9277
No log 4.25 272 0.8436 0.7143 0.8436 0.9185
No log 4.2812 274 0.7685 0.7483 0.7685 0.8766
No log 4.3125 276 0.7141 0.7662 0.7141 0.8450
No log 4.3438 278 0.6712 0.7871 0.6712 0.8193
No log 4.375 280 0.6579 0.7922 0.6579 0.8111
No log 4.4062 282 0.6485 0.7922 0.6485 0.8053
No log 4.4375 284 0.6515 0.8025 0.6515 0.8072
No log 4.4688 286 0.7002 0.8025 0.7002 0.8368
No log 4.5 288 0.7127 0.7952 0.7127 0.8442
No log 4.5312 290 0.7208 0.7674 0.7208 0.8490
No log 4.5625 292 0.6808 0.8092 0.6808 0.8251
No log 4.5938 294 0.6769 0.7976 0.6769 0.8227
No log 4.625 296 0.7016 0.7975 0.7016 0.8376
No log 4.6562 298 0.6820 0.8125 0.6820 0.8259
No log 4.6875 300 0.6781 0.8272 0.6781 0.8235
No log 4.7188 302 0.7021 0.8025 0.7021 0.8379
No log 4.75 304 0.7217 0.8025 0.7217 0.8495
No log 4.7812 306 0.8082 0.7075 0.8082 0.8990
No log 4.8125 308 0.8738 0.6853 0.8738 0.9348
No log 4.8438 310 0.8930 0.6567 0.8930 0.9450
No log 4.875 312 0.9070 0.6853 0.9070 0.9524
No log 4.9062 314 0.8745 0.6853 0.8745 0.9351
No log 4.9375 316 0.8759 0.6667 0.8759 0.9359
No log 4.9688 318 0.7816 0.7383 0.7816 0.8841
No log 5.0 320 0.7059 0.8050 0.7059 0.8402
No log 5.0312 322 0.7124 0.8075 0.7124 0.8441
No log 5.0625 324 0.7181 0.8 0.7181 0.8474
No log 5.0938 326 0.7336 0.7871 0.7336 0.8565
No log 5.125 328 0.7439 0.7712 0.7439 0.8625
No log 5.1562 330 0.8019 0.6803 0.8019 0.8955
No log 5.1875 332 0.8473 0.6528 0.8473 0.9205
No log 5.2188 334 0.8150 0.7285 0.8150 0.9028
No log 5.25 336 0.7568 0.7651 0.7568 0.8699
No log 5.2812 338 0.7690 0.7324 0.7690 0.8770
No log 5.3125 340 0.8126 0.7007 0.8126 0.9014
No log 5.3438 342 0.8670 0.6222 0.8670 0.9311
No log 5.375 344 0.8524 0.6176 0.8524 0.9233
No log 5.4062 346 0.8012 0.7172 0.8012 0.8951
No log 5.4375 348 0.7729 0.7448 0.7729 0.8791
No log 5.4688 350 0.7351 0.7895 0.7351 0.8574
No log 5.5 352 0.7026 0.7848 0.7026 0.8382
No log 5.5312 354 0.7602 0.7436 0.7602 0.8719
No log 5.5625 356 0.8272 0.6667 0.8272 0.9095
No log 5.5938 358 0.8078 0.7020 0.8078 0.8988
No log 5.625 360 0.7814 0.7613 0.7814 0.8840
No log 5.6562 362 0.7799 0.7564 0.7799 0.8831
No log 5.6875 364 0.7773 0.7595 0.7773 0.8817
No log 5.7188 366 0.7713 0.775 0.7713 0.8782
No log 5.75 368 0.7537 0.7771 0.7537 0.8682
No log 5.7812 370 0.7654 0.7722 0.7654 0.8749
No log 5.8125 372 0.7979 0.7742 0.7979 0.8933
No log 5.8438 374 0.8611 0.6974 0.8611 0.9280
No log 5.875 376 0.8684 0.7152 0.8684 0.9319
No log 5.9062 378 0.8665 0.7383 0.8665 0.9309
No log 5.9375 380 0.8910 0.6853 0.8910 0.9439
No log 5.9688 382 0.8997 0.6761 0.8997 0.9485
No log 6.0 384 0.8663 0.6667 0.8663 0.9307
No log 6.0312 386 0.8303 0.6892 0.8303 0.9112
No log 6.0625 388 0.7887 0.6892 0.7887 0.8881
No log 6.0938 390 0.7810 0.7075 0.7810 0.8837
No log 6.125 392 0.7458 0.7651 0.7458 0.8636
No log 6.1562 394 0.7193 0.7550 0.7193 0.8481
No log 6.1875 396 0.7159 0.7547 0.7159 0.8461
No log 6.2188 398 0.7183 0.7547 0.7183 0.8476
No log 6.25 400 0.7199 0.7826 0.7199 0.8485
No log 6.2812 402 0.7291 0.7904 0.7291 0.8539
No log 6.3125 404 0.7556 0.7692 0.7556 0.8693
No log 6.3438 406 0.7990 0.7485 0.7990 0.8939
No log 6.375 408 0.7736 0.7711 0.7736 0.8795
No log 6.4062 410 0.7142 0.7904 0.7142 0.8451
No log 6.4375 412 0.7053 0.7904 0.7053 0.8398
No log 6.4688 414 0.6953 0.7904 0.6953 0.8339
No log 6.5 416 0.7238 0.7683 0.7238 0.8508
No log 6.5312 418 0.7914 0.7702 0.7914 0.8896
No log 6.5625 420 0.7697 0.7702 0.7697 0.8773
No log 6.5938 422 0.7094 0.7848 0.7094 0.8423
No log 6.625 424 0.6926 0.775 0.6926 0.8322
No log 6.6562 426 0.7008 0.7826 0.7008 0.8371
No log 6.6875 428 0.6776 0.8171 0.6776 0.8232
No log 6.7188 430 0.6310 0.7950 0.6310 0.7944
No log 6.75 432 0.6410 0.7927 0.6410 0.8007
No log 6.7812 434 0.6892 0.7799 0.6892 0.8302
No log 6.8125 436 0.8088 0.7013 0.8088 0.8993
No log 6.8438 438 0.8722 0.6667 0.8722 0.9339
No log 6.875 440 0.9309 0.6577 0.9309 0.9648
No log 6.9062 442 0.8616 0.6842 0.8616 0.9282
No log 6.9375 444 0.7548 0.7848 0.7548 0.8688
No log 6.9688 446 0.6877 0.7683 0.6877 0.8293
No log 7.0 448 0.6663 0.7778 0.6663 0.8163
No log 7.0312 450 0.6690 0.7778 0.6690 0.8179
No log 7.0625 452 0.6766 0.7778 0.6766 0.8226
No log 7.0938 454 0.6555 0.7950 0.6555 0.8097
No log 7.125 456 0.6736 0.7826 0.6736 0.8207
No log 7.1562 458 0.6435 0.8171 0.6435 0.8022
No log 7.1875 460 0.6418 0.8144 0.6418 0.8011
No log 7.2188 462 0.6376 0.7901 0.6376 0.7985
No log 7.25 464 0.6587 0.8144 0.6587 0.8116
No log 7.2812 466 0.6631 0.8095 0.6631 0.8143
No log 7.3125 468 0.6457 0.8235 0.6457 0.8036
No log 7.3438 470 0.6122 0.8235 0.6122 0.7824
No log 7.375 472 0.6130 0.8166 0.6130 0.7830
No log 7.4062 474 0.6087 0.8242 0.6087 0.7802
No log 7.4375 476 0.6283 0.8148 0.6283 0.7926
No log 7.4688 478 0.6646 0.8 0.6646 0.8153
No log 7.5 480 0.6296 0.7875 0.6296 0.7935
No log 7.5312 482 0.6392 0.75 0.6392 0.7995
No log 7.5625 484 0.7081 0.7297 0.7081 0.8415
No log 7.5938 486 0.7235 0.7211 0.7235 0.8506
No log 7.625 488 0.6921 0.7517 0.6921 0.8319
No log 7.6562 490 0.6731 0.7632 0.6731 0.8204
No log 7.6875 492 0.6419 0.7643 0.6419 0.8012
No log 7.7188 494 0.6141 0.7879 0.6141 0.7836
No log 7.75 496 0.6094 0.8072 0.6094 0.7807
No log 7.7812 498 0.6406 0.7451 0.6406 0.8004
0.4994 7.8125 500 0.6866 0.7517 0.6866 0.8286
0.4994 7.8438 502 0.7427 0.7114 0.7427 0.8618
0.4994 7.875 504 0.7817 0.6712 0.7817 0.8841
0.4994 7.9062 506 0.7481 0.7248 0.7481 0.8649
0.4994 7.9375 508 0.7100 0.7333 0.7100 0.8426
0.4994 7.9688 510 0.6811 0.7625 0.6811 0.8253
0.4994 8.0 512 0.7139 0.7561 0.7139 0.8449
0.4994 8.0312 514 0.7270 0.7595 0.7270 0.8526
0.4994 8.0625 516 0.7493 0.7949 0.7493 0.8656
0.4994 8.0938 518 0.7535 0.7949 0.7535 0.8680
0.4994 8.125 520 0.7268 0.7922 0.7268 0.8526
0.4994 8.1562 522 0.7255 0.7755 0.7255 0.8518
0.4994 8.1875 524 0.7326 0.7586 0.7326 0.8559
0.4994 8.2188 526 0.6903 0.7671 0.6903 0.8308
0.4994 8.25 528 0.6783 0.7692 0.6783 0.8236
0.4994 8.2812 530 0.6704 0.7898 0.6704 0.8188
0.4994 8.3125 532 0.6392 0.7712 0.6392 0.7995
0.4994 8.3438 534 0.6932 0.7179 0.6932 0.8326
0.4994 8.375 536 0.8118 0.6835 0.8118 0.9010
0.4994 8.4062 538 0.7776 0.6839 0.7776 0.8818
0.4994 8.4375 540 0.6837 0.7324 0.6837 0.8269
0.4994 8.4688 542 0.7267 0.7222 0.7267 0.8525

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k14_task1_organization

Finetuned
(4023)
this model