ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k4_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9410
  • Qwk: 0.4289
  • Mse: 0.9410
  • Rmse: 0.9701

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1333 2 4.6100 0.0010 4.6100 2.1471
No log 0.2667 4 2.6311 -0.0559 2.6311 1.6221
No log 0.4 6 1.7611 0.0504 1.7611 1.3271
No log 0.5333 8 1.2604 0.0494 1.2604 1.1227
No log 0.6667 10 1.1834 0.2293 1.1834 1.0878
No log 0.8 12 1.1925 0.1444 1.1925 1.0920
No log 0.9333 14 1.1936 0.1711 1.1936 1.0925
No log 1.0667 16 1.2111 0.2784 1.2111 1.1005
No log 1.2 18 1.2200 0.1593 1.2200 1.1045
No log 1.3333 20 1.2727 0.1294 1.2727 1.1281
No log 1.4667 22 1.3613 0.0038 1.3613 1.1667
No log 1.6 24 1.3890 0.0254 1.3890 1.1786
No log 1.7333 26 1.1509 0.2386 1.1509 1.0728
No log 1.8667 28 1.0607 0.2188 1.0607 1.0299
No log 2.0 30 1.2052 0.1271 1.2052 1.0978
No log 2.1333 32 1.2490 0.1271 1.2490 1.1176
No log 2.2667 34 1.2688 0.1171 1.2688 1.1264
No log 2.4 36 1.4033 0.0254 1.4033 1.1846
No log 2.5333 38 1.4159 0.0317 1.4159 1.1899
No log 2.6667 40 1.2829 0.0926 1.2829 1.1326
No log 2.8 42 1.3101 0.1014 1.3101 1.1446
No log 2.9333 44 1.2781 0.1173 1.2781 1.1306
No log 3.0667 46 1.2206 0.0130 1.2206 1.1048
No log 3.2 48 1.1716 0.1076 1.1716 1.0824
No log 3.3333 50 1.1117 0.2408 1.1117 1.0543
No log 3.4667 52 1.0679 0.3414 1.0679 1.0334
No log 3.6 54 1.0683 0.3462 1.0683 1.0336
No log 3.7333 56 1.2362 0.1643 1.2362 1.1118
No log 3.8667 58 1.3061 0.1135 1.3061 1.1429
No log 4.0 60 1.2718 0.1046 1.2718 1.1278
No log 4.1333 62 1.2692 0.1395 1.2692 1.1266
No log 4.2667 64 1.1518 0.2439 1.1518 1.0732
No log 4.4 66 0.9817 0.4371 0.9817 0.9908
No log 4.5333 68 0.9208 0.3448 0.9208 0.9596
No log 4.6667 70 0.8961 0.3463 0.8961 0.9466
No log 4.8 72 0.8634 0.4804 0.8634 0.9292
No log 4.9333 74 0.8631 0.3753 0.8631 0.9291
No log 5.0667 76 0.8545 0.3753 0.8545 0.9244
No log 5.2 78 0.8374 0.4780 0.8374 0.9151
No log 5.3333 80 0.8337 0.5690 0.8337 0.9130
No log 5.4667 82 0.8456 0.5849 0.8456 0.9196
No log 5.6 84 0.7954 0.5883 0.7954 0.8918
No log 5.7333 86 0.7800 0.4701 0.7800 0.8832
No log 5.8667 88 0.8426 0.4879 0.8426 0.9180
No log 6.0 90 0.9480 0.4113 0.9480 0.9737
No log 6.1333 92 0.9410 0.4493 0.9410 0.9701
No log 6.2667 94 0.9042 0.4726 0.9042 0.9509
No log 6.4 96 0.8642 0.4964 0.8642 0.9296
No log 6.5333 98 0.8630 0.5064 0.8630 0.9290
No log 6.6667 100 0.8624 0.4784 0.8624 0.9287
No log 6.8 102 0.8630 0.5836 0.8630 0.9290
No log 6.9333 104 0.9044 0.4444 0.9044 0.9510
No log 7.0667 106 0.9009 0.5203 0.9009 0.9491
No log 7.2 108 0.8646 0.5272 0.8646 0.9299
No log 7.3333 110 0.8488 0.5458 0.8488 0.9213
No log 7.4667 112 0.8756 0.4792 0.8756 0.9357
No log 7.6 114 0.9688 0.4735 0.9688 0.9843
No log 7.7333 116 1.0087 0.4256 1.0087 1.0044
No log 7.8667 118 1.1821 0.2876 1.1821 1.0872
No log 8.0 120 1.1692 0.2702 1.1692 1.0813
No log 8.1333 122 0.9831 0.3578 0.9831 0.9915
No log 8.2667 124 0.9303 0.5044 0.9303 0.9645
No log 8.4 126 0.8983 0.4301 0.8983 0.9478
No log 8.5333 128 0.9798 0.4543 0.9798 0.9899
No log 8.6667 130 0.8984 0.5171 0.8984 0.9478
No log 8.8 132 1.0070 0.4147 1.0070 1.0035
No log 8.9333 134 1.2599 0.3805 1.2599 1.1225
No log 9.0667 136 1.3279 0.3590 1.3279 1.1524
No log 9.2 138 1.1382 0.3707 1.1382 1.0669
No log 9.3333 140 0.9972 0.3401 0.9972 0.9986
No log 9.4667 142 1.0461 0.3935 1.0461 1.0228
No log 9.6 144 0.9589 0.4230 0.9589 0.9792
No log 9.7333 146 0.9604 0.4852 0.9604 0.9800
No log 9.8667 148 0.9430 0.4829 0.9430 0.9711
No log 10.0 150 0.8909 0.4277 0.8909 0.9439
No log 10.1333 152 0.8923 0.4930 0.8923 0.9446
No log 10.2667 154 0.9482 0.4983 0.9482 0.9738
No log 10.4 156 1.0116 0.4314 1.0116 1.0058
No log 10.5333 158 1.1163 0.3832 1.1163 1.0566
No log 10.6667 160 1.0874 0.3772 1.0874 1.0428
No log 10.8 162 1.0865 0.3970 1.0865 1.0423
No log 10.9333 164 1.1005 0.4115 1.1005 1.0491
No log 11.0667 166 1.0935 0.3528 1.0935 1.0457
No log 11.2 168 1.0868 0.3445 1.0868 1.0425
No log 11.3333 170 0.9957 0.4430 0.9957 0.9978
No log 11.4667 172 1.0592 0.3243 1.0592 1.0292
No log 11.6 174 1.1153 0.3197 1.1153 1.0561
No log 11.7333 176 0.9662 0.4066 0.9662 0.9830
No log 11.8667 178 1.0411 0.3860 1.0411 1.0203
No log 12.0 180 1.1198 0.4022 1.1198 1.0582
No log 12.1333 182 1.0859 0.3892 1.0859 1.0421
No log 12.2667 184 1.0350 0.3237 1.0350 1.0174
No log 12.4 186 1.0065 0.3333 1.0065 1.0033
No log 12.5333 188 1.0059 0.3584 1.0059 1.0029
No log 12.6667 190 1.0003 0.3816 1.0003 1.0001
No log 12.8 192 1.0171 0.4080 1.0171 1.0085
No log 12.9333 194 1.0377 0.4973 1.0377 1.0187
No log 13.0667 196 0.9280 0.4983 0.9280 0.9633
No log 13.2 198 0.8238 0.4349 0.8238 0.9076
No log 13.3333 200 0.8260 0.5266 0.8260 0.9089
No log 13.4667 202 0.8298 0.5180 0.8298 0.9109
No log 13.6 204 0.7820 0.5335 0.7820 0.8843
No log 13.7333 206 0.8504 0.5041 0.8504 0.9222
No log 13.8667 208 1.0459 0.4705 1.0459 1.0227
No log 14.0 210 1.0746 0.4705 1.0746 1.0366
No log 14.1333 212 0.9603 0.4557 0.9603 0.9799
No log 14.2667 214 0.8415 0.4409 0.8415 0.9173
No log 14.4 216 0.8230 0.4662 0.8230 0.9072
No log 14.5333 218 0.8159 0.5386 0.8159 0.9033
No log 14.6667 220 0.8474 0.5326 0.8474 0.9205
No log 14.8 222 0.9364 0.5 0.9364 0.9677
No log 14.9333 224 0.9144 0.5 0.9144 0.9563
No log 15.0667 226 0.9605 0.4872 0.9605 0.9800
No log 15.2 228 1.0497 0.4902 1.0497 1.0246
No log 15.3333 230 1.0310 0.4743 1.0310 1.0154
No log 15.4667 232 0.8745 0.5192 0.8745 0.9352
No log 15.6 234 0.8047 0.5438 0.8047 0.8971
No log 15.7333 236 0.8486 0.4741 0.8486 0.9212
No log 15.8667 238 0.8371 0.5467 0.8371 0.9150
No log 16.0 240 0.8578 0.4308 0.8578 0.9261
No log 16.1333 242 1.0361 0.4465 1.0361 1.0179
No log 16.2667 244 1.0909 0.4539 1.0909 1.0445
No log 16.4 246 0.9786 0.4726 0.9786 0.9893
No log 16.5333 248 0.9184 0.4253 0.9184 0.9584
No log 16.6667 250 0.9224 0.3648 0.9224 0.9604
No log 16.8 252 0.9065 0.3957 0.9065 0.9521
No log 16.9333 254 0.9588 0.5114 0.9588 0.9792
No log 17.0667 256 0.9770 0.4902 0.9770 0.9884
No log 17.2 258 0.9259 0.4323 0.9259 0.9622
No log 17.3333 260 0.8847 0.4272 0.8847 0.9406
No log 17.4667 262 0.8620 0.4311 0.8620 0.9285
No log 17.6 264 0.8656 0.4760 0.8656 0.9304
No log 17.7333 266 0.9027 0.4037 0.9027 0.9501
No log 17.8667 268 0.9444 0.4249 0.9444 0.9718
No log 18.0 270 0.9621 0.4130 0.9621 0.9809
No log 18.1333 272 0.9627 0.3887 0.9627 0.9812
No log 18.2667 274 0.9661 0.3887 0.9661 0.9829
No log 18.4 276 0.9990 0.5114 0.9990 0.9995
No log 18.5333 278 1.0014 0.3830 1.0014 1.0007
No log 18.6667 280 0.9736 0.3734 0.9736 0.9867
No log 18.8 282 0.9281 0.4108 0.9281 0.9634
No log 18.9333 284 0.8998 0.4013 0.8998 0.9486
No log 19.0667 286 0.8711 0.5073 0.8711 0.9333
No log 19.2 288 0.8671 0.5828 0.8671 0.9312
No log 19.3333 290 0.8735 0.5753 0.8735 0.9346
No log 19.4667 292 0.8636 0.5706 0.8636 0.9293
No log 19.6 294 0.8269 0.5841 0.8269 0.9093
No log 19.7333 296 0.8170 0.5824 0.8170 0.9039
No log 19.8667 298 0.8015 0.5327 0.8015 0.8953
No log 20.0 300 0.7894 0.5854 0.7894 0.8885
No log 20.1333 302 0.8409 0.5245 0.8409 0.9170
No log 20.2667 304 0.9053 0.54 0.9053 0.9515
No log 20.4 306 0.9591 0.5287 0.9591 0.9793
No log 20.5333 308 1.0295 0.4714 1.0295 1.0146
No log 20.6667 310 1.0101 0.4508 1.0101 1.0050
No log 20.8 312 1.0052 0.5045 1.0052 1.0026
No log 20.9333 314 0.9548 0.4869 0.9548 0.9771
No log 21.0667 316 0.9312 0.5245 0.9312 0.9650
No log 21.2 318 0.9196 0.5495 0.9196 0.9589
No log 21.3333 320 0.9410 0.5426 0.9410 0.9700
No log 21.4667 322 0.9102 0.5380 0.9102 0.9541
No log 21.6 324 0.9102 0.5353 0.9102 0.9540
No log 21.7333 326 0.9335 0.5686 0.9335 0.9662
No log 21.8667 328 0.8770 0.5649 0.8770 0.9365
No log 22.0 330 0.8411 0.5102 0.8411 0.9171
No log 22.1333 332 0.8121 0.4435 0.8121 0.9012
No log 22.2667 334 0.8096 0.4644 0.8096 0.8998
No log 22.4 336 0.8402 0.5624 0.8402 0.9166
No log 22.5333 338 0.8537 0.5624 0.8537 0.9240
No log 22.6667 340 0.8495 0.5320 0.8495 0.9217
No log 22.8 342 0.8729 0.5320 0.8729 0.9343
No log 22.9333 344 0.8555 0.5320 0.8555 0.9249
No log 23.0667 346 0.8212 0.5338 0.8212 0.9062
No log 23.2 348 0.8188 0.5833 0.8188 0.9049
No log 23.3333 350 0.8342 0.5756 0.8342 0.9134
No log 23.4667 352 0.9003 0.4808 0.9003 0.9488
No log 23.6 354 0.9341 0.4987 0.9341 0.9665
No log 23.7333 356 0.8822 0.5649 0.8822 0.9393
No log 23.8667 358 0.8448 0.4966 0.8448 0.9191
No log 24.0 360 0.8318 0.4966 0.8318 0.9120
No log 24.1333 362 0.8432 0.5523 0.8432 0.9183
No log 24.2667 364 0.8601 0.5841 0.8601 0.9274
No log 24.4 366 0.8438 0.5756 0.8438 0.9186
No log 24.5333 368 0.8226 0.5756 0.8226 0.9070
No log 24.6667 370 0.8246 0.5864 0.8246 0.9081
No log 24.8 372 0.8947 0.5209 0.8947 0.9459
No log 24.9333 374 1.0001 0.4489 1.0001 1.0000
No log 25.0667 376 1.0249 0.4480 1.0249 1.0124
No log 25.2 378 0.9228 0.4902 0.9228 0.9606
No log 25.3333 380 0.8639 0.5380 0.8639 0.9295
No log 25.4667 382 0.8608 0.4975 0.8608 0.9278
No log 25.6 384 0.8589 0.5304 0.8589 0.9268
No log 25.7333 386 0.8308 0.5387 0.8308 0.9115
No log 25.8667 388 0.8239 0.5408 0.8239 0.9077
No log 26.0 390 0.7987 0.5969 0.7987 0.8937
No log 26.1333 392 0.7834 0.5607 0.7834 0.8851
No log 26.2667 394 0.7842 0.5721 0.7842 0.8856
No log 26.4 396 0.7818 0.5961 0.7818 0.8842
No log 26.5333 398 0.8347 0.5836 0.8347 0.9136
No log 26.6667 400 0.8941 0.5517 0.8941 0.9455
No log 26.8 402 0.9799 0.5068 0.9799 0.9899
No log 26.9333 404 1.0251 0.4168 1.0251 1.0125
No log 27.0667 406 0.9571 0.4685 0.9571 0.9783
No log 27.2 408 0.9266 0.48 0.9266 0.9626
No log 27.3333 410 0.9194 0.48 0.9194 0.9589
No log 27.4667 412 0.9293 0.48 0.9293 0.9640
No log 27.6 414 0.9289 0.48 0.9289 0.9638
No log 27.7333 416 0.8848 0.5320 0.8848 0.9407
No log 27.8667 418 0.8243 0.5725 0.8243 0.9079
No log 28.0 420 0.7908 0.4965 0.7908 0.8893
No log 28.1333 422 0.7964 0.5359 0.7964 0.8924
No log 28.2667 424 0.8199 0.6098 0.8199 0.9055
No log 28.4 426 0.8157 0.5545 0.8157 0.9032
No log 28.5333 428 0.8252 0.5359 0.8252 0.9084
No log 28.6667 430 0.8648 0.5905 0.8648 0.9299
No log 28.8 432 0.9107 0.5320 0.9107 0.9543
No log 28.9333 434 0.9054 0.5624 0.9054 0.9515
No log 29.0667 436 0.9135 0.5733 0.9135 0.9558
No log 29.2 438 0.9043 0.5733 0.9043 0.9509
No log 29.3333 440 0.9075 0.5759 0.9075 0.9526
No log 29.4667 442 0.9501 0.5733 0.9501 0.9747
No log 29.6 444 0.9857 0.4052 0.9857 0.9928
No log 29.7333 446 1.0037 0.4102 1.0037 1.0019
No log 29.8667 448 1.0203 0.3734 1.0203 1.0101
No log 30.0 450 0.9614 0.3647 0.9614 0.9805
No log 30.1333 452 0.9077 0.4476 0.9077 0.9528
No log 30.2667 454 0.9058 0.4413 0.9058 0.9517
No log 30.4 456 0.9281 0.4700 0.9281 0.9634
No log 30.5333 458 0.9260 0.4067 0.9260 0.9623
No log 30.6667 460 0.9250 0.3860 0.9250 0.9617
No log 30.8 462 0.9373 0.4607 0.9373 0.9681
No log 30.9333 464 0.9462 0.4790 0.9462 0.9728
No log 31.0667 466 0.8979 0.5847 0.8979 0.9476
No log 31.2 468 0.8467 0.6131 0.8467 0.9202
No log 31.3333 470 0.8431 0.5935 0.8431 0.9182
No log 31.4667 472 0.8382 0.5935 0.8382 0.9156
No log 31.6 474 0.8236 0.5848 0.8236 0.9075
No log 31.7333 476 0.8247 0.5853 0.8247 0.9081
No log 31.8667 478 0.8215 0.5853 0.8215 0.9064
No log 32.0 480 0.8315 0.5858 0.8315 0.9119
No log 32.1333 482 0.9249 0.4808 0.9249 0.9617
No log 32.2667 484 0.9736 0.4475 0.9736 0.9867
No log 32.4 486 0.9206 0.4986 0.9206 0.9595
No log 32.5333 488 0.8734 0.5650 0.8734 0.9346
No log 32.6667 490 0.8281 0.5815 0.8281 0.9100
No log 32.8 492 0.8157 0.5011 0.8157 0.9032
No log 32.9333 494 0.8305 0.5451 0.8305 0.9113
No log 33.0667 496 0.8679 0.5561 0.8679 0.9316
No log 33.2 498 0.9060 0.5313 0.9060 0.9518
0.3251 33.3333 500 0.9864 0.4890 0.9864 0.9932
0.3251 33.4667 502 1.0006 0.4890 1.0006 1.0003
0.3251 33.6 504 0.9353 0.4808 0.9353 0.9671
0.3251 33.7333 506 0.8838 0.5537 0.8838 0.9401
0.3251 33.8667 508 0.8611 0.5239 0.8611 0.9280
0.3251 34.0 510 0.8751 0.5264 0.8751 0.9355
0.3251 34.1333 512 0.9098 0.5313 0.9098 0.9538
0.3251 34.2667 514 0.9412 0.4890 0.9412 0.9701
0.3251 34.4 516 0.9317 0.4902 0.9317 0.9652
0.3251 34.5333 518 0.8852 0.5227 0.8852 0.9408
0.3251 34.6667 520 0.8406 0.5674 0.8406 0.9168
0.3251 34.8 522 0.8269 0.5259 0.8269 0.9094
0.3251 34.9333 524 0.8299 0.5202 0.8299 0.9110
0.3251 35.0667 526 0.8771 0.5537 0.8771 0.9365
0.3251 35.2 528 0.9175 0.5515 0.9175 0.9578
0.3251 35.3333 530 0.8768 0.5539 0.8768 0.9364
0.3251 35.4667 532 0.8190 0.5298 0.8190 0.9050
0.3251 35.6 534 0.8031 0.5233 0.8031 0.8961
0.3251 35.7333 536 0.8124 0.4671 0.8124 0.9013
0.3251 35.8667 538 0.8253 0.4540 0.8253 0.9084
0.3251 36.0 540 0.8763 0.5311 0.8763 0.9361
0.3251 36.1333 542 1.0204 0.4135 1.0204 1.0102
0.3251 36.2667 544 1.1368 0.3726 1.1368 1.0662
0.3251 36.4 546 1.1155 0.3719 1.1155 1.0562
0.3251 36.5333 548 1.0138 0.4219 1.0138 1.0069
0.3251 36.6667 550 0.9051 0.4983 0.9051 0.9514
0.3251 36.8 552 0.8565 0.4852 0.8565 0.9255
0.3251 36.9333 554 0.8403 0.4852 0.8403 0.9167
0.3251 37.0667 556 0.8338 0.5011 0.8338 0.9131
0.3251 37.2 558 0.8384 0.5621 0.8384 0.9157
0.3251 37.3333 560 0.8569 0.5385 0.8569 0.9257
0.3251 37.4667 562 0.9470 0.5320 0.9470 0.9731
0.3251 37.6 564 1.0215 0.4597 1.0215 1.0107
0.3251 37.7333 566 1.0229 0.4751 1.0229 1.0114
0.3251 37.8667 568 0.9416 0.52 0.9416 0.9704
0.3251 38.0 570 0.9227 0.4822 0.9227 0.9606
0.3251 38.1333 572 0.9216 0.4822 0.9216 0.9600
0.3251 38.2667 574 0.9104 0.4864 0.9104 0.9541
0.3251 38.4 576 0.9145 0.4334 0.9145 0.9563
0.3251 38.5333 578 0.9410 0.4289 0.9410 0.9701

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k4_task2_organization

Finetuned
(4019)
this model