ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k9_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8645
  • Qwk: 0.4203
  • Mse: 0.8645
  • Rmse: 0.9298

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0690 2 4.8028 0.0010 4.8028 2.1915
No log 0.1379 4 2.6949 -0.0084 2.6949 1.6416
No log 0.2069 6 1.8071 0.0198 1.8071 1.3443
No log 0.2759 8 1.3700 0.0226 1.3700 1.1705
No log 0.3448 10 1.5007 0.1143 1.5007 1.2250
No log 0.4138 12 1.2546 0.0547 1.2546 1.1201
No log 0.4828 14 1.2334 0.2023 1.2334 1.1106
No log 0.5517 16 1.2049 0.1918 1.2049 1.0977
No log 0.6207 18 1.1504 0.2369 1.1504 1.0726
No log 0.6897 20 1.1352 0.3038 1.1352 1.0655
No log 0.7586 22 1.1618 0.3131 1.1618 1.0779
No log 0.8276 24 1.1830 0.1679 1.1830 1.0877
No log 0.8966 26 1.2053 0.1076 1.2053 1.0979
No log 0.9655 28 1.2099 0.1371 1.2099 1.1000
No log 1.0345 30 1.1130 0.3476 1.1130 1.0550
No log 1.1034 32 1.0708 0.3117 1.0708 1.0348
No log 1.1724 34 1.2027 0.1864 1.2027 1.0967
No log 1.2414 36 1.1687 0.2024 1.1687 1.0811
No log 1.3103 38 1.1107 0.3514 1.1107 1.0539
No log 1.3793 40 1.1367 0.2091 1.1367 1.0661
No log 1.4483 42 1.1553 0.2230 1.1553 1.0748
No log 1.5172 44 1.0977 0.3711 1.0977 1.0477
No log 1.5862 46 1.1182 0.3062 1.1182 1.0575
No log 1.6552 48 1.1111 0.3301 1.1111 1.0541
No log 1.7241 50 1.0816 0.4079 1.0816 1.0400
No log 1.7931 52 1.0832 0.2689 1.0832 1.0407
No log 1.8621 54 1.0239 0.3511 1.0239 1.0119
No log 1.9310 56 1.0352 0.3596 1.0352 1.0174
No log 2.0 58 1.2792 0.3238 1.2792 1.1310
No log 2.0690 60 1.2630 0.3798 1.2630 1.1238
No log 2.1379 62 1.0276 0.4396 1.0276 1.0137
No log 2.2069 64 0.9486 0.4042 0.9486 0.9740
No log 2.2759 66 0.9747 0.3256 0.9747 0.9873
No log 2.3448 68 0.9687 0.3434 0.9687 0.9842
No log 2.4138 70 0.9023 0.4104 0.9023 0.9499
No log 2.4828 72 0.9282 0.3645 0.9282 0.9634
No log 2.5517 74 0.9316 0.3278 0.9316 0.9652
No log 2.6207 76 0.9459 0.3145 0.9459 0.9726
No log 2.6897 78 0.9186 0.3616 0.9186 0.9584
No log 2.7586 80 1.0329 0.3333 1.0329 1.0163
No log 2.8276 82 1.2082 0.3757 1.2082 1.0992
No log 2.8966 84 1.1056 0.3757 1.1056 1.0515
No log 2.9655 86 0.8821 0.4278 0.8821 0.9392
No log 3.0345 88 0.9693 0.4406 0.9693 0.9845
No log 3.1034 90 0.9475 0.5006 0.9475 0.9734
No log 3.1724 92 0.8476 0.4272 0.8476 0.9206
No log 3.2414 94 1.1127 0.4462 1.1127 1.0548
No log 3.3103 96 1.0877 0.4216 1.0877 1.0429
No log 3.3793 98 0.8746 0.5012 0.8746 0.9352
No log 3.4483 100 0.8346 0.5025 0.8346 0.9136
No log 3.5172 102 0.8253 0.4737 0.8253 0.9084
No log 3.5862 104 0.8888 0.5339 0.8888 0.9428
No log 3.6552 106 0.8957 0.5083 0.8957 0.9464
No log 3.7241 108 0.8262 0.5308 0.8262 0.9090
No log 3.7931 110 0.8126 0.4962 0.8126 0.9014
No log 3.8621 112 1.0756 0.4872 1.0756 1.0371
No log 3.9310 114 0.8697 0.5485 0.8697 0.9326
No log 4.0 116 0.7507 0.4393 0.7507 0.8664
No log 4.0690 118 0.7519 0.4893 0.7519 0.8671
No log 4.1379 120 0.7748 0.4444 0.7748 0.8802
No log 4.2069 122 0.8084 0.4757 0.8084 0.8991
No log 4.2759 124 0.8055 0.4902 0.8055 0.8975
No log 4.3448 126 0.9066 0.4166 0.9066 0.9522
No log 4.4138 128 1.0091 0.4567 1.0091 1.0045
No log 4.4828 130 0.9798 0.4013 0.9798 0.9898
No log 4.5517 132 0.8857 0.4757 0.8857 0.9411
No log 4.6207 134 0.9096 0.4767 0.9096 0.9537
No log 4.6897 136 1.1056 0.3552 1.1056 1.0515
No log 4.7586 138 1.3370 0.3467 1.3370 1.1563
No log 4.8276 140 1.3182 0.3026 1.3182 1.1481
No log 4.8966 142 1.1045 0.2439 1.1045 1.0510
No log 4.9655 144 1.0200 0.2028 1.0200 1.0099
No log 5.0345 146 1.0688 0.2604 1.0688 1.0338
No log 5.1034 148 1.1705 0.3046 1.1705 1.0819
No log 5.1724 150 1.2542 0.3772 1.2542 1.1199
No log 5.2414 152 1.1898 0.4186 1.1898 1.0908
No log 5.3103 154 1.0088 0.3400 1.0088 1.0044
No log 5.3793 156 0.9547 0.3948 0.9547 0.9771
No log 5.4483 158 0.9323 0.4256 0.9323 0.9655
No log 5.5172 160 0.9181 0.4749 0.9181 0.9582
No log 5.5862 162 0.9289 0.3705 0.9289 0.9638
No log 5.6552 164 0.9364 0.3914 0.9364 0.9677
No log 5.7241 166 0.8977 0.4311 0.8977 0.9475
No log 5.7931 168 0.8536 0.4916 0.8536 0.9239
No log 5.8621 170 0.8963 0.5759 0.8963 0.9467
No log 5.9310 172 0.8579 0.5409 0.8579 0.9262
No log 6.0 174 0.8392 0.4653 0.8392 0.9161
No log 6.0690 176 0.9970 0.4760 0.9970 0.9985
No log 6.1379 178 1.1829 0.4186 1.1829 1.0876
No log 6.2069 180 1.0593 0.4685 1.0593 1.0292
No log 6.2759 182 0.8568 0.4142 0.8568 0.9256
No log 6.3448 184 0.8520 0.4142 0.8520 0.9230
No log 6.4138 186 0.8932 0.4203 0.8932 0.9451
No log 6.4828 188 0.8499 0.3663 0.8499 0.9219
No log 6.5517 190 0.8398 0.3943 0.8398 0.9164
No log 6.6207 192 0.8220 0.3695 0.8220 0.9067
No log 6.6897 194 0.8244 0.3695 0.8244 0.9080
No log 6.7586 196 0.8857 0.4526 0.8857 0.9411
No log 6.8276 198 1.0403 0.4776 1.0403 1.0199
No log 6.8966 200 1.0081 0.4577 1.0081 1.0041
No log 6.9655 202 0.7911 0.5012 0.7911 0.8894
No log 7.0345 204 0.7448 0.5793 0.7448 0.8630
No log 7.1034 206 0.7406 0.5793 0.7406 0.8606
No log 7.1724 208 0.7680 0.5318 0.7680 0.8764
No log 7.2414 210 0.7841 0.5060 0.7841 0.8855
No log 7.3103 212 0.8579 0.5129 0.8579 0.9262
No log 7.3793 214 0.8458 0.4565 0.8458 0.9197
No log 7.4483 216 0.8055 0.4163 0.8055 0.8975
No log 7.5172 218 0.7980 0.3448 0.7980 0.8933
No log 7.5862 220 0.7887 0.4743 0.7887 0.8881
No log 7.6552 222 0.8017 0.4845 0.8017 0.8954
No log 7.7241 224 0.8497 0.5058 0.8497 0.9218
No log 7.7931 226 0.8842 0.5290 0.8842 0.9403
No log 7.8621 228 0.7944 0.4858 0.7944 0.8913
No log 7.9310 230 0.7619 0.3868 0.7619 0.8729
No log 8.0 232 0.8076 0.4871 0.8076 0.8987
No log 8.0690 234 0.9930 0.5501 0.9930 0.9965
No log 8.1379 236 1.0983 0.4731 1.0983 1.0480
No log 8.2069 238 0.9909 0.5247 0.9909 0.9955
No log 8.2759 240 0.8188 0.4439 0.8188 0.9049
No log 8.3448 242 0.7643 0.4772 0.7643 0.8742
No log 8.4138 244 0.7498 0.5120 0.7498 0.8659
No log 8.4828 246 0.7469 0.4993 0.7469 0.8642
No log 8.5517 248 0.7806 0.5043 0.7806 0.8835
No log 8.6207 250 0.8361 0.5182 0.8361 0.9144
No log 8.6897 252 0.8062 0.5045 0.8062 0.8979
No log 8.7586 254 0.7840 0.4491 0.7840 0.8854
No log 8.8276 256 0.8036 0.4719 0.8036 0.8965
No log 8.8966 258 0.8289 0.4752 0.8289 0.9105
No log 8.9655 260 0.8035 0.4297 0.8035 0.8964
No log 9.0345 262 0.8008 0.4724 0.8008 0.8949
No log 9.1034 264 0.9008 0.5287 0.9008 0.9491
No log 9.1724 266 0.9839 0.5707 0.9839 0.9919
No log 9.2414 268 0.9199 0.5421 0.9199 0.9591
No log 9.3103 270 0.8293 0.5239 0.8293 0.9106
No log 9.3793 272 0.8083 0.4626 0.8083 0.8991
No log 9.4483 274 0.8375 0.3762 0.8375 0.9151
No log 9.5172 276 0.9384 0.4920 0.9384 0.9687
No log 9.5862 278 1.0473 0.4513 1.0473 1.0234
No log 9.6552 280 1.0164 0.4658 1.0164 1.0082
No log 9.7241 282 0.9215 0.3401 0.9215 0.9600
No log 9.7931 284 0.8537 0.3493 0.8537 0.9240
No log 9.8621 286 0.8252 0.4220 0.8252 0.9084
No log 9.9310 288 0.8082 0.3796 0.8082 0.8990
No log 10.0 290 0.8343 0.4845 0.8343 0.9134
No log 10.0690 292 0.9317 0.4845 0.9317 0.9652
No log 10.1379 294 0.9287 0.4845 0.9287 0.9637
No log 10.2069 296 0.8422 0.5090 0.8422 0.9177
No log 10.2759 298 0.8000 0.4039 0.8000 0.8945
No log 10.3448 300 0.7924 0.4429 0.7924 0.8902
No log 10.4138 302 0.7811 0.4429 0.7811 0.8838
No log 10.4828 304 0.8008 0.4377 0.8008 0.8949
No log 10.5517 306 0.8980 0.4712 0.8980 0.9476
No log 10.6207 308 0.9254 0.4681 0.9254 0.9620
No log 10.6897 310 0.8685 0.5073 0.8685 0.9319
No log 10.7586 312 0.8593 0.4681 0.8593 0.9270
No log 10.8276 314 0.8957 0.4334 0.8957 0.9464
No log 10.8966 316 0.9753 0.4094 0.9753 0.9876
No log 10.9655 318 1.0771 0.4471 1.0771 1.0378
No log 11.0345 320 1.1521 0.4511 1.1521 1.0734
No log 11.1034 322 1.0695 0.4620 1.0695 1.0342
No log 11.1724 324 0.8887 0.5 0.8887 0.9427
No log 11.2414 326 0.7786 0.4435 0.7786 0.8824
No log 11.3103 328 0.7879 0.4197 0.7879 0.8876
No log 11.3793 330 0.7741 0.4424 0.7741 0.8799
No log 11.4483 332 0.7521 0.4846 0.7521 0.8672
No log 11.5172 334 0.8473 0.5750 0.8473 0.9205
No log 11.5862 336 1.0206 0.5392 1.0206 1.0103
No log 11.6552 338 0.9998 0.5659 0.9998 0.9999
No log 11.7241 340 0.8303 0.5427 0.8303 0.9112
No log 11.7931 342 0.7380 0.4877 0.7380 0.8591
No log 11.8621 344 0.7747 0.5376 0.7747 0.8801
No log 11.9310 346 0.8083 0.4991 0.8083 0.8990
No log 12.0 348 0.7872 0.4598 0.7872 0.8873
No log 12.0690 350 0.7804 0.4734 0.7804 0.8834
No log 12.1379 352 0.8333 0.3941 0.8333 0.9128
No log 12.2069 354 0.9432 0.5129 0.9432 0.9712
No log 12.2759 356 1.0248 0.5041 1.0248 1.0123
No log 12.3448 358 0.9964 0.5506 0.9964 0.9982
No log 12.4138 360 0.9598 0.5649 0.9598 0.9797
No log 12.4828 362 0.8938 0.5899 0.8938 0.9454
No log 12.5517 364 0.8255 0.5418 0.8255 0.9086
No log 12.6207 366 0.8012 0.4672 0.8012 0.8951
No log 12.6897 368 0.7969 0.4882 0.7969 0.8927
No log 12.7586 370 0.8025 0.4864 0.8025 0.8958
No log 12.8276 372 0.7906 0.4829 0.7906 0.8892
No log 12.8966 374 0.8001 0.5085 0.8001 0.8945
No log 12.9655 376 0.8042 0.4964 0.8042 0.8968
No log 13.0345 378 0.8125 0.4637 0.8125 0.9014
No log 13.1034 380 0.8520 0.4465 0.8520 0.9231
No log 13.1724 382 0.8690 0.4754 0.8690 0.9322
No log 13.2414 384 0.8869 0.4743 0.8869 0.9417
No log 13.3103 386 0.8560 0.4902 0.8560 0.9252
No log 13.3793 388 0.8169 0.4316 0.8169 0.9038
No log 13.4483 390 0.8281 0.4328 0.8281 0.9100
No log 13.5172 392 0.8186 0.4527 0.8186 0.9047
No log 13.5862 394 0.8018 0.4413 0.8018 0.8955
No log 13.6552 396 0.8488 0.5324 0.8488 0.9213
No log 13.7241 398 0.9151 0.5190 0.9151 0.9566
No log 13.7931 400 0.9270 0.4681 0.9270 0.9628
No log 13.8621 402 0.8934 0.5015 0.8934 0.9452
No log 13.9310 404 0.8317 0.5283 0.8317 0.9120
No log 14.0 406 0.7989 0.5211 0.7989 0.8938
No log 14.0690 408 0.7864 0.5211 0.7864 0.8868
No log 14.1379 410 0.8026 0.5012 0.8026 0.8959
No log 14.2069 412 0.8592 0.5192 0.8592 0.9269
No log 14.2759 414 0.8906 0.5303 0.8906 0.9437
No log 14.3448 416 0.9011 0.4889 0.9011 0.9493
No log 14.4138 418 0.9152 0.4202 0.9152 0.9567
No log 14.4828 420 0.9105 0.4202 0.9105 0.9542
No log 14.5517 422 0.8923 0.3854 0.8923 0.9446
No log 14.6207 424 0.8902 0.3896 0.8902 0.9435
No log 14.6897 426 0.8897 0.3896 0.8897 0.9432
No log 14.7586 428 0.8860 0.4373 0.8860 0.9413
No log 14.8276 430 0.8924 0.4465 0.8924 0.9447
No log 14.8966 432 0.8806 0.4369 0.8806 0.9384
No log 14.9655 434 0.8528 0.4601 0.8528 0.9235
No log 15.0345 436 0.8554 0.4139 0.8554 0.9249
No log 15.1034 438 0.8592 0.4042 0.8592 0.9269
No log 15.1724 440 0.8590 0.3497 0.8590 0.9268
No log 15.2414 442 0.9043 0.4242 0.9043 0.9510
No log 15.3103 444 0.9616 0.4302 0.9616 0.9806
No log 15.3793 446 0.9736 0.4553 0.9736 0.9867
No log 15.4483 448 0.9535 0.4560 0.9535 0.9765
No log 15.5172 450 0.9189 0.4560 0.9189 0.9586
No log 15.5862 452 0.9043 0.4546 0.9043 0.9509
No log 15.6552 454 0.9155 0.4968 0.9155 0.9568
No log 15.7241 456 0.9018 0.4968 0.9018 0.9496
No log 15.7931 458 0.8995 0.4968 0.8995 0.9484
No log 15.8621 460 0.8905 0.4553 0.8905 0.9436
No log 15.9310 462 0.9049 0.3861 0.9049 0.9513
No log 16.0 464 0.9273 0.3298 0.9273 0.9629
No log 16.0690 466 0.9227 0.4302 0.9227 0.9606
No log 16.1379 468 0.8876 0.4302 0.8876 0.9421
No log 16.2069 470 0.8459 0.3609 0.8459 0.9197
No log 16.2759 472 0.8147 0.3974 0.8147 0.9026
No log 16.3448 474 0.7973 0.5317 0.7973 0.8929
No log 16.4138 476 0.7729 0.5393 0.7729 0.8791
No log 16.4828 478 0.7303 0.5538 0.7303 0.8546
No log 16.5517 480 0.7222 0.5596 0.7222 0.8498
No log 16.6207 482 0.7100 0.6163 0.7100 0.8426
No log 16.6897 484 0.7038 0.6073 0.7038 0.8389
No log 16.7586 486 0.7033 0.5721 0.7033 0.8386
No log 16.8276 488 0.7151 0.5712 0.7151 0.8456
No log 16.8966 490 0.7133 0.5633 0.7133 0.8446
No log 16.9655 492 0.7619 0.5424 0.7619 0.8729
No log 17.0345 494 0.8352 0.5470 0.8352 0.9139
No log 17.1034 496 0.9234 0.4781 0.9234 0.9609
No log 17.1724 498 0.9345 0.4463 0.9345 0.9667
0.3385 17.2414 500 0.9547 0.4663 0.9547 0.9771
0.3385 17.3103 502 0.9758 0.3579 0.9758 0.9878
0.3385 17.3793 504 0.9539 0.4255 0.9539 0.9767
0.3385 17.4483 506 0.8962 0.3434 0.8962 0.9467
0.3385 17.5172 508 0.8662 0.3577 0.8662 0.9307
0.3385 17.5862 510 0.8463 0.5094 0.8463 0.9199
0.3385 17.6552 512 0.8309 0.5029 0.8309 0.9115
0.3385 17.7241 514 0.8166 0.5029 0.8166 0.9037
0.3385 17.7931 516 0.8065 0.5600 0.8065 0.8981
0.3385 17.8621 518 0.7644 0.5585 0.7644 0.8743
0.3385 17.9310 520 0.7284 0.5451 0.7284 0.8535
0.3385 18.0 522 0.7336 0.5274 0.7336 0.8565
0.3385 18.0690 524 0.7498 0.4879 0.7498 0.8659
0.3385 18.1379 526 0.7720 0.5266 0.7720 0.8786
0.3385 18.2069 528 0.7868 0.5188 0.7868 0.8870
0.3385 18.2759 530 0.8073 0.4158 0.8073 0.8985
0.3385 18.3448 532 0.8080 0.4220 0.8080 0.8989
0.3385 18.4138 534 0.8064 0.4722 0.8064 0.8980
0.3385 18.4828 536 0.8141 0.4946 0.8141 0.9023
0.3385 18.5517 538 0.8225 0.4946 0.8225 0.9069
0.3385 18.6207 540 0.8458 0.4946 0.8458 0.9197
0.3385 18.6897 542 0.8379 0.4946 0.8379 0.9154
0.3385 18.7586 544 0.8039 0.4656 0.8039 0.8966
0.3385 18.8276 546 0.8007 0.4321 0.8007 0.8948
0.3385 18.8966 548 0.8156 0.4656 0.8156 0.9031
0.3385 18.9655 550 0.8509 0.4243 0.8509 0.9224
0.3385 19.0345 552 0.8645 0.4203 0.8645 0.9298

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
4
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k9_task2_organization

Finetuned
(4023)
this model