ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k11_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8725
  • Qwk: 0.3885
  • Mse: 0.8725
  • Rmse: 0.9341

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0571 2 4.5215 -0.0103 4.5215 2.1264
No log 0.1143 4 2.6020 -0.0285 2.6020 1.6131
No log 0.1714 6 1.7886 0.0062 1.7886 1.3374
No log 0.2286 8 1.4062 0.0393 1.4062 1.1859
No log 0.2857 10 1.3003 0.0623 1.3003 1.1403
No log 0.3429 12 1.2280 0.1593 1.2280 1.1081
No log 0.4 14 1.2470 0.1507 1.2470 1.1167
No log 0.4571 16 1.3825 -0.0436 1.3825 1.1758
No log 0.5143 18 1.5021 0.0488 1.5021 1.2256
No log 0.5714 20 1.2939 0.0449 1.2939 1.1375
No log 0.6286 22 1.1634 0.3041 1.1634 1.0786
No log 0.6857 24 1.1345 0.3307 1.1345 1.0651
No log 0.7429 26 1.1323 0.3195 1.1323 1.0641
No log 0.8 28 1.1059 0.3045 1.1059 1.0516
No log 0.8571 30 1.0681 0.3474 1.0681 1.0335
No log 0.9143 32 1.0732 0.3045 1.0732 1.0360
No log 0.9714 34 1.1604 0.1848 1.1604 1.0772
No log 1.0286 36 1.3546 0.0254 1.3546 1.1639
No log 1.0857 38 1.4846 0.0403 1.4846 1.2184
No log 1.1429 40 1.4045 0.0403 1.4045 1.1851
No log 1.2 42 1.1847 0.1735 1.1847 1.0884
No log 1.2571 44 1.0782 0.4051 1.0782 1.0383
No log 1.3143 46 1.3152 0.0414 1.3152 1.1468
No log 1.3714 48 1.5326 -0.0697 1.5326 1.2380
No log 1.4286 50 1.0530 0.2408 1.0530 1.0261
No log 1.4857 52 0.8911 0.4330 0.8911 0.9440
No log 1.5429 54 1.2631 0.2058 1.2631 1.1239
No log 1.6 56 1.2420 0.2058 1.2420 1.1144
No log 1.6571 58 1.0656 0.3237 1.0656 1.0323
No log 1.7143 60 1.0395 0.2440 1.0395 1.0195
No log 1.7714 62 1.0751 0.2674 1.0751 1.0369
No log 1.8286 64 1.0308 0.2605 1.0308 1.0153
No log 1.8857 66 0.9947 0.4321 0.9947 0.9973
No log 1.9429 68 1.0780 0.2149 1.0780 1.0383
No log 2.0 70 1.1717 0.1920 1.1717 1.0824
No log 2.0571 72 1.2025 0.2014 1.2025 1.0966
No log 2.1143 74 1.2240 0.2014 1.2240 1.1064
No log 2.1714 76 1.0547 0.3577 1.0547 1.0270
No log 2.2286 78 0.9713 0.4371 0.9713 0.9855
No log 2.2857 80 0.9490 0.4075 0.9490 0.9741
No log 2.3429 82 0.9388 0.3117 0.9388 0.9689
No log 2.4 84 0.9797 0.3093 0.9797 0.9898
No log 2.4571 86 1.0101 0.4202 1.0101 1.0051
No log 2.5143 88 1.0294 0.3839 1.0294 1.0146
No log 2.5714 90 1.0944 0.3551 1.0944 1.0461
No log 2.6286 92 1.0702 0.3760 1.0702 1.0345
No log 2.6857 94 1.0153 0.4020 1.0153 1.0076
No log 2.7429 96 0.9587 0.4628 0.9587 0.9791
No log 2.8 98 0.9784 0.3869 0.9784 0.9892
No log 2.8571 100 0.9538 0.4060 0.9538 0.9766
No log 2.9143 102 0.9745 0.36 0.9745 0.9872
No log 2.9714 104 0.9285 0.3892 0.9285 0.9636
No log 3.0286 106 0.9564 0.4110 0.9564 0.9779
No log 3.0857 108 1.1193 0.3556 1.1193 1.0580
No log 3.1429 110 1.0997 0.3225 1.0997 1.0487
No log 3.2 112 1.1102 0.3705 1.1102 1.0537
No log 3.2571 114 1.1494 0.3600 1.1494 1.0721
No log 3.3143 116 1.0250 0.2976 1.0250 1.0124
No log 3.3714 118 0.9901 0.2033 0.9901 0.9950
No log 3.4286 120 1.1101 0.1443 1.1101 1.0536
No log 3.4857 122 0.9265 0.3363 0.9265 0.9625
No log 3.5429 124 0.8254 0.5503 0.8254 0.9085
No log 3.6 126 1.0503 0.4275 1.0503 1.0249
No log 3.6571 128 1.1482 0.4138 1.1482 1.0715
No log 3.7143 130 0.9075 0.5792 0.9075 0.9526
No log 3.7714 132 0.9415 0.4788 0.9415 0.9703
No log 3.8286 134 1.0943 0.5148 1.0943 1.0461
No log 3.8857 136 1.0080 0.5677 1.0080 1.0040
No log 3.9429 138 0.9098 0.5997 0.9098 0.9538
No log 4.0 140 1.0645 0.4674 1.0645 1.0317
No log 4.0571 142 1.2179 0.3534 1.2179 1.1036
No log 4.1143 144 1.1002 0.4310 1.1002 1.0489
No log 4.1714 146 0.9183 0.4420 0.9183 0.9583
No log 4.2286 148 0.8952 0.5058 0.8952 0.9461
No log 4.2857 150 0.8991 0.5076 0.8991 0.9482
No log 4.3429 152 0.9069 0.4945 0.9069 0.9523
No log 4.4 154 0.9197 0.4960 0.9197 0.9590
No log 4.4571 156 0.9418 0.4534 0.9418 0.9705
No log 4.5143 158 0.9490 0.4119 0.9490 0.9741
No log 4.5714 160 1.0037 0.3548 1.0037 1.0018
No log 4.6286 162 1.1449 0.4013 1.1449 1.0700
No log 4.6857 164 1.0436 0.3310 1.0436 1.0216
No log 4.7429 166 0.9966 0.3062 0.9966 0.9983
No log 4.8 168 0.9792 0.3608 0.9792 0.9896
No log 4.8571 170 1.0584 0.2752 1.0584 1.0288
No log 4.9143 172 1.0618 0.3281 1.0618 1.0304
No log 4.9714 174 0.9504 0.2871 0.9504 0.9749
No log 5.0286 176 0.9554 0.2864 0.9554 0.9775
No log 5.0857 178 0.9628 0.2188 0.9628 0.9812
No log 5.1429 180 1.0669 0.3816 1.0669 1.0329
No log 5.2 182 1.0706 0.3326 1.0706 1.0347
No log 5.2571 184 0.9991 0.3211 0.9991 0.9995
No log 5.3143 186 0.9487 0.2850 0.9487 0.9740
No log 5.3714 188 0.9465 0.2512 0.9465 0.9729
No log 5.4286 190 0.9655 0.3115 0.9655 0.9826
No log 5.4857 192 0.9079 0.3482 0.9079 0.9528
No log 5.5429 194 0.8951 0.4719 0.8951 0.9461
No log 5.6 196 0.8962 0.5199 0.8962 0.9467
No log 5.6571 198 0.8657 0.4381 0.8657 0.9304
No log 5.7143 200 0.9419 0.3794 0.9419 0.9705
No log 5.7714 202 0.9366 0.3931 0.9366 0.9678
No log 5.8286 204 0.9197 0.4065 0.9197 0.9590
No log 5.8857 206 0.9592 0.3289 0.9592 0.9794
No log 5.9429 208 0.9462 0.3289 0.9462 0.9728
No log 6.0 210 0.9099 0.2995 0.9099 0.9539
No log 6.0571 212 0.9636 0.3897 0.9636 0.9816
No log 6.1143 214 1.1831 0.3863 1.1831 1.0877
No log 6.1714 216 1.3910 0.2895 1.3910 1.1794
No log 6.2286 218 1.3057 0.3222 1.3057 1.1427
No log 6.2857 220 1.0215 0.4010 1.0215 1.0107
No log 6.3429 222 0.8599 0.3697 0.8599 0.9273
No log 6.4 224 0.8327 0.5343 0.8327 0.9125
No log 6.4571 226 0.8395 0.3728 0.8395 0.9162
No log 6.5143 228 0.9203 0.3656 0.9203 0.9593
No log 6.5714 230 1.0192 0.4003 1.0192 1.0095
No log 6.6286 232 1.0037 0.4007 1.0037 1.0018
No log 6.6857 234 0.8898 0.4128 0.8898 0.9433
No log 6.7429 236 0.9085 0.4 0.9085 0.9532
No log 6.8 238 0.9882 0.4220 0.9882 0.9941
No log 6.8571 240 0.9632 0.4219 0.9632 0.9814
No log 6.9143 242 0.9226 0.4010 0.9226 0.9605
No log 6.9714 244 1.0263 0.4890 1.0263 1.0131
No log 7.0286 246 1.0888 0.4835 1.0888 1.0434
No log 7.0857 248 0.9062 0.4458 0.9062 0.9519
No log 7.1429 250 0.7694 0.4778 0.7694 0.8771
No log 7.2 252 0.7823 0.5312 0.7823 0.8845
No log 7.2571 254 0.7827 0.5186 0.7827 0.8847
No log 7.3143 256 0.7934 0.4499 0.7934 0.8908
No log 7.3714 258 0.8124 0.4808 0.8124 0.9014
No log 7.4286 260 0.8755 0.4258 0.8755 0.9357
No log 7.4857 262 0.8081 0.3970 0.8081 0.8990
No log 7.5429 264 0.7954 0.4337 0.7954 0.8918
No log 7.6 266 0.8534 0.4412 0.8534 0.9238
No log 7.6571 268 1.0221 0.3982 1.0221 1.0110
No log 7.7143 270 1.0773 0.3982 1.0773 1.0379
No log 7.7714 272 0.9874 0.4099 0.9874 0.9937
No log 7.8286 274 0.9865 0.4219 0.9865 0.9932
No log 7.8857 276 1.0665 0.4693 1.0665 1.0327
No log 7.9429 278 0.9963 0.3928 0.9963 0.9981
No log 8.0 280 0.9094 0.3841 0.9094 0.9536
No log 8.0571 282 0.9318 0.3211 0.9318 0.9653
No log 8.1143 284 1.0120 0.2998 1.0120 1.0060
No log 8.1714 286 0.9915 0.2843 0.9915 0.9957
No log 8.2286 288 0.9491 0.2843 0.9491 0.9742
No log 8.2857 290 0.8698 0.3196 0.8698 0.9326
No log 8.3429 292 0.8115 0.3890 0.8115 0.9008
No log 8.4 294 0.7884 0.4902 0.7884 0.8879
No log 8.4571 296 0.8711 0.4914 0.8711 0.9333
No log 8.5143 298 0.9619 0.5040 0.9619 0.9807
No log 8.5714 300 0.8813 0.4685 0.8813 0.9388
No log 8.6286 302 0.7786 0.5410 0.7786 0.8824
No log 8.6857 304 0.7844 0.4806 0.7844 0.8857
No log 8.7429 306 0.7853 0.4220 0.7853 0.8862
No log 8.8 308 0.8696 0.3820 0.8696 0.9325
No log 8.8571 310 1.0242 0.5171 1.0242 1.0120
No log 8.9143 312 1.1143 0.4544 1.1143 1.0556
No log 8.9714 314 1.0184 0.3068 1.0184 1.0092
No log 9.0286 316 1.0153 0.2762 1.0153 1.0076
No log 9.0857 318 0.9708 0.3223 0.9708 0.9853
No log 9.1429 320 0.9817 0.3223 0.9817 0.9908
No log 9.2 322 1.0940 0.2857 1.0940 1.0460
No log 9.2571 324 1.1624 0.3341 1.1624 1.0781
No log 9.3143 326 1.1035 0.3138 1.1035 1.0505
No log 9.3714 328 1.1255 0.3231 1.1255 1.0609
No log 9.4286 330 1.1342 0.3474 1.1342 1.0650
No log 9.4857 332 1.1670 0.3726 1.1670 1.0803
No log 9.5429 334 1.1181 0.3138 1.1181 1.0574
No log 9.6 336 1.0463 0.2975 1.0463 1.0229
No log 9.6571 338 0.9874 0.3289 0.9874 0.9937
No log 9.7143 340 0.9502 0.3961 0.9502 0.9748
No log 9.7714 342 0.9422 0.3590 0.9422 0.9706
No log 9.8286 344 0.8834 0.4203 0.8834 0.9399
No log 9.8857 346 0.8382 0.4646 0.8382 0.9155
No log 9.9429 348 0.8354 0.4465 0.8354 0.9140
No log 10.0 350 0.8972 0.4954 0.8972 0.9472
No log 10.0571 352 0.9073 0.5029 0.9073 0.9525
No log 10.1143 354 0.8477 0.4465 0.8477 0.9207
No log 10.1714 356 0.8332 0.4620 0.8332 0.9128
No log 10.2286 358 0.8473 0.4620 0.8473 0.9205
No log 10.2857 360 0.8672 0.3820 0.8672 0.9313
No log 10.3429 362 0.8589 0.3820 0.8589 0.9268
No log 10.4 364 0.8428 0.4009 0.8428 0.9181
No log 10.4571 366 0.8238 0.4239 0.8238 0.9076
No log 10.5143 368 0.8178 0.4459 0.8178 0.9043
No log 10.5714 370 0.8533 0.4202 0.8533 0.9237
No log 10.6286 372 0.8191 0.4948 0.8191 0.9050
No log 10.6857 374 0.8032 0.4734 0.8032 0.8962
No log 10.7429 376 0.8173 0.4603 0.8173 0.9041
No log 10.8 378 0.7980 0.5094 0.7980 0.8933
No log 10.8571 380 0.7953 0.5203 0.7953 0.8918
No log 10.9143 382 0.7919 0.4828 0.7919 0.8899
No log 10.9714 384 0.7775 0.5073 0.7775 0.8818
No log 11.0286 386 0.7624 0.5916 0.7624 0.8731
No log 11.0857 388 0.7892 0.4910 0.7892 0.8883
No log 11.1429 390 0.8770 0.4706 0.8770 0.9365
No log 11.2 392 1.0010 0.5094 1.0010 1.0005
No log 11.2571 394 0.9748 0.5124 0.9748 0.9873
No log 11.3143 396 0.8880 0.4501 0.8880 0.9423
No log 11.3714 398 0.7844 0.5358 0.7844 0.8857
No log 11.4286 400 0.7648 0.5072 0.7648 0.8745
No log 11.4857 402 0.7772 0.5621 0.7772 0.8816
No log 11.5429 404 0.8402 0.4764 0.8402 0.9166
No log 11.6 406 0.8318 0.4631 0.8318 0.9120
No log 11.6571 408 0.8201 0.4425 0.8201 0.9056
No log 11.7143 410 0.8319 0.4425 0.8319 0.9121
No log 11.7714 412 0.8479 0.4639 0.8479 0.9208
No log 11.8286 414 0.8299 0.4334 0.8299 0.9110
No log 11.8857 416 0.8111 0.4334 0.8111 0.9006
No log 11.9429 418 0.8184 0.4465 0.8184 0.9046
No log 12.0 420 0.8759 0.4507 0.8759 0.9359
No log 12.0571 422 0.9011 0.4501 0.9011 0.9493
No log 12.1143 424 0.9134 0.4501 0.9134 0.9557
No log 12.1714 426 0.8969 0.4507 0.8969 0.9470
No log 12.2286 428 0.9267 0.4501 0.9267 0.9627
No log 12.2857 430 0.9204 0.4501 0.9204 0.9594
No log 12.3429 432 0.9210 0.4703 0.9210 0.9597
No log 12.4 434 0.9383 0.4703 0.9383 0.9687
No log 12.4571 436 1.0233 0.4986 1.0233 1.0116
No log 12.5143 438 1.0414 0.4986 1.0414 1.0205
No log 12.5714 440 1.0366 0.4986 1.0366 1.0182
No log 12.6286 442 0.9696 0.4583 0.9696 0.9847
No log 12.6857 444 0.9048 0.4295 0.9048 0.9512
No log 12.7429 446 0.8757 0.4620 0.8757 0.9358
No log 12.8 448 0.8828 0.3583 0.8828 0.9396
No log 12.8571 450 0.8909 0.4079 0.8909 0.9439
No log 12.9143 452 0.9219 0.3868 0.9219 0.9602
No log 12.9714 454 0.9567 0.4425 0.9567 0.9781
No log 13.0286 456 0.9701 0.3478 0.9701 0.9849
No log 13.0857 458 0.9612 0.3729 0.9612 0.9804
No log 13.1429 460 0.9695 0.4072 0.9695 0.9846
No log 13.2 462 0.9511 0.4072 0.9511 0.9752
No log 13.2571 464 0.9654 0.4712 0.9654 0.9826
No log 13.3143 466 0.9773 0.4703 0.9773 0.9886
No log 13.3714 468 0.9695 0.4583 0.9695 0.9846
No log 13.4286 470 0.9208 0.4526 0.9208 0.9596
No log 13.4857 472 0.8974 0.4526 0.8974 0.9473
No log 13.5429 474 0.8661 0.4700 0.8661 0.9307
No log 13.6 476 0.8381 0.4646 0.8381 0.9155
No log 13.6571 478 0.8598 0.4983 0.8598 0.9272
No log 13.7143 480 0.9208 0.4576 0.9208 0.9596
No log 13.7714 482 0.9319 0.4576 0.9319 0.9654
No log 13.8286 484 0.9153 0.4576 0.9153 0.9567
No log 13.8857 486 0.9469 0.4576 0.9469 0.9731
No log 13.9429 488 0.9690 0.4576 0.9690 0.9844
No log 14.0 490 0.9520 0.4598 0.9520 0.9757
No log 14.0571 492 0.9480 0.4606 0.9480 0.9737
No log 14.1143 494 0.9615 0.4259 0.9615 0.9806
No log 14.1714 496 0.9984 0.3989 0.9984 0.9992
No log 14.2286 498 1.0263 0.4473 1.0263 1.0131
0.3609 14.2857 500 1.0320 0.4576 1.0320 1.0159
0.3609 14.3429 502 0.9552 0.4712 0.9552 0.9773
0.3609 14.4 504 0.8891 0.4540 0.8891 0.9429
0.3609 14.4571 506 0.8644 0.4270 0.8644 0.9297
0.3609 14.5143 508 0.8768 0.4304 0.8768 0.9364
0.3609 14.5714 510 0.8606 0.4124 0.8606 0.9277
0.3609 14.6286 512 0.8725 0.3885 0.8725 0.9341

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k11_task2_organization

Finetuned
(4023)
this model