ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k14_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0894
  • Qwk: 0.1094
  • Mse: 1.0894
  • Rmse: 1.0438

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0465 2 4.6189 0.0010 4.6189 2.1492
No log 0.0930 4 2.5452 0.0239 2.5452 1.5954
No log 0.1395 6 1.7772 0.0372 1.7772 1.3331
No log 0.1860 8 1.5012 0.0438 1.5012 1.2252
No log 0.2326 10 1.2198 0.1863 1.2198 1.1044
No log 0.2791 12 1.1438 0.2290 1.1438 1.0695
No log 0.3256 14 1.1677 0.1706 1.1677 1.0806
No log 0.3721 16 1.2376 0.0349 1.2376 1.1125
No log 0.4186 18 1.4512 -0.0284 1.4512 1.2047
No log 0.4651 20 1.5961 0.0084 1.5961 1.2634
No log 0.5116 22 1.5077 0.0403 1.5077 1.2279
No log 0.5581 24 1.2904 0.1339 1.2904 1.1359
No log 0.6047 26 1.0956 0.4760 1.0956 1.0467
No log 0.6512 28 1.0773 0.1335 1.0773 1.0379
No log 0.6977 30 1.0841 0.2351 1.0841 1.0412
No log 0.7442 32 1.2314 0.0454 1.2314 1.1097
No log 0.7907 34 1.2359 0.0872 1.2359 1.1117
No log 0.8372 36 1.2358 0.1507 1.2358 1.1117
No log 0.8837 38 1.2095 0.0599 1.2095 1.0998
No log 0.9302 40 1.2087 0.1042 1.2087 1.0994
No log 0.9767 42 1.2614 0.0353 1.2614 1.1231
No log 1.0233 44 1.3070 0.0454 1.3070 1.1432
No log 1.0698 46 1.3457 0.0454 1.3457 1.1601
No log 1.1163 48 1.2916 0.0771 1.2916 1.1365
No log 1.1628 50 1.3011 -0.0011 1.3011 1.1407
No log 1.2093 52 1.3129 -0.0011 1.3129 1.1458
No log 1.2558 54 1.3320 -0.0085 1.3320 1.1541
No log 1.3023 56 1.3397 0.0032 1.3397 1.1575
No log 1.3488 58 1.3683 0.0671 1.3683 1.1697
No log 1.3953 60 1.3652 0.0222 1.3652 1.1684
No log 1.4419 62 1.3882 -0.0042 1.3882 1.1782
No log 1.4884 64 1.4099 -0.0207 1.4099 1.1874
No log 1.5349 66 1.3987 0.0365 1.3987 1.1827
No log 1.5814 68 1.3900 0.1043 1.3900 1.1790
No log 1.6279 70 1.3764 0.0648 1.3764 1.1732
No log 1.6744 72 1.3919 0.0648 1.3919 1.1798
No log 1.7209 74 1.3831 0.0951 1.3831 1.1761
No log 1.7674 76 1.3997 -0.0688 1.3997 1.1831
No log 1.8140 78 1.4707 0.0389 1.4707 1.2127
No log 1.8605 80 1.4359 0.0622 1.4359 1.1983
No log 1.9070 82 1.4176 -0.0534 1.4176 1.1906
No log 1.9535 84 1.4254 -0.0046 1.4254 1.1939
No log 2.0 86 1.4437 0.1282 1.4437 1.2015
No log 2.0465 88 1.4208 0.1549 1.4208 1.1920
No log 2.0930 90 1.2939 0.1556 1.2939 1.1375
No log 2.1395 92 1.2305 0.1943 1.2305 1.1093
No log 2.1860 94 1.1414 0.2380 1.1414 1.0684
No log 2.2326 96 1.1080 0.1722 1.1080 1.0526
No log 2.2791 98 1.0953 0.0741 1.0953 1.0466
No log 2.3256 100 1.0381 0.3345 1.0381 1.0189
No log 2.3721 102 1.0015 0.3890 1.0015 1.0007
No log 2.4186 104 0.9938 0.3117 0.9938 0.9969
No log 2.4651 106 1.0118 0.3318 1.0118 1.0059
No log 2.5116 108 1.0381 0.3578 1.0381 1.0189
No log 2.5581 110 1.0420 0.4023 1.0420 1.0208
No log 2.6047 112 1.0241 0.4434 1.0241 1.0120
No log 2.6512 114 0.9561 0.4498 0.9561 0.9778
No log 2.6977 116 0.9607 0.4418 0.9607 0.9801
No log 2.7442 118 1.0266 0.2522 1.0266 1.0132
No log 2.7907 120 1.1281 0.1950 1.1281 1.0621
No log 2.8372 122 1.3136 0.1416 1.3136 1.1461
No log 2.8837 124 1.4192 0.1515 1.4192 1.1913
No log 2.9302 126 1.4909 0.2043 1.4909 1.2210
No log 2.9767 128 1.5016 0.2240 1.5016 1.2254
No log 3.0233 130 1.4449 0.1597 1.4449 1.2020
No log 3.0698 132 1.3865 0.1168 1.3865 1.1775
No log 3.1163 134 1.3433 0.1168 1.3433 1.1590
No log 3.1628 136 1.1885 0.1761 1.1885 1.0902
No log 3.2093 138 1.0634 0.2154 1.0634 1.0312
No log 3.2558 140 0.9951 0.3508 0.9951 0.9976
No log 3.3023 142 0.9992 0.3448 0.9992 0.9996
No log 3.3488 144 1.0642 0.2260 1.0642 1.0316
No log 3.3953 146 1.1850 0.1312 1.1850 1.0886
No log 3.4419 148 1.2916 0.0830 1.2916 1.1365
No log 3.4884 150 1.3520 0.1112 1.3520 1.1628
No log 3.5349 152 1.3759 0.2092 1.3759 1.1730
No log 3.5814 154 1.3172 0.3289 1.3172 1.1477
No log 3.6279 156 1.1715 0.2840 1.1715 1.0824
No log 3.6744 158 1.0354 0.3421 1.0354 1.0176
No log 3.7209 160 1.0137 0.3328 1.0137 1.0068
No log 3.7674 162 1.0563 0.3937 1.0563 1.0278
No log 3.8140 164 1.0143 0.3714 1.0143 1.0071
No log 3.8605 166 1.0034 0.2621 1.0034 1.0017
No log 3.9070 168 1.1291 0.2028 1.1291 1.0626
No log 3.9535 170 1.3018 0.2343 1.3018 1.1410
No log 4.0 172 1.3817 0.3082 1.3817 1.1755
No log 4.0465 174 1.4237 0.3082 1.4237 1.1932
No log 4.0930 176 1.3934 0.2436 1.3934 1.1804
No log 4.1395 178 1.3477 0.3289 1.3477 1.1609
No log 4.1860 180 1.2507 0.3570 1.2507 1.1183
No log 4.2326 182 1.1158 0.3294 1.1158 1.0563
No log 4.2791 184 1.0198 0.4340 1.0198 1.0099
No log 4.3256 186 1.1771 0.3907 1.1771 1.0850
No log 4.3721 188 1.5923 0.3291 1.5923 1.2619
No log 4.4186 190 1.6947 0.2711 1.6947 1.3018
No log 4.4651 192 1.6308 0.2775 1.6308 1.2770
No log 4.5116 194 1.4537 0.2968 1.4537 1.2057
No log 4.5581 196 1.2486 0.4041 1.2486 1.1174
No log 4.6047 198 0.9505 0.6032 0.9505 0.9749
No log 4.6512 200 0.8728 0.3945 0.8728 0.9342
No log 4.6977 202 0.8754 0.4440 0.8754 0.9356
No log 4.7442 204 0.8901 0.3819 0.8901 0.9435
No log 4.7907 206 0.9216 0.4272 0.9216 0.9600
No log 4.8372 208 1.0353 0.4130 1.0353 1.0175
No log 4.8837 210 1.0710 0.2154 1.0710 1.0349
No log 4.9302 212 1.0025 0.3045 1.0025 1.0012
No log 4.9767 214 0.9692 0.3478 0.9692 0.9845
No log 5.0233 216 0.9283 0.4181 0.9283 0.9635
No log 5.0698 218 0.8780 0.4555 0.8780 0.9370
No log 5.1163 220 0.8648 0.4930 0.8648 0.9300
No log 5.1628 222 0.8497 0.5171 0.8497 0.9218
No log 5.2093 224 0.8817 0.6085 0.8817 0.9390
No log 5.2558 226 0.8663 0.6193 0.8663 0.9307
No log 5.3023 228 0.8639 0.4409 0.8639 0.9295
No log 5.3488 230 0.9183 0.3478 0.9183 0.9583
No log 5.3953 232 0.9572 0.3191 0.9572 0.9784
No log 5.4419 234 0.9821 0.3389 0.9821 0.9910
No log 5.4884 236 0.9799 0.3389 0.9799 0.9899
No log 5.5349 238 0.9975 0.3489 0.9975 0.9988
No log 5.5814 240 1.0284 0.3489 1.0284 1.0141
No log 5.6279 242 0.9835 0.3122 0.9835 0.9917
No log 5.6744 244 0.9494 0.3270 0.9494 0.9743
No log 5.7209 246 0.9394 0.3590 0.9394 0.9692
No log 5.7674 248 0.9381 0.4722 0.9381 0.9686
No log 5.8140 250 0.9652 0.4200 0.9652 0.9824
No log 5.8605 252 0.9969 0.3930 0.9969 0.9984
No log 5.9070 254 0.9708 0.4345 0.9708 0.9853
No log 5.9535 256 0.9402 0.4774 0.9402 0.9696
No log 6.0 258 0.9219 0.5029 0.9219 0.9601
No log 6.0465 260 0.9178 0.4280 0.9178 0.9580
No log 6.0930 262 0.9386 0.3674 0.9386 0.9688
No log 6.1395 264 0.9656 0.3535 0.9656 0.9826
No log 6.1860 266 1.0071 0.3762 1.0071 1.0036
No log 6.2326 268 1.0239 0.3394 1.0239 1.0119
No log 6.2791 270 1.0250 0.3451 1.0250 1.0124
No log 6.3256 272 1.0052 0.5058 1.0052 1.0026
No log 6.3721 274 0.9926 0.5253 0.9926 0.9963
No log 6.4186 276 0.9549 0.5012 0.9549 0.9772
No log 6.4651 278 0.9581 0.4553 0.9581 0.9788
No log 6.5116 280 0.9463 0.3621 0.9463 0.9728
No log 6.5581 282 0.9393 0.3621 0.9393 0.9692
No log 6.6047 284 0.9354 0.3762 0.9354 0.9672
No log 6.6512 286 0.9400 0.3762 0.9400 0.9695
No log 6.6977 288 0.9491 0.3237 0.9491 0.9742
No log 6.7442 290 0.9964 0.2009 0.9964 0.9982
No log 6.7907 292 1.0697 0.2953 1.0697 1.0343
No log 6.8372 294 1.1178 0.2367 1.1178 1.0572
No log 6.8837 296 1.1428 0.2721 1.1428 1.0690
No log 6.9302 298 1.1005 0.2772 1.1005 1.0490
No log 6.9767 300 1.0437 0.3317 1.0437 1.0216
No log 7.0233 302 1.0835 0.2090 1.0835 1.0409
No log 7.0698 304 1.1382 0.3129 1.1382 1.0669
No log 7.1163 306 1.1319 0.3355 1.1319 1.0639
No log 7.1628 308 1.0957 0.4099 1.0957 1.0467
No log 7.2093 310 1.0198 0.3908 1.0198 1.0099
No log 7.2558 312 1.0013 0.3046 1.0013 1.0007
No log 7.3023 314 0.9821 0.2260 0.9821 0.9910
No log 7.3488 316 0.9976 0.2357 0.9976 0.9988
No log 7.3953 318 1.0495 0.2475 1.0495 1.0245
No log 7.4419 320 1.1485 0.2503 1.1485 1.0717
No log 7.4884 322 1.2308 0.2617 1.2308 1.1094
No log 7.5349 324 1.3184 0.1713 1.3184 1.1482
No log 7.5814 326 1.4761 0.2348 1.4761 1.2150
No log 7.6279 328 1.4921 0.2348 1.4921 1.2215
No log 7.6744 330 1.4148 0.2084 1.4148 1.1894
No log 7.7209 332 1.2913 0.1708 1.2913 1.1363
No log 7.7674 334 1.1730 0.1896 1.1730 1.0831
No log 7.8140 336 1.0485 0.1502 1.0485 1.0240
No log 7.8605 338 0.9742 0.2947 0.9742 0.9870
No log 7.9070 340 0.9498 0.3478 0.9498 0.9746
No log 7.9535 342 0.9436 0.3868 0.9436 0.9714
No log 8.0 344 0.9342 0.4859 0.9342 0.9666
No log 8.0465 346 0.9312 0.4893 0.9312 0.9650
No log 8.0930 348 0.9359 0.4859 0.9359 0.9674
No log 8.1395 350 0.9605 0.3747 0.9605 0.9800
No log 8.1860 352 0.9994 0.4295 0.9994 0.9997
No log 8.2326 354 0.9979 0.4031 0.9979 0.9990
No log 8.2791 356 1.0070 0.3891 1.0070 1.0035
No log 8.3256 358 1.0178 0.3891 1.0178 1.0088
No log 8.3721 360 1.0667 0.3162 1.0667 1.0328
No log 8.4186 362 1.1195 0.2958 1.1195 1.0581
No log 8.4651 364 1.0862 0.3162 1.0862 1.0422
No log 8.5116 366 1.0274 0.3560 1.0274 1.0136
No log 8.5581 368 0.9849 0.3505 0.9849 0.9924
No log 8.6047 370 0.9919 0.3505 0.9919 0.9960
No log 8.6512 372 1.0089 0.2871 1.0089 1.0044
No log 8.6977 374 1.0459 0.3645 1.0459 1.0227
No log 8.7442 376 1.0771 0.2046 1.0771 1.0378
No log 8.7907 378 1.1241 0.1889 1.1241 1.0602
No log 8.8372 380 1.2013 0.2065 1.2013 1.0960
No log 8.8837 382 1.2299 0.2325 1.2299 1.1090
No log 8.9302 384 1.2217 0.2325 1.2217 1.1053
No log 8.9767 386 1.2348 0.1886 1.2348 1.1112
No log 9.0233 388 1.1606 0.2074 1.1606 1.0773
No log 9.0698 390 1.0513 0.3885 1.0513 1.0253
No log 9.1163 392 0.9700 0.2921 0.9700 0.9849
No log 9.1628 394 0.9533 0.3744 0.9533 0.9764
No log 9.2093 396 0.9412 0.3663 0.9412 0.9702
No log 9.2558 398 0.9276 0.3868 0.9276 0.9631
No log 9.3023 400 0.9179 0.4568 0.9179 0.9580
No log 9.3488 402 0.9065 0.4006 0.9065 0.9521
No log 9.3953 404 0.9005 0.4006 0.9005 0.9489
No log 9.4419 406 0.9011 0.4048 0.9011 0.9492
No log 9.4884 408 0.9119 0.4006 0.9119 0.9550
No log 9.5349 410 0.9070 0.3908 0.9070 0.9524
No log 9.5814 412 0.9050 0.4142 0.9050 0.9513
No log 9.6279 414 0.9258 0.3868 0.9258 0.9622
No log 9.6744 416 0.9610 0.3687 0.9610 0.9803
No log 9.7209 418 1.0598 0.3902 1.0598 1.0295
No log 9.7674 420 1.1612 0.3440 1.1612 1.0776
No log 9.8140 422 1.2598 0.2633 1.2598 1.1224
No log 9.8605 424 1.2942 0.2658 1.2942 1.1376
No log 9.9070 426 1.3358 0.2633 1.3358 1.1558
No log 9.9535 428 1.3100 0.2406 1.3100 1.1446
No log 10.0 430 1.1785 0.2589 1.1785 1.0856
No log 10.0465 432 1.0892 0.1753 1.0892 1.0436
No log 10.0930 434 1.0906 0.1753 1.0906 1.0443
No log 10.1395 436 1.1184 0.1753 1.1184 1.0575
No log 10.1860 438 1.1861 0.3155 1.1861 1.0891
No log 10.2326 440 1.2057 0.3174 1.2057 1.0980
No log 10.2791 442 1.2089 0.3174 1.2089 1.0995
No log 10.3256 444 1.2231 0.3174 1.2231 1.1060
No log 10.3721 446 1.1720 0.2032 1.1720 1.0826
No log 10.4186 448 1.1389 0.1611 1.1389 1.0672
No log 10.4651 450 1.1412 0.1989 1.1412 1.0683
No log 10.5116 452 1.1449 0.1943 1.1449 1.0700
No log 10.5581 454 1.1504 0.1554 1.1504 1.0726
No log 10.6047 456 1.1189 0.0607 1.1189 1.0578
No log 10.6512 458 1.1369 0.1026 1.1369 1.0662
No log 10.6977 460 1.1435 0.1596 1.1435 1.0693
No log 10.7442 462 1.1427 0.1703 1.1427 1.0690
No log 10.7907 464 1.1052 0.1989 1.1052 1.0513
No log 10.8372 466 1.0809 0.3699 1.0809 1.0397
No log 10.8837 468 0.9809 0.4129 0.9809 0.9904
No log 10.9302 470 0.8829 0.4637 0.8829 0.9396
No log 10.9767 472 0.8539 0.4583 0.8539 0.9241
No log 11.0233 474 0.8673 0.4794 0.8673 0.9313
No log 11.0698 476 0.8963 0.4941 0.8963 0.9467
No log 11.1163 478 0.9008 0.4941 0.9008 0.9491
No log 11.1628 480 0.8842 0.4926 0.8842 0.9403
No log 11.2093 482 0.8685 0.4847 0.8685 0.9319
No log 11.2558 484 0.8893 0.4311 0.8893 0.9430
No log 11.3023 486 0.9530 0.4615 0.9530 0.9762
No log 11.3488 488 1.0445 0.3772 1.0445 1.0220
No log 11.3953 490 1.0451 0.3962 1.0451 1.0223
No log 11.4419 492 0.9959 0.3119 0.9959 0.9979
No log 11.4884 494 0.9690 0.3325 0.9690 0.9844
No log 11.5349 496 0.9566 0.3354 0.9566 0.9781
No log 11.5814 498 0.9601 0.3354 0.9601 0.9799
0.3771 11.6279 500 0.9764 0.3474 0.9764 0.9881
0.3771 11.6744 502 1.0012 0.3735 1.0012 1.0006
0.3771 11.7209 504 1.0569 0.2525 1.0569 1.0280
0.3771 11.7674 506 1.0614 0.2195 1.0614 1.0302
0.3771 11.8140 508 1.0513 0.2363 1.0513 1.0253
0.3771 11.8605 510 1.0648 0.2363 1.0648 1.0319
0.3771 11.9070 512 1.0894 0.1094 1.0894 1.0438

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k14_task2_organization

Finetuned
(4023)
this model