ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k20_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9493
  • Qwk: 0.3648
  • Mse: 0.9493
  • Rmse: 0.9743

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0274 2 4.4856 0.0010 4.4856 2.1179
No log 0.0548 4 2.6290 0.0179 2.6290 1.6214
No log 0.0822 6 1.6935 0.0372 1.6935 1.3013
No log 0.1096 8 1.8833 0.0041 1.8833 1.3723
No log 0.1370 10 1.4180 0.0421 1.4180 1.1908
No log 0.1644 12 1.3442 0.0203 1.3442 1.1594
No log 0.1918 14 1.3492 0.1021 1.3492 1.1616
No log 0.2192 16 1.3935 0.0101 1.3935 1.1805
No log 0.2466 18 1.4201 0.0104 1.4201 1.1917
No log 0.2740 20 1.3218 0.2108 1.3218 1.1497
No log 0.3014 22 1.5031 0.1956 1.5031 1.2260
No log 0.3288 24 2.0352 0.1414 2.0352 1.4266
No log 0.3562 26 2.6391 0.0699 2.6391 1.6245
No log 0.3836 28 2.0525 0.1086 2.0525 1.4327
No log 0.4110 30 1.1851 0.2088 1.1851 1.0886
No log 0.4384 32 1.0903 0.2333 1.0903 1.0442
No log 0.4658 34 1.1352 0.2289 1.1352 1.0655
No log 0.4932 36 1.4032 0.0512 1.4032 1.1846
No log 0.5205 38 2.0461 0.1814 2.0461 1.4304
No log 0.5479 40 2.6614 0.0759 2.6614 1.6314
No log 0.5753 42 2.8593 0.0902 2.8593 1.6910
No log 0.6027 44 2.4515 0.1324 2.4515 1.5657
No log 0.6301 46 1.8742 0.1612 1.8742 1.3690
No log 0.6575 48 1.5443 0.2048 1.5443 1.2427
No log 0.6849 50 1.4143 0.2840 1.4143 1.1893
No log 0.7123 52 1.5926 0.2458 1.5926 1.2620
No log 0.7397 54 1.8511 0.1892 1.8511 1.3606
No log 0.7671 56 1.7524 0.2228 1.7524 1.3238
No log 0.7945 58 1.3297 0.2237 1.3297 1.1531
No log 0.8219 60 1.3105 0.1842 1.3105 1.1448
No log 0.8493 62 1.2096 0.1639 1.2096 1.0998
No log 0.8767 64 1.2864 0.0121 1.2864 1.1342
No log 0.9041 66 1.2525 0.0269 1.2525 1.1192
No log 0.9315 68 1.1498 0.1959 1.1498 1.0723
No log 0.9589 70 1.0852 0.3284 1.0852 1.0417
No log 0.9863 72 1.1204 0.2565 1.1204 1.0585
No log 1.0137 74 1.2787 0.2885 1.2787 1.1308
No log 1.0411 76 1.5029 0.2299 1.5029 1.2259
No log 1.0685 78 1.6798 0.3151 1.6798 1.2961
No log 1.0959 80 1.4956 0.1966 1.4956 1.2230
No log 1.1233 82 1.1365 0.2203 1.1365 1.0661
No log 1.1507 84 0.9346 0.3708 0.9346 0.9668
No log 1.1781 86 0.9294 0.4072 0.9294 0.9640
No log 1.2055 88 0.9955 0.3236 0.9955 0.9977
No log 1.2329 90 1.1895 0.2014 1.1895 1.0906
No log 1.2603 92 1.2262 0.2675 1.2262 1.1073
No log 1.2877 94 1.0536 0.3238 1.0536 1.0265
No log 1.3151 96 1.2103 0.3761 1.2103 1.1001
No log 1.3425 98 1.7548 0.2652 1.7547 1.3247
No log 1.3699 100 2.0141 0.2151 2.0141 1.4192
No log 1.3973 102 1.7704 0.2746 1.7704 1.3306
No log 1.4247 104 1.3311 0.3787 1.3311 1.1537
No log 1.4521 106 1.1944 0.4121 1.1944 1.0929
No log 1.4795 108 1.2055 0.3113 1.2055 1.0979
No log 1.5068 110 0.9499 0.2815 0.9499 0.9746
No log 1.5342 112 0.8369 0.5375 0.8369 0.9148
No log 1.5616 114 0.8753 0.4879 0.8753 0.9356
No log 1.5890 116 1.3574 0.4237 1.3574 1.1651
No log 1.6164 118 1.7708 0.2980 1.7708 1.3307
No log 1.6438 120 1.8345 0.2757 1.8345 1.3544
No log 1.6712 122 1.6588 0.3472 1.6588 1.2879
No log 1.6986 124 1.5828 0.2873 1.5828 1.2581
No log 1.7260 126 1.3110 0.3348 1.3110 1.1450
No log 1.7534 128 1.0275 0.3429 1.0275 1.0136
No log 1.7808 130 0.9749 0.4668 0.9749 0.9874
No log 1.8082 132 0.9882 0.4528 0.9882 0.9941
No log 1.8356 134 1.1549 0.4310 1.1549 1.0746
No log 1.8630 136 1.3159 0.4561 1.3159 1.1471
No log 1.8904 138 1.1161 0.5048 1.1161 1.0564
No log 1.9178 140 0.9690 0.3787 0.9690 0.9844
No log 1.9452 142 0.9650 0.4389 0.9650 0.9824
No log 1.9726 144 1.0034 0.3345 1.0034 1.0017
No log 2.0 146 1.0927 0.3714 1.0927 1.0453
No log 2.0274 148 1.1607 0.4258 1.1607 1.0774
No log 2.0548 150 1.0285 0.3720 1.0285 1.0141
No log 2.0822 152 0.9809 0.3538 0.9809 0.9904
No log 2.1096 154 0.9946 0.4209 0.9946 0.9973
No log 2.1370 156 1.1009 0.5451 1.1009 1.0493
No log 2.1644 158 0.9769 0.3888 0.9769 0.9884
No log 2.1918 160 0.9793 0.4297 0.9793 0.9896
No log 2.2192 162 0.9888 0.4211 0.9888 0.9944
No log 2.2466 164 0.9439 0.3940 0.9439 0.9715
No log 2.2740 166 0.9415 0.3699 0.9415 0.9703
No log 2.3014 168 0.9440 0.4032 0.9440 0.9716
No log 2.3288 170 1.0067 0.3728 1.0067 1.0034
No log 2.3562 172 1.2240 0.4137 1.2240 1.1064
No log 2.3836 174 1.4299 0.3164 1.4299 1.1958
No log 2.4110 176 1.2118 0.4867 1.2118 1.1008
No log 2.4384 178 1.0207 0.3501 1.0207 1.0103
No log 2.4658 180 1.0490 0.4783 1.0490 1.0242
No log 2.4932 182 1.0236 0.4559 1.0236 1.0117
No log 2.5205 184 1.0870 0.4110 1.0870 1.0426
No log 2.5479 186 1.1103 0.4140 1.1103 1.0537
No log 2.5753 188 1.0062 0.3076 1.0062 1.0031
No log 2.6027 190 0.9755 0.4128 0.9755 0.9877
No log 2.6301 192 0.9680 0.4421 0.9680 0.9839
No log 2.6575 194 0.9841 0.3684 0.9841 0.9920
No log 2.6849 196 1.1580 0.4561 1.1580 1.0761
No log 2.7123 198 1.1352 0.4265 1.1352 1.0654
No log 2.7397 200 0.9987 0.4323 0.9987 0.9993
No log 2.7671 202 0.9815 0.4536 0.9815 0.9907
No log 2.7945 204 1.0535 0.4096 1.0535 1.0264
No log 2.8219 206 1.2470 0.3752 1.2470 1.1167
No log 2.8493 208 1.1301 0.3959 1.1301 1.0631
No log 2.8767 210 1.0263 0.3720 1.0263 1.0131
No log 2.9041 212 1.1013 0.3805 1.1013 1.0495
No log 2.9315 214 1.2938 0.3630 1.2938 1.1375
No log 2.9589 216 1.2624 0.3745 1.2624 1.1236
No log 2.9863 218 1.0201 0.3993 1.0201 1.0100
No log 3.0137 220 0.9807 0.4331 0.9807 0.9903
No log 3.0411 222 1.0494 0.4003 1.0494 1.0244
No log 3.0685 224 1.0026 0.4712 1.0026 1.0013
No log 3.0959 226 0.9498 0.3811 0.9498 0.9746
No log 3.1233 228 1.0260 0.3606 1.0260 1.0129
No log 3.1507 230 1.0896 0.3652 1.0896 1.0439
No log 3.1781 232 1.0470 0.3484 1.0470 1.0232
No log 3.2055 234 0.9923 0.4362 0.9923 0.9961
No log 3.2329 236 0.9595 0.3888 0.9595 0.9795
No log 3.2603 238 0.9488 0.4847 0.9488 0.9741
No log 3.2877 240 0.9640 0.3922 0.9640 0.9818
No log 3.3151 242 0.9672 0.3636 0.9672 0.9835
No log 3.3425 244 0.9891 0.3554 0.9891 0.9945
No log 3.3699 246 1.1090 0.3777 1.1090 1.0531
No log 3.3973 248 1.1923 0.4121 1.1923 1.0919
No log 3.4247 250 1.2645 0.4067 1.2645 1.1245
No log 3.4521 252 1.1914 0.3660 1.1914 1.0915
No log 3.4795 254 1.1093 0.3682 1.1093 1.0532
No log 3.5068 256 1.0659 0.3277 1.0659 1.0324
No log 3.5342 258 1.1124 0.3450 1.1124 1.0547
No log 3.5616 260 1.1587 0.4154 1.1587 1.0764
No log 3.5890 262 1.1011 0.3967 1.1011 1.0493
No log 3.6164 264 1.1333 0.3500 1.1333 1.0646
No log 3.6438 266 1.1382 0.3500 1.1382 1.0669
No log 3.6712 268 1.1411 0.3791 1.1411 1.0682
No log 3.6986 270 1.1875 0.3621 1.1875 1.0897
No log 3.7260 272 1.2238 0.3913 1.2238 1.1062
No log 3.7534 274 1.2011 0.4172 1.2011 1.0960
No log 3.7808 276 1.2611 0.4293 1.2611 1.1230
No log 3.8082 278 1.2402 0.4265 1.2402 1.1136
No log 3.8356 280 1.5098 0.2801 1.5098 1.2287
No log 3.8630 282 1.7568 0.2019 1.7568 1.3255
No log 3.8904 284 1.7103 0.2403 1.7103 1.3078
No log 3.9178 286 1.2910 0.3148 1.2910 1.1362
No log 3.9452 288 0.9464 0.4102 0.9464 0.9728
No log 3.9726 290 0.9284 0.4146 0.9284 0.9635
No log 4.0 292 0.9303 0.4413 0.9303 0.9645
No log 4.0274 294 1.0181 0.3009 1.0181 1.0090
No log 4.0548 296 1.2025 0.2601 1.2025 1.0966
No log 4.0822 298 1.3266 0.2578 1.3266 1.1518
No log 4.1096 300 1.4374 0.2558 1.4374 1.1989
No log 4.1370 302 1.3039 0.3232 1.3039 1.1419
No log 4.1644 304 1.0981 0.3601 1.0981 1.0479
No log 4.1918 306 0.9982 0.2936 0.9982 0.9991
No log 4.2192 308 1.0217 0.3179 1.0217 1.0108
No log 4.2466 310 1.0615 0.3128 1.0615 1.0303
No log 4.2740 312 1.1048 0.3458 1.1048 1.0511
No log 4.3014 314 1.0320 0.3805 1.0320 1.0159
No log 4.3288 316 0.9481 0.3991 0.9481 0.9737
No log 4.3562 318 0.9500 0.4363 0.9500 0.9747
No log 4.3836 320 0.9355 0.3961 0.9355 0.9672
No log 4.4110 322 0.9275 0.3618 0.9275 0.9631
No log 4.4384 324 0.9353 0.4607 0.9353 0.9671
No log 4.4658 326 0.9549 0.3636 0.9549 0.9772
No log 4.4932 328 0.9995 0.3561 0.9995 0.9998
No log 4.5205 330 1.0449 0.3828 1.0449 1.0222
No log 4.5479 332 1.0063 0.3653 1.0063 1.0032
No log 4.5753 334 0.9585 0.3865 0.9585 0.9790
No log 4.6027 336 0.9524 0.3541 0.9524 0.9759
No log 4.6301 338 0.9863 0.3703 0.9863 0.9931
No log 4.6575 340 1.1592 0.4767 1.1592 1.0767
No log 4.6849 342 1.2739 0.4302 1.2739 1.1287
No log 4.7123 344 1.1346 0.4544 1.1346 1.0652
No log 4.7397 346 1.0011 0.3789 1.0011 1.0005
No log 4.7671 348 0.9600 0.3527 0.9600 0.9798
No log 4.7945 350 0.9532 0.3305 0.9532 0.9763
No log 4.8219 352 1.0100 0.3902 1.0100 1.0050
No log 4.8493 354 1.1097 0.3380 1.1097 1.0534
No log 4.8767 356 1.1410 0.3590 1.1410 1.0682
No log 4.9041 358 1.0609 0.3237 1.0609 1.0300
No log 4.9315 360 0.9788 0.2966 0.9788 0.9893
No log 4.9589 362 0.9682 0.4002 0.9682 0.9840
No log 4.9863 364 0.9624 0.4039 0.9624 0.9810
No log 5.0137 366 0.9720 0.3103 0.9720 0.9859
No log 5.0411 368 1.0515 0.3237 1.0515 1.0254
No log 5.0685 370 1.1020 0.3523 1.1020 1.0498
No log 5.0959 372 1.1046 0.3523 1.1046 1.0510
No log 5.1233 374 1.0677 0.3701 1.0677 1.0333
No log 5.1507 376 0.9925 0.3371 0.9925 0.9962
No log 5.1781 378 0.9634 0.2991 0.9634 0.9815
No log 5.2055 380 0.9685 0.3421 0.9685 0.9841
No log 5.2329 382 0.9738 0.3421 0.9738 0.9868
No log 5.2603 384 1.0364 0.3705 1.0364 1.0180
No log 5.2877 386 1.1811 0.3777 1.1811 1.0868
No log 5.3151 388 1.4438 0.2745 1.4438 1.2016
No log 5.3425 390 1.4872 0.28 1.4872 1.2195
No log 5.3699 392 1.3168 0.2875 1.3168 1.1475
No log 5.3973 394 1.0809 0.3323 1.0809 1.0397
No log 5.4247 396 0.9810 0.3218 0.9810 0.9905
No log 5.4521 398 0.9601 0.3697 0.9601 0.9798
No log 5.4795 400 0.9721 0.3697 0.9721 0.9859
No log 5.5068 402 0.9573 0.3960 0.9573 0.9784
No log 5.5342 404 0.9745 0.4518 0.9745 0.9872
No log 5.5616 406 1.0449 0.4804 1.0449 1.0222
No log 5.5890 408 1.1040 0.3989 1.1040 1.0507
No log 5.6164 410 1.0681 0.4119 1.0681 1.0335
No log 5.6438 412 0.9512 0.4176 0.9512 0.9753
No log 5.6712 414 0.9235 0.3723 0.9235 0.9610
No log 5.6986 416 0.9343 0.4045 0.9343 0.9666
No log 5.7260 418 0.9348 0.3728 0.9348 0.9669
No log 5.7534 420 0.9110 0.3723 0.9110 0.9545
No log 5.7808 422 0.9185 0.3859 0.9185 0.9584
No log 5.8082 424 0.9279 0.3989 0.9279 0.9633
No log 5.8356 426 0.9425 0.4645 0.9425 0.9708
No log 5.8630 428 0.9580 0.4556 0.9580 0.9788
No log 5.8904 430 0.9150 0.4932 0.9150 0.9566
No log 5.9178 432 0.8611 0.4711 0.8611 0.9279
No log 5.9452 434 0.8664 0.4685 0.8664 0.9308
No log 5.9726 436 0.8628 0.4982 0.8628 0.9288
No log 6.0 438 0.9437 0.4685 0.9437 0.9714
No log 6.0274 440 1.2458 0.3485 1.2458 1.1161
No log 6.0548 442 1.4278 0.3521 1.4278 1.1949
No log 6.0822 444 1.3004 0.3485 1.3004 1.1403
No log 6.1096 446 1.0802 0.4567 1.0802 1.0393
No log 6.1370 448 0.9696 0.3877 0.9696 0.9847
No log 6.1644 450 0.9607 0.3796 0.9607 0.9802
No log 6.1918 452 0.9690 0.3490 0.9690 0.9844
No log 6.2192 454 1.0174 0.3532 1.0174 1.0087
No log 6.2466 456 1.0359 0.3624 1.0359 1.0178
No log 6.2740 458 1.0721 0.4050 1.0721 1.0354
No log 6.3014 460 1.0505 0.4137 1.0505 1.0250
No log 6.3288 462 0.9646 0.4003 0.9646 0.9822
No log 6.3562 464 0.9515 0.4825 0.9515 0.9755
No log 6.3836 466 0.9437 0.4449 0.9437 0.9714
No log 6.4110 468 0.9387 0.4898 0.9387 0.9689
No log 6.4384 470 0.9433 0.4694 0.9433 0.9713
No log 6.4658 472 1.0037 0.4224 1.0037 1.0018
No log 6.4932 474 1.1640 0.4036 1.1640 1.0789
No log 6.5205 476 1.2375 0.3667 1.2375 1.1124
No log 6.5479 478 1.1804 0.4036 1.1804 1.0865
No log 6.5753 480 1.0490 0.3557 1.0490 1.0242
No log 6.6027 482 0.9810 0.3467 0.9810 0.9904
No log 6.6301 484 0.9402 0.4407 0.9402 0.9697
No log 6.6575 486 0.9374 0.4906 0.9374 0.9682
No log 6.6849 488 0.9376 0.4610 0.9376 0.9683
No log 6.7123 490 0.9621 0.3493 0.9621 0.9809
No log 6.7397 492 0.9842 0.3768 0.9842 0.9920
No log 6.7671 494 1.0210 0.4172 1.0210 1.0104
No log 6.7945 496 1.0360 0.4172 1.0360 1.0178
No log 6.8219 498 1.0102 0.3761 1.0102 1.0051
0.3762 6.8493 500 0.9807 0.3685 0.9807 0.9903
0.3762 6.8767 502 0.9712 0.4121 0.9712 0.9855
0.3762 6.9041 504 1.0044 0.4156 1.0044 1.0022
0.3762 6.9315 506 1.0925 0.4487 1.0925 1.0452
0.3762 6.9589 508 1.1464 0.3786 1.1464 1.0707
0.3762 6.9863 510 1.1700 0.3585 1.1700 1.0817
0.3762 7.0137 512 1.0276 0.4172 1.0276 1.0137
0.3762 7.0411 514 0.9164 0.4742 0.9164 0.9573
0.3762 7.0685 516 0.9087 0.4949 0.9087 0.9533
0.3762 7.0959 518 0.9128 0.3836 0.9128 0.9554
0.3762 7.1233 520 0.9370 0.3490 0.9370 0.9680
0.3762 7.1507 522 0.9726 0.3798 0.9726 0.9862
0.3762 7.1781 524 0.9918 0.3930 0.9918 0.9959
0.3762 7.2055 526 1.0307 0.3339 1.0307 1.0152
0.3762 7.2329 528 1.0032 0.3521 1.0032 1.0016
0.3762 7.2603 530 0.9527 0.2824 0.9527 0.9761
0.3762 7.2877 532 0.9491 0.3174 0.9491 0.9742
0.3762 7.3151 534 0.9493 0.3648 0.9493 0.9743

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k20_task2_organization

Finetuned
(4019)
this model