ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k2_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0237
  • Qwk: 0.5263
  • Mse: 1.0237
  • Rmse: 1.0118

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.2 2 6.9261 0.0114 6.9261 2.6318
No log 0.4 4 4.4570 0.0526 4.4570 2.1112
No log 0.6 6 2.9124 0.0629 2.9124 1.7066
No log 0.8 8 2.3103 0.0690 2.3103 1.5200
No log 1.0 10 2.0207 0.1575 2.0207 1.4215
No log 1.2 12 1.8326 0.2124 1.8326 1.3537
No log 1.4 14 1.7653 0.2000 1.7653 1.3287
No log 1.6 16 1.7733 0.2124 1.7733 1.3317
No log 1.8 18 1.9041 0.2542 1.9041 1.3799
No log 2.0 20 2.2318 0.1481 2.2318 1.4939
No log 2.2 22 2.2764 0.2044 2.2764 1.5088
No log 2.4 24 2.1714 0.1940 2.1714 1.4736
No log 2.6 26 2.1959 0.1143 2.1959 1.4818
No log 2.8 28 2.0699 0.1985 2.0699 1.4387
No log 3.0 30 1.7995 0.2381 1.7995 1.3415
No log 3.2 32 1.6708 0.3360 1.6708 1.2926
No log 3.4 34 1.7448 0.3359 1.7448 1.3209
No log 3.6 36 1.7161 0.3692 1.7161 1.3100
No log 3.8 38 1.6070 0.3226 1.6070 1.2677
No log 4.0 40 1.7378 0.256 1.7378 1.3182
No log 4.2 42 1.7304 0.2439 1.7304 1.3154
No log 4.4 44 1.5903 0.35 1.5903 1.2611
No log 4.6 46 1.5227 0.384 1.5227 1.2340
No log 4.8 48 1.6189 0.3731 1.6189 1.2724
No log 5.0 50 1.5396 0.3846 1.5396 1.2408
No log 5.2 52 1.5450 0.3710 1.5450 1.2430
No log 5.4 54 1.6106 0.2969 1.6106 1.2691
No log 5.6 56 1.4100 0.4677 1.4100 1.1874
No log 5.8 58 1.3484 0.3780 1.3484 1.1612
No log 6.0 60 1.5100 0.3664 1.5100 1.2288
No log 6.2 62 1.3219 0.4 1.3219 1.1498
No log 6.4 64 1.2054 0.5079 1.2054 1.0979
No log 6.6 66 1.4448 0.3607 1.4448 1.2020
No log 6.8 68 1.3252 0.4603 1.3252 1.1512
No log 7.0 70 1.2303 0.5469 1.2303 1.1092
No log 7.2 72 1.2526 0.5 1.2526 1.1192
No log 7.4 74 1.3586 0.4478 1.3586 1.1656
No log 7.6 76 1.2950 0.4885 1.2950 1.1380
No log 7.8 78 1.3122 0.5197 1.3122 1.1455
No log 8.0 80 1.4384 0.4962 1.4384 1.1993
No log 8.2 82 1.5036 0.4511 1.5036 1.2262
No log 8.4 84 1.3563 0.4928 1.3563 1.1646
No log 8.6 86 1.1975 0.5441 1.1975 1.0943
No log 8.8 88 1.1901 0.5401 1.1901 1.0909
No log 9.0 90 1.1675 0.5606 1.1675 1.0805
No log 9.2 92 1.1829 0.5496 1.1829 1.0876
No log 9.4 94 1.2179 0.5303 1.2179 1.1036
No log 9.6 96 1.3693 0.4706 1.3693 1.1702
No log 9.8 98 1.2987 0.4892 1.2987 1.1396
No log 10.0 100 1.1475 0.6043 1.1475 1.0712
No log 10.2 102 1.1061 0.5578 1.1061 1.0517
No log 10.4 104 1.1468 0.5769 1.1468 1.0709
No log 10.6 106 1.0836 0.6069 1.0836 1.0410
No log 10.8 108 1.0776 0.6522 1.0776 1.0381
No log 11.0 110 1.1058 0.5882 1.1058 1.0516
No log 11.2 112 1.1637 0.4878 1.1637 1.0788
No log 11.4 114 1.1607 0.5 1.1607 1.0774
No log 11.6 116 1.1496 0.5839 1.1496 1.0722
No log 11.8 118 1.3626 0.4828 1.3626 1.1673
No log 12.0 120 1.3249 0.5170 1.3249 1.1510
No log 12.2 122 1.0931 0.5616 1.0931 1.0455
No log 12.4 124 1.0495 0.6043 1.0495 1.0245
No log 12.6 126 1.1005 0.5899 1.1005 1.0491
No log 12.8 128 1.1188 0.5630 1.1188 1.0578
No log 13.0 130 1.1495 0.5735 1.1495 1.0721
No log 13.2 132 1.2745 0.4714 1.2745 1.1289
No log 13.4 134 1.2692 0.5 1.2692 1.1266
No log 13.6 136 1.1968 0.5 1.1968 1.0940
No log 13.8 138 1.1140 0.5414 1.1140 1.0555
No log 14.0 140 1.0260 0.5714 1.0260 1.0129
No log 14.2 142 0.9762 0.6260 0.9762 0.9880
No log 14.4 144 0.9603 0.6522 0.9603 0.9800
No log 14.6 146 0.9752 0.6176 0.9752 0.9875
No log 14.8 148 1.0526 0.5970 1.0526 1.0260
No log 15.0 150 1.0716 0.5507 1.0716 1.0352
No log 15.2 152 0.9910 0.5899 0.9910 0.9955
No log 15.4 154 0.9438 0.6277 0.9438 0.9715
No log 15.6 156 0.9471 0.6522 0.9471 0.9732
No log 15.8 158 0.9844 0.6429 0.9844 0.9922
No log 16.0 160 1.0299 0.5942 1.0299 1.0148
No log 16.2 162 1.0782 0.5821 1.0782 1.0383
No log 16.4 164 1.1136 0.5344 1.1136 1.0553
No log 16.6 166 1.0424 0.6015 1.0424 1.0210
No log 16.8 168 0.9737 0.6269 0.9737 0.9868
No log 17.0 170 0.9693 0.6370 0.9693 0.9845
No log 17.2 172 0.9641 0.6522 0.9641 0.9819
No log 17.4 174 1.0480 0.5899 1.0480 1.0237
No log 17.6 176 1.1168 0.5 1.1168 1.0568
No log 17.8 178 1.1074 0.5606 1.1074 1.0523
No log 18.0 180 1.0479 0.5865 1.0479 1.0237
No log 18.2 182 1.0024 0.6277 1.0024 1.0012
No log 18.4 184 1.0240 0.6029 1.0240 1.0119
No log 18.6 186 1.0530 0.5985 1.0530 1.0261
No log 18.8 188 1.0961 0.5630 1.0961 1.0469
No log 19.0 190 1.2356 0.4741 1.2356 1.1116
No log 19.2 192 1.2863 0.4615 1.2863 1.1342
No log 19.4 194 1.2621 0.4733 1.2621 1.1234
No log 19.6 196 1.1766 0.4812 1.1766 1.0847
No log 19.8 198 1.1034 0.5652 1.1034 1.0504
No log 20.0 200 1.0771 0.5816 1.0771 1.0379
No log 20.2 202 1.0643 0.5816 1.0643 1.0316
No log 20.4 204 1.0604 0.5816 1.0604 1.0297
No log 20.6 206 1.0524 0.5816 1.0524 1.0259
No log 20.8 208 1.0388 0.5816 1.0388 1.0192
No log 21.0 210 1.0910 0.5714 1.0910 1.0445
No log 21.2 212 1.2180 0.4964 1.2180 1.1036
No log 21.4 214 1.2305 0.4651 1.2305 1.1093
No log 21.6 216 1.1301 0.5344 1.1301 1.0631
No log 21.8 218 1.0945 0.5606 1.0945 1.0462
No log 22.0 220 1.1184 0.5455 1.1184 1.0576
No log 22.2 222 1.2195 0.5217 1.2195 1.1043
No log 22.4 224 1.3058 0.4091 1.3058 1.1427
No log 22.6 226 1.4502 0.4088 1.4502 1.2042
No log 22.8 228 1.4159 0.4179 1.4159 1.1899
No log 23.0 230 1.2765 0.4308 1.2765 1.1298
No log 23.2 232 1.1131 0.5865 1.1131 1.0550
No log 23.4 234 1.1188 0.5455 1.1188 1.0577
No log 23.6 236 1.1310 0.5672 1.1310 1.0635
No log 23.8 238 1.1904 0.5 1.1904 1.0910
No log 24.0 240 1.2160 0.5526 1.2160 1.1027
No log 24.2 242 1.1713 0.5621 1.1713 1.0823
No log 24.4 244 1.0369 0.5517 1.0369 1.0183
No log 24.6 246 0.9958 0.6119 0.9958 0.9979
No log 24.8 248 0.9876 0.6165 0.9876 0.9938
No log 25.0 250 1.0400 0.5865 1.0400 1.0198
No log 25.2 252 1.1752 0.4806 1.1752 1.0840
No log 25.4 254 1.3807 0.4545 1.3807 1.1750
No log 25.6 256 1.4893 0.4276 1.4893 1.2204
No log 25.8 258 1.3911 0.4583 1.3911 1.1795
No log 26.0 260 1.1995 0.5077 1.1995 1.0952
No log 26.2 262 1.0313 0.6119 1.0313 1.0155
No log 26.4 264 0.9857 0.6370 0.9857 0.9928
No log 26.6 266 1.0335 0.5960 1.0335 1.0166
No log 26.8 268 1.1071 0.5270 1.1071 1.0522
No log 27.0 270 1.1477 0.5068 1.1477 1.0713
No log 27.2 272 1.1311 0.5373 1.1311 1.0635
No log 27.4 274 1.1924 0.5224 1.1924 1.0920
No log 27.6 276 1.2788 0.4478 1.2788 1.1308
No log 27.8 278 1.3251 0.4361 1.3251 1.1511
No log 28.0 280 1.1863 0.5152 1.1863 1.0892
No log 28.2 282 1.0272 0.5778 1.0272 1.0135
No log 28.4 284 0.9555 0.6522 0.9555 0.9775
No log 28.6 286 0.9911 0.6479 0.9911 0.9955
No log 28.8 288 1.0908 0.5401 1.0908 1.0444
No log 29.0 290 1.1621 0.5147 1.1621 1.0780
No log 29.2 292 1.2594 0.4889 1.2594 1.1222
No log 29.4 294 1.2575 0.4769 1.2575 1.1214
No log 29.6 296 1.1656 0.4806 1.1656 1.0796
No log 29.8 298 1.0847 0.5970 1.0847 1.0415
No log 30.0 300 1.0997 0.5758 1.0997 1.0487
No log 30.2 302 1.1620 0.5649 1.1620 1.0780
No log 30.4 304 1.2119 0.4806 1.2119 1.1009
No log 30.6 306 1.1884 0.5077 1.1884 1.0901
No log 30.8 308 1.1477 0.5455 1.1477 1.0713
No log 31.0 310 1.0694 0.5778 1.0694 1.0341
No log 31.2 312 1.0252 0.6029 1.0252 1.0125
No log 31.4 314 0.9870 0.6370 0.9870 0.9935
No log 31.6 316 0.9528 0.6715 0.9528 0.9761
No log 31.8 318 0.9400 0.6715 0.9400 0.9695
No log 32.0 320 0.9744 0.6232 0.9744 0.9871
No log 32.2 322 1.0402 0.5571 1.0402 1.0199
No log 32.4 324 1.0845 0.5616 1.0845 1.0414
No log 32.6 326 1.0510 0.5455 1.0510 1.0252
No log 32.8 328 0.9891 0.6154 0.9891 0.9945
No log 33.0 330 0.9735 0.6143 0.9735 0.9867
No log 33.2 332 0.9953 0.6277 0.9953 0.9977
No log 33.4 334 1.0294 0.6015 1.0294 1.0146
No log 33.6 336 1.1006 0.5271 1.1006 1.0491
No log 33.8 338 1.2506 0.4444 1.2506 1.1183
No log 34.0 340 1.4836 0.4361 1.4836 1.2180
No log 34.2 342 1.4308 0.4444 1.4308 1.1962
No log 34.4 344 1.2396 0.4651 1.2396 1.1134
No log 34.6 346 1.0687 0.5312 1.0687 1.0338
No log 34.8 348 1.0143 0.6212 1.0143 1.0071
No log 35.0 350 1.0111 0.6212 1.0111 1.0055
No log 35.2 352 1.0243 0.5954 1.0243 1.0121
No log 35.4 354 1.0386 0.5985 1.0386 1.0191
No log 35.6 356 1.0435 0.5255 1.0435 1.0215
No log 35.8 358 1.0847 0.5455 1.0847 1.0415
No log 36.0 360 1.0505 0.5655 1.0505 1.0249
No log 36.2 362 0.9803 0.6087 0.9803 0.9901
No log 36.4 364 0.9439 0.5942 0.9439 0.9715
No log 36.6 366 0.9358 0.6074 0.9358 0.9674
No log 36.8 368 0.9540 0.6043 0.9540 0.9767
No log 37.0 370 1.0379 0.5493 1.0379 1.0188
No log 37.2 372 1.1102 0.5867 1.1102 1.0537
No log 37.4 374 1.1237 0.6013 1.1237 1.0601
No log 37.6 376 1.0813 0.5867 1.0813 1.0399
No log 37.8 378 0.9973 0.5816 0.9973 0.9986
No log 38.0 380 0.9386 0.5839 0.9386 0.9688
No log 38.2 382 0.9436 0.6269 0.9436 0.9714
No log 38.4 384 0.9770 0.6222 0.9770 0.9884
No log 38.6 386 1.0015 0.6269 1.0015 1.0008
No log 38.8 388 0.9970 0.6324 0.9970 0.9985
No log 39.0 390 1.0370 0.5714 1.0370 1.0183
No log 39.2 392 1.1428 0.4806 1.1428 1.0690
No log 39.4 394 1.3313 0.5038 1.3313 1.1538
No log 39.6 396 1.4467 0.4361 1.4467 1.2028
No log 39.8 398 1.4115 0.4769 1.4115 1.1881
No log 40.0 400 1.3087 0.4923 1.3087 1.1440
No log 40.2 402 1.1201 0.4844 1.1201 1.0583
No log 40.4 404 1.0188 0.5758 1.0188 1.0094
No log 40.6 406 0.9732 0.5985 0.9732 0.9865
No log 40.8 408 0.9729 0.5839 0.9729 0.9864
No log 41.0 410 1.0278 0.5211 1.0278 1.0138
No log 41.2 412 1.0646 0.5695 1.0646 1.0318
No log 41.4 414 1.0282 0.5655 1.0282 1.0140
No log 41.6 416 0.9938 0.5333 0.9938 0.9969
No log 41.8 418 0.9858 0.5588 0.9858 0.9929
No log 42.0 420 1.0037 0.5185 1.0037 1.0018
No log 42.2 422 1.0137 0.5441 1.0137 1.0068
No log 42.4 424 1.0542 0.5075 1.0542 1.0268
No log 42.6 426 1.1448 0.5113 1.1448 1.0700
No log 42.8 428 1.1686 0.4962 1.1686 1.0810
No log 43.0 430 1.1633 0.5362 1.1633 1.0786
No log 43.2 432 1.0927 0.5147 1.0927 1.0453
No log 43.4 434 1.0244 0.5224 1.0244 1.0121
No log 43.6 436 1.0011 0.5547 1.0011 1.0005
No log 43.8 438 0.9867 0.5547 0.9867 0.9933
No log 44.0 440 1.0078 0.5802 1.0078 1.0039
No log 44.2 442 1.0378 0.5496 1.0378 1.0187
No log 44.4 444 1.0496 0.5116 1.0496 1.0245
No log 44.6 446 1.0525 0.5156 1.0525 1.0259
No log 44.8 448 1.0622 0.5039 1.0622 1.0306
No log 45.0 450 1.0576 0.5 1.0576 1.0284
No log 45.2 452 1.0213 0.5231 1.0213 1.0106
No log 45.4 454 0.9636 0.6119 0.9636 0.9816
No log 45.6 456 0.9482 0.6667 0.9482 0.9738
No log 45.8 458 0.9509 0.6667 0.9509 0.9751
No log 46.0 460 0.9636 0.6716 0.9636 0.9816
No log 46.2 462 0.9883 0.6212 0.9883 0.9941
No log 46.4 464 0.9876 0.6074 0.9876 0.9938
No log 46.6 466 1.0156 0.5839 1.0156 1.0078
No log 46.8 468 1.0152 0.5333 1.0152 1.0076
No log 47.0 470 1.0085 0.5441 1.0085 1.0043
No log 47.2 472 0.9895 0.5839 0.9895 0.9947
No log 47.4 474 0.9620 0.5942 0.9620 0.9808
No log 47.6 476 0.9288 0.6286 0.9288 0.9637
No log 47.8 478 0.9179 0.6906 0.9179 0.9581
No log 48.0 480 0.9166 0.6906 0.9166 0.9574
No log 48.2 482 0.9224 0.6761 0.9224 0.9604
No log 48.4 484 0.9449 0.6525 0.9449 0.9720
No log 48.6 486 0.9952 0.5882 0.9952 0.9976
No log 48.8 488 1.0772 0.5753 1.0772 1.0379
No log 49.0 490 1.1182 0.5229 1.1182 1.0574
No log 49.2 492 1.1119 0.5342 1.1119 1.0545
No log 49.4 494 1.0398 0.5526 1.0398 1.0197
No log 49.6 496 0.9698 0.6056 0.9698 0.9848
No log 49.8 498 0.9163 0.6525 0.9163 0.9572
0.2391 50.0 500 0.9038 0.6525 0.9038 0.9507
0.2391 50.2 502 0.9123 0.6429 0.9123 0.9552
0.2391 50.4 504 0.9487 0.6475 0.9487 0.9740
0.2391 50.6 506 0.9960 0.6131 0.9960 0.9980
0.2391 50.8 508 1.0175 0.5630 1.0175 1.0087
0.2391 51.0 510 1.0237 0.5263 1.0237 1.0118

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k2_task1_organization

Finetuned
(4019)
this model