ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run2_AugV5_k5_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8204
  • Qwk: 0.5286
  • Mse: 0.8204
  • Rmse: 0.9057

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0769 2 5.3148 0.0119 5.3148 2.3054
No log 0.1538 4 3.2932 0.0602 3.2932 1.8147
No log 0.2308 6 2.2446 0.0108 2.2446 1.4982
No log 0.3077 8 1.8599 0.0481 1.8599 1.3638
No log 0.3846 10 1.3178 0.1259 1.3178 1.1479
No log 0.4615 12 1.1589 0.2166 1.1589 1.0765
No log 0.5385 14 1.1510 0.2470 1.1510 1.0729
No log 0.6154 16 1.1385 0.1928 1.1385 1.0670
No log 0.6923 18 1.2879 0.1721 1.2879 1.1348
No log 0.7692 20 1.4453 0.0746 1.4453 1.2022
No log 0.8462 22 1.1494 0.3407 1.1494 1.0721
No log 0.9231 24 1.0541 0.2650 1.0541 1.0267
No log 1.0 26 1.1148 0.1868 1.1148 1.0559
No log 1.0769 28 1.1623 0.2249 1.1623 1.0781
No log 1.1538 30 1.1824 0.3116 1.1824 1.0874
No log 1.2308 32 1.0948 0.3389 1.0948 1.0463
No log 1.3077 34 1.1468 0.3332 1.1468 1.0709
No log 1.3846 36 1.1181 0.3730 1.1181 1.0574
No log 1.4615 38 1.0346 0.3260 1.0346 1.0171
No log 1.5385 40 1.0320 0.2908 1.0320 1.0159
No log 1.6154 42 1.0251 0.3212 1.0251 1.0125
No log 1.6923 44 0.9816 0.4133 0.9816 0.9908
No log 1.7692 46 1.0084 0.4587 1.0084 1.0042
No log 1.8462 48 0.9978 0.4384 0.9978 0.9989
No log 1.9231 50 1.0281 0.4507 1.0281 1.0139
No log 2.0 52 0.9984 0.4534 0.9984 0.9992
No log 2.0769 54 0.9511 0.4369 0.9511 0.9752
No log 2.1538 56 0.9429 0.4279 0.9429 0.9710
No log 2.2308 58 0.9287 0.4397 0.9287 0.9637
No log 2.3077 60 0.9325 0.4012 0.9325 0.9657
No log 2.3846 62 0.9268 0.3985 0.9268 0.9627
No log 2.4615 64 0.9309 0.4682 0.9309 0.9648
No log 2.5385 66 0.8860 0.4219 0.8860 0.9413
No log 2.6154 68 0.8869 0.4360 0.8869 0.9418
No log 2.6923 70 0.8855 0.4866 0.8855 0.9410
No log 2.7692 72 0.9086 0.5130 0.9086 0.9532
No log 2.8462 74 0.9535 0.4604 0.9535 0.9765
No log 2.9231 76 0.9683 0.4937 0.9683 0.9840
No log 3.0 78 0.9823 0.5015 0.9823 0.9911
No log 3.0769 80 0.9551 0.4307 0.9551 0.9773
No log 3.1538 82 0.9761 0.4816 0.9761 0.9880
No log 3.2308 84 1.0172 0.4733 1.0172 1.0086
No log 3.3077 86 1.0031 0.4723 1.0031 1.0016
No log 3.3846 88 1.0333 0.4881 1.0333 1.0165
No log 3.4615 90 0.9469 0.5368 0.9469 0.9731
No log 3.5385 92 0.8839 0.5579 0.8839 0.9402
No log 3.6154 94 0.8907 0.5862 0.8907 0.9438
No log 3.6923 96 0.9274 0.5663 0.9274 0.9630
No log 3.7692 98 1.0297 0.5510 1.0297 1.0147
No log 3.8462 100 1.0337 0.5650 1.0337 1.0167
No log 3.9231 102 0.8941 0.5294 0.8941 0.9456
No log 4.0 104 0.8108 0.5914 0.8108 0.9005
No log 4.0769 106 0.8024 0.5502 0.8024 0.8958
No log 4.1538 108 0.8104 0.5483 0.8104 0.9002
No log 4.2308 110 0.8877 0.5616 0.8877 0.9422
No log 4.3077 112 0.9310 0.5741 0.9310 0.9649
No log 4.3846 114 0.9176 0.5832 0.9176 0.9579
No log 4.4615 116 0.9246 0.6079 0.9246 0.9615
No log 4.5385 118 0.9778 0.6084 0.9778 0.9888
No log 4.6154 120 1.0144 0.5835 1.0144 1.0072
No log 4.6923 122 0.9751 0.5986 0.9751 0.9875
No log 4.7692 124 1.0018 0.5938 1.0018 1.0009
No log 4.8462 126 1.3015 0.4981 1.3015 1.1408
No log 4.9231 128 1.3340 0.4571 1.3340 1.1550
No log 5.0 130 1.2259 0.4968 1.2259 1.1072
No log 5.0769 132 1.0892 0.4904 1.0892 1.0436
No log 5.1538 134 1.0490 0.5519 1.0490 1.0242
No log 5.2308 136 1.0796 0.5160 1.0796 1.0391
No log 5.3077 138 0.9860 0.5517 0.9860 0.9930
No log 5.3846 140 0.9697 0.5752 0.9697 0.9847
No log 5.4615 142 0.8845 0.6096 0.8845 0.9405
No log 5.5385 144 0.8583 0.6292 0.8583 0.9265
No log 5.6154 146 0.9290 0.5825 0.9290 0.9639
No log 5.6923 148 1.0084 0.5666 1.0084 1.0042
No log 5.7692 150 0.9783 0.5801 0.9783 0.9891
No log 5.8462 152 0.9515 0.5623 0.9515 0.9754
No log 5.9231 154 0.9424 0.5752 0.9424 0.9708
No log 6.0 156 0.8557 0.5338 0.8557 0.9250
No log 6.0769 158 0.8668 0.5603 0.8668 0.9310
No log 6.1538 160 1.1379 0.5414 1.1379 1.0667
No log 6.2308 162 1.5209 0.4704 1.5209 1.2332
No log 6.3077 164 1.5108 0.4713 1.5108 1.2291
No log 6.3846 166 1.3026 0.5060 1.3026 1.1413
No log 6.4615 168 1.1531 0.5333 1.1531 1.0738
No log 6.5385 170 1.1089 0.5672 1.1089 1.0530
No log 6.6154 172 1.2302 0.5218 1.2302 1.1091
No log 6.6923 174 1.4789 0.4520 1.4789 1.2161
No log 6.7692 176 1.4854 0.4520 1.4854 1.2188
No log 6.8462 178 1.4377 0.4435 1.4377 1.1990
No log 6.9231 180 1.1266 0.5079 1.1266 1.0614
No log 7.0 182 0.9500 0.5755 0.9500 0.9747
No log 7.0769 184 0.9235 0.6077 0.9235 0.9610
No log 7.1538 186 1.0361 0.5471 1.0361 1.0179
No log 7.2308 188 1.2710 0.4782 1.2710 1.1274
No log 7.3077 190 1.3260 0.4698 1.3260 1.1515
No log 7.3846 192 1.1578 0.5686 1.1578 1.0760
No log 7.4615 194 1.1159 0.5832 1.1159 1.0564
No log 7.5385 196 1.1393 0.5594 1.1393 1.0674
No log 7.6154 198 1.0290 0.6036 1.0290 1.0144
No log 7.6923 200 1.0641 0.5591 1.0641 1.0316
No log 7.7692 202 1.1833 0.5228 1.1833 1.0878
No log 7.8462 204 1.1321 0.5391 1.1321 1.0640
No log 7.9231 206 0.9575 0.5192 0.9575 0.9785
No log 8.0 208 0.8981 0.5520 0.8981 0.9477
No log 8.0769 210 0.9455 0.5514 0.9455 0.9724
No log 8.1538 212 1.0051 0.5187 1.0051 1.0026
No log 8.2308 214 1.0025 0.5667 1.0025 1.0013
No log 8.3077 216 0.9514 0.5517 0.9514 0.9754
No log 8.3846 218 0.8391 0.5844 0.8391 0.9160
No log 8.4615 220 0.8864 0.6322 0.8864 0.9415
No log 8.5385 222 1.1634 0.5503 1.1634 1.0786
No log 8.6154 224 1.4303 0.5083 1.4303 1.1959
No log 8.6923 226 1.3972 0.4866 1.3972 1.1820
No log 8.7692 228 1.2078 0.5415 1.2078 1.0990
No log 8.8462 230 1.0755 0.5907 1.0755 1.0371
No log 8.9231 232 0.9823 0.5864 0.9823 0.9911
No log 9.0 234 0.9491 0.5946 0.9491 0.9742
No log 9.0769 236 0.9899 0.5777 0.9899 0.9949
No log 9.1538 238 1.0692 0.5754 1.0692 1.0340
No log 9.2308 240 1.0300 0.6177 1.0300 1.0149
No log 9.3077 242 1.0370 0.6032 1.0370 1.0183
No log 9.3846 244 1.0746 0.5531 1.0746 1.0366
No log 9.4615 246 1.1948 0.4720 1.1948 1.0931
No log 9.5385 248 1.2234 0.5142 1.2234 1.1061
No log 9.6154 250 1.2767 0.5035 1.2767 1.1299
No log 9.6923 252 1.2916 0.5022 1.2916 1.1365
No log 9.7692 254 1.2577 0.5155 1.2577 1.1215
No log 9.8462 256 1.1329 0.5305 1.1329 1.0644
No log 9.9231 258 1.1249 0.5279 1.1249 1.0606
No log 10.0 260 1.0864 0.4949 1.0864 1.0423
No log 10.0769 262 1.1177 0.5037 1.1177 1.0572
No log 10.1538 264 1.0044 0.5193 1.0044 1.0022
No log 10.2308 266 0.8596 0.5956 0.8596 0.9271
No log 10.3077 268 0.8502 0.5974 0.8502 0.9221
No log 10.3846 270 0.9665 0.5861 0.9665 0.9831
No log 10.4615 272 1.0332 0.5686 1.0332 1.0165
No log 10.5385 274 1.1339 0.5448 1.1339 1.0649
No log 10.6154 276 1.1066 0.5629 1.1066 1.0520
No log 10.6923 278 1.0421 0.5614 1.0421 1.0209
No log 10.7692 280 1.0736 0.5438 1.0736 1.0362
No log 10.8462 282 1.1078 0.5266 1.1078 1.0525
No log 10.9231 284 1.1206 0.5430 1.1206 1.0586
No log 11.0 286 1.1875 0.5369 1.1875 1.0897
No log 11.0769 288 1.1487 0.5326 1.1487 1.0718
No log 11.1538 290 1.0840 0.5177 1.0840 1.0411
No log 11.2308 292 1.0144 0.5400 1.0144 1.0072
No log 11.3077 294 1.0005 0.5422 1.0005 1.0003
No log 11.3846 296 1.1828 0.5664 1.1828 1.0876
No log 11.4615 298 1.2148 0.5623 1.2148 1.1022
No log 11.5385 300 1.2732 0.5666 1.2732 1.1284
No log 11.6154 302 1.2366 0.5333 1.2366 1.1120
No log 11.6923 304 1.3137 0.5264 1.3137 1.1462
No log 11.7692 306 1.2327 0.5482 1.2327 1.1103
No log 11.8462 308 1.0813 0.5913 1.0813 1.0398
No log 11.9231 310 1.0479 0.6086 1.0479 1.0237
No log 12.0 312 1.1015 0.5945 1.1015 1.0495
No log 12.0769 314 1.2767 0.5612 1.2767 1.1299
No log 12.1538 316 1.3220 0.5282 1.3220 1.1498
No log 12.2308 318 1.2076 0.5614 1.2076 1.0989
No log 12.3077 320 1.0801 0.5938 1.0801 1.0393
No log 12.3846 322 1.0710 0.5506 1.0710 1.0349
No log 12.4615 324 1.0874 0.5416 1.0874 1.0428
No log 12.5385 326 1.2211 0.5138 1.2211 1.1050
No log 12.6154 328 1.1880 0.5613 1.1880 1.0899
No log 12.6923 330 1.0064 0.5834 1.0064 1.0032
No log 12.7692 332 0.9152 0.5912 0.9152 0.9566
No log 12.8462 334 0.9842 0.6043 0.9842 0.9921
No log 12.9231 336 1.1544 0.5731 1.1544 1.0744
No log 13.0 338 1.3469 0.5417 1.3469 1.1606
No log 13.0769 340 1.3689 0.5365 1.3689 1.1700
No log 13.1538 342 1.2310 0.5940 1.2310 1.1095
No log 13.2308 344 1.0467 0.6031 1.0467 1.0231
No log 13.3077 346 0.8493 0.6107 0.8493 0.9216
No log 13.3846 348 0.8398 0.6142 0.8398 0.9164
No log 13.4615 350 0.9853 0.6185 0.9853 0.9926
No log 13.5385 352 1.1060 0.5654 1.1060 1.0516
No log 13.6154 354 1.0715 0.5731 1.0715 1.0352
No log 13.6923 356 1.0287 0.5719 1.0287 1.0142
No log 13.7692 358 0.8908 0.5833 0.8908 0.9438
No log 13.8462 360 0.8106 0.5727 0.8106 0.9003
No log 13.9231 362 0.8055 0.4879 0.8055 0.8975
No log 14.0 364 0.8374 0.4879 0.8374 0.9151
No log 14.0769 366 0.9160 0.5152 0.9160 0.9571
No log 14.1538 368 0.9398 0.5223 0.9398 0.9694
No log 14.2308 370 0.9690 0.5309 0.9690 0.9844
No log 14.3077 372 1.0198 0.5582 1.0198 1.0099
No log 14.3846 374 0.9758 0.5766 0.9758 0.9878
No log 14.4615 376 1.0035 0.5879 1.0035 1.0018
No log 14.5385 378 1.0905 0.5822 1.0905 1.0443
No log 14.6154 380 1.0645 0.5879 1.0645 1.0318
No log 14.6923 382 1.0756 0.5659 1.0756 1.0371
No log 14.7692 384 1.1074 0.5361 1.1074 1.0523
No log 14.8462 386 1.1062 0.5161 1.1062 1.0517
No log 14.9231 388 1.1659 0.5492 1.1659 1.0798
No log 15.0 390 1.2507 0.5232 1.2507 1.1184
No log 15.0769 392 1.2498 0.5154 1.2498 1.1179
No log 15.1538 394 1.3067 0.5218 1.3067 1.1431
No log 15.2308 396 1.2145 0.5129 1.2145 1.1021
No log 15.3077 398 1.1506 0.5405 1.1506 1.0727
No log 15.3846 400 1.2186 0.5394 1.2186 1.1039
No log 15.4615 402 1.3172 0.4936 1.3172 1.1477
No log 15.5385 404 1.2573 0.5447 1.2573 1.1213
No log 15.6154 406 1.0514 0.5780 1.0514 1.0254
No log 15.6923 408 0.8581 0.5859 0.8581 0.9263
No log 15.7692 410 0.7658 0.6252 0.7658 0.8751
No log 15.8462 412 0.7589 0.6547 0.7589 0.8711
No log 15.9231 414 0.8086 0.6159 0.8086 0.8992
No log 16.0 416 0.8984 0.5740 0.8984 0.9478
No log 16.0769 418 0.8809 0.5875 0.8809 0.9385
No log 16.1538 420 0.8578 0.5640 0.8578 0.9262
No log 16.2308 422 0.9047 0.5886 0.9047 0.9511
No log 16.3077 424 1.0312 0.5708 1.0312 1.0155
No log 16.3846 426 1.0778 0.5687 1.0778 1.0382
No log 16.4615 428 0.9581 0.5813 0.9581 0.9788
No log 16.5385 430 0.8433 0.5997 0.8433 0.9183
No log 16.6154 432 0.8934 0.5947 0.8934 0.9452
No log 16.6923 434 0.9938 0.5973 0.9938 0.9969
No log 16.7692 436 1.1424 0.6078 1.1424 1.0688
No log 16.8462 438 1.1610 0.6052 1.1610 1.0775
No log 16.9231 440 1.1350 0.5994 1.1350 1.0654
No log 17.0 442 1.0350 0.5887 1.0350 1.0173
No log 17.0769 444 0.9929 0.5767 0.9929 0.9965
No log 17.1538 446 0.9719 0.5767 0.9719 0.9858
No log 17.2308 448 0.9669 0.5641 0.9669 0.9833
No log 17.3077 450 0.8916 0.5369 0.8916 0.9442
No log 17.3846 452 0.8748 0.5309 0.8748 0.9353
No log 17.4615 454 0.9197 0.5396 0.9197 0.9590
No log 17.5385 456 0.9360 0.5730 0.9360 0.9675
No log 17.6154 458 0.9066 0.5863 0.9066 0.9521
No log 17.6923 460 0.9571 0.6224 0.9571 0.9783
No log 17.7692 462 0.9195 0.6513 0.9195 0.9589
No log 17.8462 464 0.8112 0.6492 0.8112 0.9007
No log 17.9231 466 0.8110 0.6495 0.8110 0.9006
No log 18.0 468 0.9569 0.6186 0.9569 0.9782
No log 18.0769 470 1.0097 0.6060 1.0097 1.0048
No log 18.1538 472 0.9909 0.5986 0.9909 0.9954
No log 18.2308 474 0.9041 0.5962 0.9041 0.9508
No log 18.3077 476 0.9193 0.6003 0.9193 0.9588
No log 18.3846 478 0.9779 0.5449 0.9779 0.9889
No log 18.4615 480 1.0177 0.4871 1.0177 1.0088
No log 18.5385 482 0.9362 0.5603 0.9362 0.9676
No log 18.6154 484 0.9542 0.5359 0.9542 0.9768
No log 18.6923 486 1.0163 0.5101 1.0163 1.0081
No log 18.7692 488 1.0293 0.5891 1.0293 1.0146
No log 18.8462 490 1.0851 0.5525 1.0851 1.0417
No log 18.9231 492 1.2342 0.5230 1.2342 1.1110
No log 19.0 494 1.3032 0.5065 1.3032 1.1416
No log 19.0769 496 1.2041 0.5091 1.2041 1.0973
No log 19.1538 498 1.1026 0.5435 1.1026 1.0500
0.3897 19.2308 500 1.0719 0.5426 1.0719 1.0353
0.3897 19.3077 502 1.1502 0.5513 1.1502 1.0725
0.3897 19.3846 504 1.2250 0.5336 1.2250 1.1068
0.3897 19.4615 506 1.1305 0.5216 1.1305 1.0633
0.3897 19.5385 508 0.9312 0.5202 0.9312 0.9650
0.3897 19.6154 510 0.8445 0.4807 0.8445 0.9190
0.3897 19.6923 512 0.8204 0.5286 0.8204 0.9057

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run2_AugV5_k5_task1_organization

Finetuned
(4023)
this model