ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k19_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8062
  • Qwk: 0.2593
  • Mse: 0.8062
  • Rmse: 0.8979

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0211 2 2.7982 -0.0084 2.7982 1.6728
No log 0.0421 4 1.6062 0.0572 1.6062 1.2674
No log 0.0632 6 1.0298 -0.0550 1.0298 1.0148
No log 0.0842 8 0.9947 -0.1866 0.9947 0.9973
No log 0.1053 10 1.0621 0.0473 1.0621 1.0306
No log 0.1263 12 1.1772 -0.0096 1.1772 1.0850
No log 0.1474 14 1.2296 -0.1153 1.2296 1.1089
No log 0.1684 16 1.4220 -0.0870 1.4220 1.1925
No log 0.1895 18 1.3772 -0.0036 1.3772 1.1735
No log 0.2105 20 1.0971 0.0230 1.0971 1.0474
No log 0.2316 22 1.0140 0.0821 1.0140 1.0070
No log 0.2526 24 1.0700 0.0391 1.0700 1.0344
No log 0.2737 26 1.0161 0.1183 1.0161 1.0080
No log 0.2947 28 0.9053 0.0846 0.9053 0.9515
No log 0.3158 30 0.9998 -0.0257 0.9998 0.9999
No log 0.3368 32 0.9799 -0.0320 0.9799 0.9899
No log 0.3579 34 0.8792 0.1187 0.8792 0.9377
No log 0.3789 36 0.8959 0.0181 0.8959 0.9465
No log 0.4 38 0.8880 0.0652 0.8880 0.9423
No log 0.4211 40 1.0090 0.0893 1.0090 1.0045
No log 0.4421 42 1.2633 0.0221 1.2633 1.1240
No log 0.4632 44 1.3453 -0.0755 1.3453 1.1599
No log 0.4842 46 1.1964 -0.0156 1.1964 1.0938
No log 0.5053 48 1.1075 -0.0619 1.1075 1.0524
No log 0.5263 50 1.0058 0.0 1.0058 1.0029
No log 0.5474 52 0.9873 0.0474 0.9873 0.9936
No log 0.5684 54 0.9667 0.0 0.9667 0.9832
No log 0.5895 56 0.9497 0.0 0.9497 0.9745
No log 0.6105 58 0.9251 0.0428 0.9251 0.9618
No log 0.6316 60 0.9716 0.0428 0.9716 0.9857
No log 0.6526 62 1.0435 0.0949 1.0435 1.0215
No log 0.6737 64 1.0221 0.0846 1.0221 1.0110
No log 0.6947 66 1.0187 0.0181 1.0187 1.0093
No log 0.7158 68 0.9958 0.0509 0.9958 0.9979
No log 0.7368 70 0.9877 0.0509 0.9877 0.9938
No log 0.7579 72 0.9906 0.0993 0.9906 0.9953
No log 0.7789 74 1.0363 0.2032 1.0363 1.0180
No log 0.8 76 1.0195 0.2032 1.0195 1.0097
No log 0.8211 78 0.9723 0.0129 0.9723 0.9860
No log 0.8421 80 0.9397 -0.0127 0.9397 0.9694
No log 0.8632 82 0.9266 -0.0149 0.9266 0.9626
No log 0.8842 84 0.9483 0.0643 0.9483 0.9738
No log 0.9053 86 0.9676 0.0643 0.9676 0.9837
No log 0.9263 88 1.0588 0.0511 1.0588 1.0290
No log 0.9474 90 1.1849 0.0274 1.1849 1.0885
No log 0.9684 92 1.2084 0.0287 1.2084 1.0993
No log 0.9895 94 1.0998 0.0341 1.0998 1.0487
No log 1.0105 96 1.0950 0.0341 1.0950 1.0464
No log 1.0316 98 1.0704 0.0 1.0704 1.0346
No log 1.0526 100 1.0510 -0.0351 1.0510 1.0252
No log 1.0737 102 1.0983 0.0494 1.0983 1.0480
No log 1.0947 104 1.1796 -0.0831 1.1796 1.0861
No log 1.1158 106 1.0806 -0.0186 1.0806 1.0395
No log 1.1368 108 0.9953 -0.0241 0.9953 0.9976
No log 1.1579 110 1.0532 0.1214 1.0532 1.0263
No log 1.1789 112 1.0559 0.1178 1.0559 1.0276
No log 1.2 114 1.0157 0.0691 1.0157 1.0078
No log 1.2211 116 1.0219 0.1092 1.0219 1.0109
No log 1.2421 118 1.0282 0.0662 1.0282 1.0140
No log 1.2632 120 1.0107 0.0418 1.0107 1.0053
No log 1.2842 122 1.0045 0.1525 1.0045 1.0022
No log 1.3053 124 0.9865 0.1251 0.9865 0.9932
No log 1.3263 126 0.9912 0.1940 0.9912 0.9956
No log 1.3474 128 1.0061 0.1856 1.0061 1.0031
No log 1.3684 130 0.9290 0.2012 0.9290 0.9639
No log 1.3895 132 0.8961 0.0361 0.8961 0.9466
No log 1.4105 134 0.8852 0.1093 0.8852 0.9408
No log 1.4316 136 0.9061 0.1141 0.9061 0.9519
No log 1.4526 138 0.9850 0.1884 0.9850 0.9925
No log 1.4737 140 1.0904 0.1356 1.0904 1.0442
No log 1.4947 142 1.0964 0.1287 1.0964 1.0471
No log 1.5158 144 1.0220 0.1109 1.0220 1.0110
No log 1.5368 146 0.9911 0.1557 0.9911 0.9955
No log 1.5579 148 0.9538 0.1396 0.9538 0.9766
No log 1.5789 150 0.9083 0.0864 0.9083 0.9531
No log 1.6 152 0.8711 0.0670 0.8711 0.9333
No log 1.6211 154 0.9211 0.2150 0.9211 0.9597
No log 1.6421 156 0.9081 0.1960 0.9081 0.9529
No log 1.6632 158 0.8120 0.1550 0.8120 0.9011
No log 1.6842 160 0.7839 0.1094 0.7839 0.8854
No log 1.7053 162 0.7876 0.1723 0.7876 0.8875
No log 1.7263 164 0.7892 0.1353 0.7892 0.8884
No log 1.7474 166 0.8011 0.1935 0.8011 0.8951
No log 1.7684 168 0.8217 0.1353 0.8217 0.9065
No log 1.7895 170 0.8563 0.1373 0.8563 0.9254
No log 1.8105 172 0.8827 0.1979 0.8827 0.9395
No log 1.8316 174 0.8738 0.1463 0.8738 0.9348
No log 1.8526 176 0.8382 0.0975 0.8382 0.9155
No log 1.8737 178 0.8805 0.2633 0.8805 0.9384
No log 1.8947 180 0.9883 0.2394 0.9883 0.9941
No log 1.9158 182 1.0059 0.2267 1.0059 1.0029
No log 1.9368 184 1.0308 0.1646 1.0308 1.0153
No log 1.9579 186 0.9675 0.2747 0.9675 0.9836
No log 1.9789 188 1.1839 0.1932 1.1839 1.0881
No log 2.0 190 1.4119 0.0601 1.4119 1.1882
No log 2.0211 192 1.2765 0.1920 1.2765 1.1298
No log 2.0421 194 1.0057 0.1500 1.0057 1.0029
No log 2.0632 196 0.9150 0.2071 0.9150 0.9566
No log 2.0842 198 0.9197 0.1472 0.9197 0.9590
No log 2.1053 200 0.9481 0.0534 0.9481 0.9737
No log 2.1263 202 1.0070 0.0982 1.0070 1.0035
No log 2.1474 204 1.0034 0.1672 1.0034 1.0017
No log 2.1684 206 0.9651 -0.0528 0.9651 0.9824
No log 2.1895 208 0.9981 0.2724 0.9981 0.9990
No log 2.2105 210 1.0169 0.2616 1.0169 1.0084
No log 2.2316 212 0.9870 0.1978 0.9870 0.9935
No log 2.2526 214 0.9758 0.0812 0.9758 0.9878
No log 2.2737 216 0.9729 0.1205 0.9729 0.9864
No log 2.2947 218 0.9779 0.1203 0.9779 0.9889
No log 2.3158 220 1.0034 0.1133 1.0034 1.0017
No log 2.3368 222 1.0546 0.1813 1.0546 1.0269
No log 2.3579 224 1.0573 0.1472 1.0573 1.0282
No log 2.3789 226 0.9946 0.1620 0.9946 0.9973
No log 2.4 228 0.9930 0.1131 0.9930 0.9965
No log 2.4211 230 1.0496 0.2154 1.0496 1.0245
No log 2.4421 232 1.0119 0.1476 1.0119 1.0059
No log 2.4632 234 0.9639 0.2006 0.9639 0.9818
No log 2.4842 236 0.9552 0.1672 0.9552 0.9774
No log 2.5053 238 0.9311 0.2040 0.9311 0.9649
No log 2.5263 240 0.9239 0.1171 0.9239 0.9612
No log 2.5474 242 0.9366 0.1573 0.9366 0.9678
No log 2.5684 244 0.9177 0.0496 0.9177 0.9580
No log 2.5895 246 0.9274 0.1137 0.9274 0.9630
No log 2.6105 248 0.9484 0.1992 0.9484 0.9739
No log 2.6316 250 1.0031 0.1589 1.0031 1.0016
No log 2.6526 252 1.0503 0.1541 1.0503 1.0249
No log 2.6737 254 1.1222 0.1699 1.1222 1.0593
No log 2.6947 256 1.1507 0.1734 1.1507 1.0727
No log 2.7158 258 1.1793 0.1532 1.1793 1.0860
No log 2.7368 260 1.2064 0.1519 1.2064 1.0984
No log 2.7579 262 1.2250 0.1799 1.2250 1.1068
No log 2.7789 264 1.2402 0.1453 1.2402 1.1136
No log 2.8 266 1.3503 0.0520 1.3503 1.1620
No log 2.8211 268 1.3358 0.0714 1.3358 1.1558
No log 2.8421 270 1.2605 0.1344 1.2605 1.1227
No log 2.8632 272 1.1967 0.1808 1.1967 1.0939
No log 2.8842 274 1.1224 0.1310 1.1224 1.0594
No log 2.9053 276 1.0640 0.0428 1.0640 1.0315
No log 2.9263 278 1.0007 0.0453 1.0007 1.0003
No log 2.9474 280 0.9650 0.1992 0.9650 0.9824
No log 2.9684 282 0.9312 0.2109 0.9312 0.9650
No log 2.9895 284 0.9027 0.0573 0.9027 0.9501
No log 3.0105 286 0.8740 -0.0112 0.8740 0.9349
No log 3.0316 288 0.8467 -0.0112 0.8467 0.9201
No log 3.0526 290 0.8500 -0.0112 0.8500 0.9220
No log 3.0737 292 0.8665 0.2212 0.8665 0.9309
No log 3.0947 294 0.9141 0.1066 0.9141 0.9561
No log 3.1158 296 0.9328 0.0453 0.9328 0.9658
No log 3.1368 298 0.9279 -0.0177 0.9279 0.9633
No log 3.1579 300 0.9226 0.0227 0.9226 0.9605
No log 3.1789 302 0.8925 0.1992 0.8925 0.9447
No log 3.2 304 0.9513 0.2467 0.9513 0.9754
No log 3.2211 306 0.9052 0.2566 0.9052 0.9514
No log 3.2421 308 0.8377 0.2129 0.8377 0.9153
No log 3.2632 310 0.8503 0.0269 0.8503 0.9221
No log 3.2842 312 0.8525 0.0269 0.8525 0.9233
No log 3.3053 314 0.8460 0.2349 0.8460 0.9198
No log 3.3263 316 0.8394 0.2606 0.8394 0.9162
No log 3.3474 318 0.8238 0.2747 0.8238 0.9076
No log 3.3684 320 0.8232 0.2152 0.8232 0.9073
No log 3.3895 322 0.8424 0.1246 0.8424 0.9178
No log 3.4105 324 0.8729 0.1672 0.8729 0.9343
No log 3.4316 326 0.8590 0.1424 0.8590 0.9268
No log 3.4526 328 0.8927 0.2335 0.8927 0.9448
No log 3.4737 330 0.9118 0.2183 0.9118 0.9549
No log 3.4947 332 0.9191 0.1760 0.9191 0.9587
No log 3.5158 334 0.9970 0.2042 0.9970 0.9985
No log 3.5368 336 1.0382 0.2267 1.0382 1.0189
No log 3.5579 338 1.0693 0.2377 1.0693 1.0341
No log 3.5789 340 0.9534 0.2651 0.9534 0.9764
No log 3.6 342 0.8355 0.2291 0.8355 0.9140
No log 3.6211 344 0.7880 0.2661 0.7880 0.8877
No log 3.6421 346 0.7809 0.2479 0.7809 0.8837
No log 3.6632 348 0.9208 0.2991 0.9208 0.9596
No log 3.6842 350 0.9728 0.2763 0.9728 0.9863
No log 3.7053 352 0.8963 0.1809 0.8963 0.9468
No log 3.7263 354 0.8292 0.1410 0.8292 0.9106
No log 3.7474 356 0.7868 0.2419 0.7868 0.8870
No log 3.7684 358 0.7942 0.2605 0.7942 0.8912
No log 3.7895 360 0.7952 0.2862 0.7952 0.8917
No log 3.8105 362 0.7913 0.2232 0.7913 0.8895
No log 3.8316 364 0.7713 0.2232 0.7713 0.8782
No log 3.8526 366 0.7663 0.2576 0.7663 0.8754
No log 3.8737 368 0.7502 0.2633 0.7502 0.8662
No log 3.8947 370 0.7539 0.2633 0.7539 0.8683
No log 3.9158 372 0.7256 0.2953 0.7256 0.8518
No log 3.9368 374 0.7251 0.2953 0.7251 0.8515
No log 3.9579 376 0.7269 0.2953 0.7269 0.8526
No log 3.9789 378 0.7120 0.3366 0.7120 0.8438
No log 4.0 380 0.7262 0.3385 0.7262 0.8522
No log 4.0211 382 0.7432 0.3291 0.7432 0.8621
No log 4.0421 384 0.7178 0.3258 0.7178 0.8472
No log 4.0632 386 0.6662 0.3324 0.6662 0.8162
No log 4.0842 388 0.7645 0.3280 0.7645 0.8744
No log 4.1053 390 0.8791 0.3560 0.8791 0.9376
No log 4.1263 392 0.8202 0.2832 0.8202 0.9056
No log 4.1474 394 0.7267 0.2953 0.7267 0.8525
No log 4.1684 396 0.7319 0.4160 0.7319 0.8555
No log 4.1895 398 0.7587 0.4074 0.7587 0.8710
No log 4.2105 400 0.8017 0.3357 0.8017 0.8954
No log 4.2316 402 0.7689 0.3859 0.7689 0.8769
No log 4.2526 404 0.7427 0.3007 0.7427 0.8618
No log 4.2737 406 0.7892 0.2424 0.7892 0.8884
No log 4.2947 408 0.8246 0.2328 0.8246 0.9081
No log 4.3158 410 0.8335 0.2342 0.8335 0.9130
No log 4.3368 412 0.9280 0.2570 0.9280 0.9633
No log 4.3579 414 1.0055 0.1671 1.0055 1.0027
No log 4.3789 416 0.9601 0.1633 0.9601 0.9798
No log 4.4 418 0.8757 0.2715 0.8757 0.9358
No log 4.4211 420 0.8544 0.2745 0.8544 0.9243
No log 4.4421 422 0.8418 0.2634 0.8418 0.9175
No log 4.4632 424 0.8476 0.2511 0.8476 0.9207
No log 4.4842 426 0.8899 0.2252 0.8899 0.9433
No log 4.5053 428 0.9020 0.2826 0.9020 0.9497
No log 4.5263 430 0.8803 0.2824 0.8803 0.9382
No log 4.5474 432 0.8960 0.2693 0.8960 0.9466
No log 4.5684 434 0.8799 0.2693 0.8799 0.9380
No log 4.5895 436 0.8859 0.3321 0.8859 0.9412
No log 4.6105 438 0.8830 0.2769 0.8830 0.9397
No log 4.6316 440 0.8212 0.2867 0.8212 0.9062
No log 4.6526 442 0.8176 0.1961 0.8176 0.9042
No log 4.6737 444 0.8389 0.2521 0.8389 0.9159
No log 4.6947 446 0.7927 0.2009 0.7927 0.8903
No log 4.7158 448 0.7849 0.1393 0.7849 0.8859
No log 4.7368 450 0.8364 0.0771 0.8364 0.9146
No log 4.7579 452 0.8450 0.0334 0.8450 0.9192
No log 4.7789 454 0.8310 0.1331 0.8310 0.9116
No log 4.8 456 0.8155 0.1986 0.8155 0.9031
No log 4.8211 458 0.7966 0.1558 0.7966 0.8925
No log 4.8421 460 0.7945 0.1558 0.7945 0.8913
No log 4.8632 462 0.8330 0.2318 0.8330 0.9127
No log 4.8842 464 0.9281 0.1873 0.9281 0.9634
No log 4.9053 466 0.9455 0.1891 0.9455 0.9724
No log 4.9263 468 0.8748 0.1906 0.8748 0.9353
No log 4.9474 470 0.7802 0.1649 0.7802 0.8833
No log 4.9684 472 0.7618 0.2475 0.7618 0.8728
No log 4.9895 474 0.7803 0.2475 0.7803 0.8833
No log 5.0105 476 0.8126 0.2505 0.8126 0.9014
No log 5.0316 478 0.8283 0.2505 0.8283 0.9101
No log 5.0526 480 0.8395 0.2023 0.8395 0.9162
No log 5.0737 482 0.8403 0.2023 0.8403 0.9167
No log 5.0947 484 0.8320 0.2023 0.8320 0.9122
No log 5.1158 486 0.8116 0.1988 0.8116 0.9009
No log 5.1368 488 0.8086 0.1900 0.8086 0.8992
No log 5.1579 490 0.7901 0.1856 0.7901 0.8889
No log 5.1789 492 0.7635 0.1589 0.7635 0.8738
No log 5.2 494 0.7507 0.2606 0.7507 0.8664
No log 5.2211 496 0.7569 0.3622 0.7569 0.8700
No log 5.2421 498 0.7514 0.2914 0.7514 0.8668
0.3842 5.2632 500 0.7579 0.2633 0.7579 0.8706
0.3842 5.2842 502 0.7746 0.3554 0.7746 0.8801
0.3842 5.3053 504 0.7809 0.3311 0.7809 0.8837
0.3842 5.3263 506 0.7894 0.2342 0.7894 0.8885
0.3842 5.3474 508 0.7896 0.2633 0.7896 0.8886
0.3842 5.3684 510 0.7903 0.1672 0.7903 0.8890
0.3842 5.3895 512 0.7976 0.2545 0.7976 0.8931
0.3842 5.4105 514 0.8062 0.2593 0.8062 0.8979

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k19_task7_organization

Finetuned
(4019)
this model