ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k1_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.1758
  • Qwk: 0.4837
  • Mse: 1.1758
  • Rmse: 1.0843

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.25 2 5.3705 -0.0034 5.3705 2.3174
No log 0.5 4 3.2916 0.0629 3.2916 1.8143
No log 0.75 6 1.9511 0.0228 1.9511 1.3968
No log 1.0 8 1.4549 0.1190 1.4549 1.2062
No log 1.25 10 1.2247 0.2841 1.2247 1.1066
No log 1.5 12 1.2206 0.3148 1.2206 1.1048
No log 1.75 14 1.1559 0.2272 1.1559 1.0751
No log 2.0 16 1.1702 0.2802 1.1702 1.0817
No log 2.25 18 1.2006 0.3229 1.2006 1.0957
No log 2.5 20 1.2484 0.4146 1.2484 1.1173
No log 2.75 22 1.3627 0.3625 1.3627 1.1673
No log 3.0 24 1.4026 0.3478 1.4026 1.1843
No log 3.25 26 1.1352 0.4064 1.1352 1.0654
No log 3.5 28 1.4863 0.0403 1.4863 1.2192
No log 3.75 30 1.8654 -0.1767 1.8654 1.3658
No log 4.0 32 1.8413 -0.1390 1.8413 1.3570
No log 4.25 34 1.4557 -0.0167 1.4557 1.2065
No log 4.5 36 1.3132 0.0185 1.3132 1.1459
No log 4.75 38 1.1922 0.1452 1.1922 1.0919
No log 5.0 40 1.3175 0.2318 1.3175 1.1478
No log 5.25 42 1.6629 0.1299 1.6629 1.2895
No log 5.5 44 1.3039 0.1700 1.3039 1.1419
No log 5.75 46 1.0444 0.2908 1.0444 1.0220
No log 6.0 48 1.1462 0.1283 1.1462 1.0706
No log 6.25 50 1.2610 0.0465 1.2610 1.1229
No log 6.5 52 1.2384 0.0465 1.2384 1.1128
No log 6.75 54 1.1618 0.0738 1.1618 1.0779
No log 7.0 56 1.1069 0.1539 1.1069 1.0521
No log 7.25 58 1.0472 0.2378 1.0472 1.0233
No log 7.5 60 0.9789 0.3747 0.9789 0.9894
No log 7.75 62 0.9910 0.4023 0.9910 0.9955
No log 8.0 64 1.1736 0.3756 1.1736 1.0833
No log 8.25 66 1.2165 0.3903 1.2165 1.1030
No log 8.5 68 1.0621 0.4537 1.0621 1.0306
No log 8.75 70 0.9742 0.4203 0.9742 0.9870
No log 9.0 72 0.9818 0.3913 0.9818 0.9908
No log 9.25 74 0.9297 0.4810 0.9297 0.9642
No log 9.5 76 0.9519 0.5493 0.9519 0.9756
No log 9.75 78 0.9908 0.5263 0.9908 0.9954
No log 10.0 80 0.9574 0.5442 0.9574 0.9785
No log 10.25 82 1.0278 0.5366 1.0278 1.0138
No log 10.5 84 0.9656 0.5480 0.9656 0.9826
No log 10.75 86 1.0046 0.5366 1.0046 1.0023
No log 11.0 88 1.1815 0.5103 1.1815 1.0870
No log 11.25 90 1.4300 0.4147 1.4300 1.1958
No log 11.5 92 1.2649 0.4319 1.2649 1.1247
No log 11.75 94 0.9381 0.5607 0.9381 0.9685
No log 12.0 96 0.8481 0.5129 0.8481 0.9209
No log 12.25 98 0.8269 0.5575 0.8269 0.9093
No log 12.5 100 0.9697 0.5674 0.9697 0.9847
No log 12.75 102 1.3356 0.4328 1.3356 1.1557
No log 13.0 104 1.5820 0.3410 1.5820 1.2578
No log 13.25 106 1.5341 0.3775 1.5341 1.2386
No log 13.5 108 1.1639 0.4664 1.1639 1.0788
No log 13.75 110 0.8399 0.5843 0.8399 0.9165
No log 14.0 112 0.8283 0.5711 0.8283 0.9101
No log 14.25 114 0.8812 0.5496 0.8812 0.9387
No log 14.5 116 1.0013 0.5654 1.0013 1.0006
No log 14.75 118 1.1976 0.4885 1.1976 1.0944
No log 15.0 120 1.2150 0.5025 1.2150 1.1023
No log 15.25 122 1.0165 0.5720 1.0165 1.0082
No log 15.5 124 0.9352 0.5896 0.9352 0.9671
No log 15.75 126 0.9438 0.6029 0.9438 0.9715
No log 16.0 128 0.9711 0.5852 0.9711 0.9854
No log 16.25 130 0.9393 0.5976 0.9393 0.9692
No log 16.5 132 0.8775 0.5717 0.8775 0.9367
No log 16.75 134 1.0160 0.5811 1.0160 1.0080
No log 17.0 136 1.1530 0.4903 1.1530 1.0738
No log 17.25 138 1.1434 0.5289 1.1434 1.0693
No log 17.5 140 0.9848 0.5452 0.9848 0.9924
No log 17.75 142 0.8251 0.5702 0.8251 0.9083
No log 18.0 144 0.7688 0.6094 0.7688 0.8768
No log 18.25 146 0.8215 0.5800 0.8215 0.9064
No log 18.5 148 1.0365 0.5377 1.0365 1.0181
No log 18.75 150 1.1682 0.5017 1.1682 1.0809
No log 19.0 152 1.0815 0.5281 1.0815 1.0399
No log 19.25 154 1.1351 0.5118 1.1351 1.0654
No log 19.5 156 1.0397 0.5538 1.0397 1.0196
No log 19.75 158 0.9184 0.5947 0.9184 0.9583
No log 20.0 160 0.9734 0.5798 0.9734 0.9866
No log 20.25 162 1.0499 0.5237 1.0499 1.0246
No log 20.5 164 1.2011 0.4572 1.2011 1.0959
No log 20.75 166 1.2286 0.4560 1.2286 1.1084
No log 21.0 168 1.0780 0.5117 1.0780 1.0383
No log 21.25 170 0.8870 0.5615 0.8870 0.9418
No log 21.5 172 0.7906 0.5802 0.7906 0.8891
No log 21.75 174 0.8271 0.5656 0.8271 0.9094
No log 22.0 176 1.0281 0.5766 1.0281 1.0140
No log 22.25 178 1.3259 0.4883 1.3259 1.1515
No log 22.5 180 1.2562 0.5341 1.2562 1.1208
No log 22.75 182 1.0721 0.5183 1.0721 1.0354
No log 23.0 184 0.9352 0.5360 0.9352 0.9671
No log 23.25 186 0.8992 0.5541 0.8992 0.9482
No log 23.5 188 1.0157 0.5814 1.0157 1.0078
No log 23.75 190 1.0187 0.5543 1.0187 1.0093
No log 24.0 192 0.9881 0.5255 0.9881 0.9940
No log 24.25 194 0.9953 0.5173 0.9953 0.9977
No log 24.5 196 1.0614 0.5185 1.0614 1.0302
No log 24.75 198 1.0586 0.5307 1.0586 1.0289
No log 25.0 200 1.0036 0.5360 1.0036 1.0018
No log 25.25 202 1.0706 0.5357 1.0706 1.0347
No log 25.5 204 1.1717 0.5045 1.1717 1.0825
No log 25.75 206 1.1065 0.5088 1.1065 1.0519
No log 26.0 208 0.9453 0.5719 0.9453 0.9723
No log 26.25 210 0.8487 0.6147 0.8487 0.9212
No log 26.5 212 0.8747 0.6173 0.8747 0.9352
No log 26.75 214 0.9697 0.5852 0.9697 0.9848
No log 27.0 216 1.1385 0.4980 1.1385 1.0670
No log 27.25 218 1.1991 0.4603 1.1991 1.0950
No log 27.5 220 1.1088 0.5069 1.1088 1.0530
No log 27.75 222 0.9270 0.5676 0.9270 0.9628
No log 28.0 224 0.8946 0.5969 0.8946 0.9458
No log 28.25 226 0.9992 0.5475 0.9992 0.9996
No log 28.5 228 1.1376 0.5317 1.1376 1.0666
No log 28.75 230 1.0821 0.5281 1.0821 1.0402
No log 29.0 232 0.9613 0.5720 0.9613 0.9805
No log 29.25 234 0.8166 0.5910 0.8166 0.9037
No log 29.5 236 0.8283 0.5991 0.8283 0.9101
No log 29.75 238 0.9634 0.5668 0.9634 0.9815
No log 30.0 240 1.0987 0.5586 1.0987 1.0482
No log 30.25 242 1.0938 0.5542 1.0938 1.0458
No log 30.5 244 0.9618 0.6329 0.9618 0.9807
No log 30.75 246 0.8859 0.5851 0.8859 0.9412
No log 31.0 248 0.8881 0.5915 0.8881 0.9424
No log 31.25 250 0.8562 0.6288 0.8562 0.9253
No log 31.5 252 0.9531 0.5879 0.9531 0.9763
No log 31.75 254 0.9800 0.5555 0.9800 0.9900
No log 32.0 256 0.9824 0.5263 0.9824 0.9912
No log 32.25 258 0.9863 0.5176 0.9863 0.9931
No log 32.5 260 1.0305 0.5111 1.0305 1.0151
No log 32.75 262 1.0212 0.5085 1.0212 1.0106
No log 33.0 264 1.0223 0.5216 1.0223 1.0111
No log 33.25 266 0.9758 0.5508 0.9758 0.9878
No log 33.5 268 0.9339 0.5988 0.9339 0.9664
No log 33.75 270 0.8271 0.6321 0.8271 0.9095
No log 34.0 272 0.7760 0.6292 0.7760 0.8809
No log 34.25 274 0.7832 0.6474 0.7832 0.8850
No log 34.5 276 0.8443 0.6292 0.8443 0.9188
No log 34.75 278 0.8492 0.6228 0.8492 0.9215
No log 35.0 280 0.9248 0.5940 0.9248 0.9616
No log 35.25 282 1.0609 0.5276 1.0609 1.0300
No log 35.5 284 1.0645 0.5385 1.0645 1.0317
No log 35.75 286 1.0359 0.5565 1.0359 1.0178
No log 36.0 288 0.9931 0.5536 0.9931 0.9965
No log 36.25 290 1.0873 0.5024 1.0873 1.0427
No log 36.5 292 1.2168 0.4871 1.2168 1.1031
No log 36.75 294 1.1593 0.5002 1.1593 1.0767
No log 37.0 296 1.0488 0.5475 1.0488 1.0241
No log 37.25 298 0.8983 0.5974 0.8983 0.9478
No log 37.5 300 0.8892 0.5974 0.8892 0.9430
No log 37.75 302 0.9518 0.5912 0.9518 0.9756
No log 38.0 304 0.9965 0.6048 0.9965 0.9982
No log 38.25 306 1.0051 0.5744 1.0051 1.0025
No log 38.5 308 0.9599 0.5889 0.9599 0.9797
No log 38.75 310 0.9620 0.6031 0.9620 0.9808
No log 39.0 312 1.0411 0.5499 1.0411 1.0203
No log 39.25 314 1.1273 0.4979 1.1273 1.0618
No log 39.5 316 1.0450 0.5369 1.0450 1.0222
No log 39.75 318 0.9312 0.5801 0.9312 0.9650
No log 40.0 320 0.8756 0.5902 0.8756 0.9357
No log 40.25 322 0.8972 0.5801 0.8972 0.9472
No log 40.5 324 1.0106 0.5548 1.0106 1.0053
No log 40.75 326 1.1207 0.5046 1.1207 1.0586
No log 41.0 328 1.0768 0.4965 1.0768 1.0377
No log 41.25 330 0.9781 0.5666 0.9781 0.9890
No log 41.5 332 0.9954 0.5359 0.9954 0.9977
No log 41.75 334 1.0159 0.5289 1.0159 1.0079
No log 42.0 336 1.0781 0.5215 1.0781 1.0383
No log 42.25 338 1.1442 0.5045 1.1442 1.0697
No log 42.5 340 1.0461 0.5336 1.0461 1.0228
No log 42.75 342 0.9940 0.5759 0.9940 0.9970
No log 43.0 344 0.9964 0.5653 0.9964 0.9982
No log 43.25 346 1.0287 0.5536 1.0287 1.0143
No log 43.5 348 1.0645 0.5266 1.0645 1.0318
No log 43.75 350 1.1618 0.4959 1.1618 1.0778
No log 44.0 352 1.1967 0.4861 1.1967 1.0939
No log 44.25 354 1.0751 0.5200 1.0751 1.0369
No log 44.5 356 1.0169 0.5449 1.0169 1.0084
No log 44.75 358 0.9947 0.5950 0.9947 0.9973
No log 45.0 360 1.0771 0.4915 1.0771 1.0379
No log 45.25 362 1.2200 0.4414 1.2200 1.1046
No log 45.5 364 1.2735 0.4360 1.2735 1.1285
No log 45.75 366 1.2058 0.4545 1.2058 1.0981
No log 46.0 368 1.1127 0.4940 1.1127 1.0548
No log 46.25 370 1.0711 0.5184 1.0711 1.0349
No log 46.5 372 0.9522 0.5745 0.9522 0.9758
No log 46.75 374 0.8825 0.5942 0.8825 0.9394
No log 47.0 376 0.8975 0.5731 0.8975 0.9473
No log 47.25 378 0.9276 0.5690 0.9276 0.9631
No log 47.5 380 1.0332 0.5285 1.0332 1.0165
No log 47.75 382 1.1185 0.4853 1.1185 1.0576
No log 48.0 384 1.1057 0.4853 1.1057 1.0515
No log 48.25 386 1.0395 0.5619 1.0395 1.0196
No log 48.5 388 0.9192 0.5787 0.9192 0.9588
No log 48.75 390 0.8627 0.6054 0.8627 0.9288
No log 49.0 392 0.8697 0.6054 0.8697 0.9326
No log 49.25 394 0.9504 0.5867 0.9504 0.9749
No log 49.5 396 1.0485 0.5338 1.0485 1.0239
No log 49.75 398 1.1529 0.4681 1.1529 1.0737
No log 50.0 400 1.1297 0.4642 1.1297 1.0629
No log 50.25 402 1.0790 0.5184 1.0790 1.0387
No log 50.5 404 0.9757 0.5692 0.9757 0.9878
No log 50.75 406 0.9238 0.5745 0.9238 0.9612
No log 51.0 408 0.9102 0.5840 0.9102 0.9540
No log 51.25 410 0.9438 0.5745 0.9438 0.9715
No log 51.5 412 0.9844 0.5903 0.9844 0.9922
No log 51.75 414 0.9785 0.5758 0.9785 0.9892
No log 52.0 416 0.9967 0.5693 0.9967 0.9984
No log 52.25 418 0.9537 0.5903 0.9537 0.9766
No log 52.5 420 0.9749 0.5693 0.9749 0.9874
No log 52.75 422 0.9677 0.5851 0.9677 0.9837
No log 53.0 424 0.9423 0.5838 0.9423 0.9707
No log 53.25 426 0.9500 0.5551 0.9500 0.9747
No log 53.5 428 0.9969 0.5508 0.9969 0.9985
No log 53.75 430 1.0124 0.5356 1.0124 1.0062
No log 54.0 432 1.0223 0.5159 1.0223 1.0111
No log 54.25 434 0.9867 0.5512 0.9867 0.9933
No log 54.5 436 0.9255 0.5595 0.9255 0.9620
No log 54.75 438 0.9288 0.5636 0.9288 0.9638
No log 55.0 440 1.0086 0.5578 1.0086 1.0043
No log 55.25 442 1.0645 0.5392 1.0645 1.0317
No log 55.5 444 1.0525 0.5402 1.0525 1.0259
No log 55.75 446 1.0189 0.5384 1.0189 1.0094
No log 56.0 448 0.9675 0.5514 0.9675 0.9836
No log 56.25 450 0.8980 0.5759 0.8980 0.9476
No log 56.5 452 0.8613 0.6135 0.8613 0.9281
No log 56.75 454 0.8441 0.6147 0.8441 0.9187
No log 57.0 456 0.8893 0.5963 0.8893 0.9430
No log 57.25 458 0.9789 0.5718 0.9789 0.9894
No log 57.5 460 1.0877 0.5539 1.0877 1.0429
No log 57.75 462 1.0964 0.5539 1.0964 1.0471
No log 58.0 464 1.0405 0.5758 1.0405 1.0201
No log 58.25 466 1.0174 0.5932 1.0174 1.0087
No log 58.5 468 1.0123 0.5879 1.0123 1.0061
No log 58.75 470 0.9999 0.5609 0.9999 0.9999
No log 59.0 472 0.9540 0.5745 0.9540 0.9768
No log 59.25 474 0.9201 0.5481 0.9201 0.9592
No log 59.5 476 0.9075 0.5422 0.9075 0.9526
No log 59.75 478 0.9301 0.5336 0.9301 0.9644
No log 60.0 480 0.9937 0.5445 0.9937 0.9969
No log 60.25 482 1.0360 0.5348 1.0360 1.0178
No log 60.5 484 1.0669 0.5338 1.0669 1.0329
No log 60.75 486 1.0658 0.5420 1.0658 1.0324
No log 61.0 488 1.0148 0.5548 1.0148 1.0074
No log 61.25 490 0.9355 0.6163 0.9355 0.9672
No log 61.5 492 0.8565 0.5700 0.8565 0.9255
No log 61.75 494 0.8448 0.5969 0.8448 0.9191
No log 62.0 496 0.8695 0.5916 0.8695 0.9325
No log 62.25 498 0.9138 0.6011 0.9138 0.9560
0.3348 62.5 500 1.0067 0.5988 1.0067 1.0033
0.3348 62.75 502 1.1522 0.5107 1.1522 1.0734
0.3348 63.0 504 1.2567 0.4816 1.2567 1.1210
0.3348 63.25 506 1.3304 0.4576 1.3304 1.1534
0.3348 63.5 508 1.2839 0.4414 1.2839 1.1331
0.3348 63.75 510 1.1758 0.4837 1.1758 1.0843

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k1_task1_organization

Finetuned
(4023)
this model