ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k3_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0212
  • Qwk: 0.4718
  • Mse: 1.0212
  • Rmse: 1.0106

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.125 2 5.4040 -0.0316 5.4040 2.3247
No log 0.25 4 3.1172 0.0245 3.1172 1.7656
No log 0.375 6 2.0753 0.0252 2.0753 1.4406
No log 0.5 8 1.5545 0.1278 1.5545 1.2468
No log 0.625 10 1.3117 0.2170 1.3117 1.1453
No log 0.75 12 1.1928 0.2232 1.1928 1.0922
No log 0.875 14 1.2822 0.1630 1.2822 1.1323
No log 1.0 16 1.5897 0.0569 1.5897 1.2608
No log 1.125 18 2.4327 -0.0215 2.4327 1.5597
No log 1.25 20 2.4883 0.0673 2.4883 1.5774
No log 1.375 22 1.6973 0.0339 1.6973 1.3028
No log 1.5 24 1.2215 0.2932 1.2215 1.1052
No log 1.625 26 1.3757 0.3162 1.3757 1.1729
No log 1.75 28 1.3324 0.2396 1.3324 1.1543
No log 1.875 30 1.1412 0.0963 1.1412 1.0683
No log 2.0 32 1.1363 0.1822 1.1363 1.0660
No log 2.125 34 1.1419 0.3318 1.1419 1.0686
No log 2.25 36 1.0775 0.3461 1.0775 1.0380
No log 2.375 38 1.0001 0.2961 1.0001 1.0001
No log 2.5 40 1.0043 0.2788 1.0043 1.0021
No log 2.625 42 1.0380 0.2470 1.0380 1.0188
No log 2.75 44 1.0748 0.2509 1.0748 1.0367
No log 2.875 46 1.1402 0.2498 1.1402 1.0678
No log 3.0 48 1.0747 0.2420 1.0747 1.0367
No log 3.125 50 0.9938 0.3635 0.9938 0.9969
No log 3.25 52 0.9969 0.3484 0.9969 0.9985
No log 3.375 54 0.9585 0.4215 0.9585 0.9790
No log 3.5 56 0.9809 0.4032 0.9809 0.9904
No log 3.625 58 1.1322 0.3344 1.1322 1.0640
No log 3.75 60 1.0559 0.4397 1.0559 1.0276
No log 3.875 62 1.0101 0.4343 1.0101 1.0050
No log 4.0 64 1.1545 0.3349 1.1545 1.0745
No log 4.125 66 1.1290 0.3095 1.1290 1.0626
No log 4.25 68 1.0184 0.4756 1.0184 1.0091
No log 4.375 70 1.0812 0.4666 1.0812 1.0398
No log 4.5 72 1.0396 0.4676 1.0396 1.0196
No log 4.625 74 0.9706 0.5066 0.9706 0.9852
No log 4.75 76 0.9139 0.5238 0.9139 0.9560
No log 4.875 78 0.9220 0.5323 0.9220 0.9602
No log 5.0 80 0.8290 0.6292 0.8290 0.9105
No log 5.125 82 0.8340 0.6175 0.8340 0.9132
No log 5.25 84 0.8495 0.5878 0.8495 0.9217
No log 5.375 86 0.8592 0.5162 0.8592 0.9269
No log 5.5 88 0.9179 0.5048 0.9179 0.9581
No log 5.625 90 1.1003 0.4570 1.1003 1.0489
No log 5.75 92 1.1842 0.4588 1.1842 1.0882
No log 5.875 94 1.1354 0.4580 1.1354 1.0656
No log 6.0 96 0.8711 0.5462 0.8711 0.9334
No log 6.125 98 0.7704 0.5923 0.7704 0.8777
No log 6.25 100 0.7389 0.6141 0.7389 0.8596
No log 6.375 102 0.8484 0.5945 0.8484 0.9211
No log 6.5 104 1.2639 0.4843 1.2639 1.1242
No log 6.625 106 1.2367 0.4771 1.2367 1.1121
No log 6.75 108 0.8656 0.6160 0.8656 0.9304
No log 6.875 110 0.8022 0.6407 0.8022 0.8957
No log 7.0 112 0.8206 0.5898 0.8206 0.9059
No log 7.125 114 0.8173 0.6166 0.8173 0.9041
No log 7.25 116 0.9418 0.5082 0.9418 0.9705
No log 7.375 118 0.9405 0.5075 0.9405 0.9698
No log 7.5 120 0.8074 0.5708 0.8074 0.8985
No log 7.625 122 0.9024 0.5617 0.9024 0.9499
No log 7.75 124 0.9325 0.5484 0.9325 0.9656
No log 7.875 126 0.8582 0.5842 0.8582 0.9264
No log 8.0 128 0.9395 0.5551 0.9395 0.9693
No log 8.125 130 0.9194 0.5313 0.9194 0.9588
No log 8.25 132 0.8821 0.5460 0.8821 0.9392
No log 8.375 134 0.9138 0.5335 0.9138 0.9559
No log 8.5 136 0.9082 0.5269 0.9082 0.9530
No log 8.625 138 0.9744 0.5034 0.9744 0.9871
No log 8.75 140 1.0213 0.4559 1.0213 1.0106
No log 8.875 142 0.9274 0.5366 0.9274 0.9630
No log 9.0 144 0.9178 0.5278 0.9178 0.9580
No log 9.125 146 0.9140 0.5463 0.9140 0.9560
No log 9.25 148 1.0295 0.4752 1.0295 1.0146
No log 9.375 150 1.1659 0.4482 1.1659 1.0798
No log 9.5 152 1.0157 0.4946 1.0157 1.0078
No log 9.625 154 0.8584 0.6000 0.8584 0.9265
No log 9.75 156 0.8692 0.5979 0.8692 0.9323
No log 9.875 158 0.9415 0.5189 0.9415 0.9703
No log 10.0 160 0.8824 0.5769 0.8824 0.9394
No log 10.125 162 0.8420 0.5923 0.8420 0.9176
No log 10.25 164 0.8452 0.5923 0.8452 0.9193
No log 10.375 166 0.9015 0.5315 0.9015 0.9495
No log 10.5 168 0.8980 0.5526 0.8980 0.9476
No log 10.625 170 0.8454 0.5966 0.8454 0.9195
No log 10.75 172 0.8448 0.6060 0.8448 0.9191
No log 10.875 174 0.9165 0.5430 0.9165 0.9573
No log 11.0 176 0.8849 0.5534 0.8849 0.9407
No log 11.125 178 0.8250 0.6008 0.8250 0.9083
No log 11.25 180 0.8409 0.5953 0.8409 0.9170
No log 11.375 182 1.0072 0.4974 1.0072 1.0036
No log 11.5 184 1.1069 0.4780 1.1069 1.0521
No log 11.625 186 1.0173 0.5115 1.0173 1.0086
No log 11.75 188 0.8894 0.5598 0.8894 0.9431
No log 11.875 190 0.8698 0.5712 0.8698 0.9326
No log 12.0 192 0.9361 0.5475 0.9361 0.9675
No log 12.125 194 0.9915 0.4905 0.9915 0.9957
No log 12.25 196 0.8888 0.5608 0.8888 0.9428
No log 12.375 198 0.7734 0.6109 0.7734 0.8794
No log 12.5 200 0.7577 0.5947 0.7577 0.8705
No log 12.625 202 0.8071 0.6072 0.8071 0.8984
No log 12.75 204 0.8662 0.5865 0.8662 0.9307
No log 12.875 206 0.8184 0.6016 0.8184 0.9047
No log 13.0 208 0.7835 0.5635 0.7835 0.8851
No log 13.125 210 0.7831 0.5824 0.7831 0.8849
No log 13.25 212 0.7946 0.5916 0.7946 0.8914
No log 13.375 214 0.9032 0.5144 0.9032 0.9503
No log 13.5 216 1.0533 0.5178 1.0533 1.0263
No log 13.625 218 0.9551 0.5212 0.9551 0.9773
No log 13.75 220 0.8154 0.5828 0.8154 0.9030
No log 13.875 222 0.8079 0.5752 0.8079 0.8988
No log 14.0 224 0.8675 0.5987 0.8675 0.9314
No log 14.125 226 0.9016 0.5816 0.9016 0.9495
No log 14.25 228 0.8630 0.5891 0.8630 0.9290
No log 14.375 230 0.8300 0.5922 0.8300 0.9110
No log 14.5 232 0.8293 0.5768 0.8293 0.9106
No log 14.625 234 0.8742 0.5345 0.8742 0.9350
No log 14.75 236 0.9250 0.4976 0.9250 0.9618
No log 14.875 238 0.8729 0.4941 0.8729 0.9343
No log 15.0 240 0.8143 0.5724 0.8143 0.9024
No log 15.125 242 0.8542 0.5395 0.8542 0.9242
No log 15.25 244 0.9439 0.4841 0.9439 0.9715
No log 15.375 246 0.9094 0.5440 0.9094 0.9536
No log 15.5 248 0.9019 0.5080 0.9019 0.9497
No log 15.625 250 0.9896 0.4886 0.9896 0.9948
No log 15.75 252 0.9914 0.4766 0.9914 0.9957
No log 15.875 254 0.8468 0.5470 0.8468 0.9202
No log 16.0 256 0.8080 0.5872 0.8080 0.8989
No log 16.125 258 0.8025 0.6024 0.8025 0.8958
No log 16.25 260 0.7948 0.5681 0.7948 0.8915
No log 16.375 262 0.9183 0.5433 0.9183 0.9583
No log 16.5 264 1.0022 0.4959 1.0022 1.0011
No log 16.625 266 0.9948 0.4723 0.9948 0.9974
No log 16.75 268 0.8770 0.5453 0.8770 0.9365
No log 16.875 270 0.8548 0.5790 0.8548 0.9245
No log 17.0 272 0.8725 0.5565 0.8725 0.9341
No log 17.125 274 0.8766 0.5573 0.8766 0.9363
No log 17.25 276 1.0214 0.5027 1.0214 1.0106
No log 17.375 278 1.1846 0.4705 1.1846 1.0884
No log 17.5 280 1.1167 0.5224 1.1167 1.0567
No log 17.625 282 0.9284 0.5124 0.9284 0.9636
No log 17.75 284 0.8112 0.5879 0.8112 0.9006
No log 17.875 286 0.7948 0.6018 0.7948 0.8915
No log 18.0 288 0.8000 0.5899 0.8000 0.8944
No log 18.125 290 0.9158 0.5250 0.9158 0.9570
No log 18.25 292 1.1610 0.5272 1.1610 1.0775
No log 18.375 294 1.1604 0.5212 1.1604 1.0772
No log 18.5 296 0.9622 0.5075 0.9622 0.9809
No log 18.625 298 0.8461 0.6013 0.8461 0.9198
No log 18.75 300 0.8317 0.5986 0.8317 0.9120
No log 18.875 302 0.8382 0.5825 0.8382 0.9155
No log 19.0 304 0.9320 0.5192 0.9320 0.9654
No log 19.125 306 1.1031 0.4657 1.1031 1.0503
No log 19.25 308 1.1527 0.4553 1.1527 1.0736
No log 19.375 310 1.0294 0.4675 1.0294 1.0146
No log 19.5 312 0.9354 0.5527 0.9354 0.9672
No log 19.625 314 0.9191 0.5487 0.9191 0.9587
No log 19.75 316 0.9235 0.5458 0.9235 0.9610
No log 19.875 318 0.9865 0.5148 0.9865 0.9932
No log 20.0 320 0.9778 0.4994 0.9778 0.9888
No log 20.125 322 0.9002 0.5768 0.9002 0.9488
No log 20.25 324 0.8784 0.5768 0.8784 0.9372
No log 20.375 326 0.9581 0.4979 0.9581 0.9788
No log 20.5 328 1.0545 0.4959 1.0545 1.0269
No log 20.625 330 1.0195 0.5036 1.0195 1.0097
No log 20.75 332 0.8778 0.5874 0.8778 0.9369
No log 20.875 334 0.8534 0.5816 0.8534 0.9238
No log 21.0 336 0.8912 0.5874 0.8912 0.9440
No log 21.125 338 0.9805 0.6151 0.9805 0.9902
No log 21.25 340 1.1384 0.5445 1.1384 1.0670
No log 21.375 342 1.1055 0.5354 1.1055 1.0514
No log 21.5 344 0.9807 0.5463 0.9807 0.9903
No log 21.625 346 0.8355 0.5995 0.8355 0.9140
No log 21.75 348 0.8473 0.5995 0.8473 0.9205
No log 21.875 350 0.8982 0.5630 0.8982 0.9478
No log 22.0 352 0.9766 0.4971 0.9766 0.9882
No log 22.125 354 1.0505 0.4656 1.0505 1.0249
No log 22.25 356 0.9573 0.5269 0.9573 0.9784
No log 22.375 358 0.8913 0.6139 0.8913 0.9441
No log 22.5 360 0.8730 0.5768 0.8730 0.9343
No log 22.625 362 0.9101 0.5686 0.9101 0.9540
No log 22.75 364 0.8534 0.5527 0.8534 0.9238
No log 22.875 366 0.8205 0.5652 0.8205 0.9058
No log 23.0 368 0.8028 0.5782 0.8028 0.8960
No log 23.125 370 0.8052 0.5868 0.8052 0.8973
No log 23.25 372 0.8364 0.5866 0.8364 0.9145
No log 23.375 374 0.8662 0.5825 0.8662 0.9307
No log 23.5 376 0.8351 0.5741 0.8351 0.9138
No log 23.625 378 0.8199 0.5968 0.8199 0.9055
No log 23.75 380 0.8304 0.6153 0.8304 0.9113
No log 23.875 382 0.8705 0.5684 0.8705 0.9330
No log 24.0 384 0.8598 0.5598 0.8598 0.9272
No log 24.125 386 0.8284 0.6104 0.8284 0.9102
No log 24.25 388 0.8210 0.6288 0.8210 0.9061
No log 24.375 390 0.8551 0.5853 0.8551 0.9247
No log 24.5 392 0.9009 0.5741 0.9009 0.9492
No log 24.625 394 0.8819 0.5755 0.8819 0.9391
No log 24.75 396 0.8175 0.5536 0.8175 0.9042
No log 24.875 398 0.7928 0.5563 0.7928 0.8904
No log 25.0 400 0.7898 0.5449 0.7898 0.8887
No log 25.125 402 0.8016 0.5708 0.8016 0.8953
No log 25.25 404 0.8180 0.5513 0.8180 0.9044
No log 25.375 406 0.8567 0.5273 0.8567 0.9256
No log 25.5 408 0.8538 0.5295 0.8538 0.9240
No log 25.625 410 0.8476 0.5411 0.8476 0.9207
No log 25.75 412 0.8784 0.5428 0.8784 0.9372
No log 25.875 414 0.9011 0.5428 0.9011 0.9492
No log 26.0 416 0.9309 0.5281 0.9309 0.9648
No log 26.125 418 0.8880 0.5339 0.8880 0.9423
No log 26.25 420 0.8449 0.5657 0.8449 0.9192
No log 26.375 422 0.7996 0.5556 0.7996 0.8942
No log 26.5 424 0.8101 0.5319 0.8101 0.9001
No log 26.625 426 0.8345 0.5421 0.8345 0.9135
No log 26.75 428 0.8363 0.5421 0.8363 0.9145
No log 26.875 430 0.8388 0.5370 0.8388 0.9159
No log 27.0 432 0.8565 0.5641 0.8565 0.9255
No log 27.125 434 0.8285 0.5341 0.8285 0.9102
No log 27.25 436 0.8066 0.5478 0.8066 0.8981
No log 27.375 438 0.8091 0.5541 0.8091 0.8995
No log 27.5 440 0.8455 0.5640 0.8455 0.9195
No log 27.625 442 0.9747 0.5262 0.9747 0.9873
No log 27.75 444 1.0629 0.4747 1.0629 1.0310
No log 27.875 446 1.0444 0.5014 1.0444 1.0220
No log 28.0 448 0.9269 0.5250 0.9269 0.9628
No log 28.125 450 0.8675 0.5570 0.8675 0.9314
No log 28.25 452 0.8332 0.5758 0.8332 0.9128
No log 28.375 454 0.8160 0.5758 0.8160 0.9033
No log 28.5 456 0.7810 0.5770 0.7810 0.8838
No log 28.625 458 0.7685 0.5884 0.7685 0.8766
No log 28.75 460 0.7795 0.6146 0.7795 0.8829
No log 28.875 462 0.7850 0.5985 0.7850 0.8860
No log 29.0 464 0.7700 0.5778 0.7700 0.8775
No log 29.125 466 0.7660 0.5636 0.7660 0.8752
No log 29.25 468 0.8018 0.6183 0.8018 0.8954
No log 29.375 470 0.9007 0.5840 0.9007 0.9491
No log 29.5 472 0.9294 0.5826 0.9294 0.9641
No log 29.625 474 0.9567 0.5514 0.9567 0.9781
No log 29.75 476 0.8887 0.5478 0.8887 0.9427
No log 29.875 478 0.8168 0.5728 0.8168 0.9038
No log 30.0 480 0.8120 0.6243 0.8120 0.9011
No log 30.125 482 0.8579 0.6036 0.8579 0.9263
No log 30.25 484 0.8450 0.6019 0.8450 0.9192
No log 30.375 486 0.8070 0.6146 0.8070 0.8983
No log 30.5 488 0.8057 0.6025 0.8057 0.8976
No log 30.625 490 0.8387 0.5859 0.8387 0.9158
No log 30.75 492 0.9128 0.5281 0.9128 0.9554
No log 30.875 494 0.9485 0.5261 0.9485 0.9739
No log 31.0 496 0.8894 0.5313 0.8894 0.9431
No log 31.125 498 0.8229 0.5916 0.8229 0.9072
0.3536 31.25 500 0.7853 0.5770 0.7853 0.8862
0.3536 31.375 502 0.7869 0.5842 0.7869 0.8871
0.3536 31.5 504 0.7902 0.5445 0.7902 0.8889
0.3536 31.625 506 0.7994 0.5439 0.7994 0.8941
0.3536 31.75 508 0.8332 0.5568 0.8332 0.9128
0.3536 31.875 510 0.9256 0.5002 0.9256 0.9621
0.3536 32.0 512 1.0212 0.4718 1.0212 1.0106

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k3_task1_organization

Finetuned
(4023)
this model