ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k3_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.1799
  • Qwk: 0.5781
  • Mse: 1.1799
  • Rmse: 1.0862

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1538 2 6.8790 0.0116 6.8790 2.6228
No log 0.3077 4 4.6895 0.0711 4.6895 2.1655
No log 0.4615 6 3.3853 -0.0109 3.3853 1.8399
No log 0.6154 8 2.2973 0.2695 2.2973 1.5157
No log 0.7692 10 2.0260 0.2927 2.0260 1.4234
No log 0.9231 12 2.0932 0.1860 2.0932 1.4468
No log 1.0769 14 1.7590 0.2833 1.7590 1.3263
No log 1.2308 16 1.4991 0.2632 1.4991 1.2244
No log 1.3846 18 1.4607 0.2655 1.4607 1.2086
No log 1.5385 20 1.4211 0.3276 1.4211 1.1921
No log 1.6923 22 1.4139 0.3089 1.4139 1.1891
No log 1.8462 24 1.4126 0.3492 1.4126 1.1885
No log 2.0 26 1.3899 0.3780 1.3899 1.1790
No log 2.1538 28 1.4500 0.3051 1.4500 1.2042
No log 2.3077 30 1.3809 0.3390 1.3809 1.1751
No log 2.4615 32 1.2324 0.4762 1.2324 1.1101
No log 2.6154 34 1.1829 0.5039 1.1829 1.0876
No log 2.7692 36 1.1377 0.5496 1.1377 1.0666
No log 2.9231 38 1.1134 0.5581 1.1134 1.0552
No log 3.0769 40 1.3854 0.4167 1.3854 1.1770
No log 3.2308 42 1.3597 0.3902 1.3597 1.1661
No log 3.3846 44 1.1029 0.5778 1.1029 1.0502
No log 3.5385 46 1.6821 0.3556 1.6821 1.2969
No log 3.6923 48 1.6703 0.3623 1.6703 1.2924
No log 3.8462 50 1.3287 0.4580 1.3287 1.1527
No log 4.0 52 1.2160 0.4964 1.2160 1.1027
No log 4.1538 54 1.3166 0.4697 1.3166 1.1474
No log 4.3077 56 1.1117 0.6074 1.1117 1.0544
No log 4.4615 58 1.0477 0.5899 1.0477 1.0236
No log 4.6154 60 1.0644 0.5652 1.0644 1.0317
No log 4.7692 62 1.1102 0.5672 1.1102 1.0537
No log 4.9231 64 1.1073 0.5778 1.1073 1.0523
No log 5.0769 66 0.9882 0.5714 0.9882 0.9941
No log 5.2308 68 1.0010 0.5865 1.0010 1.0005
No log 5.3846 70 1.0202 0.5414 1.0202 1.0100
No log 5.5385 72 1.2144 0.5564 1.2144 1.1020
No log 5.6923 74 1.6299 0.3165 1.6299 1.2767
No log 5.8462 76 1.7209 0.3000 1.7209 1.3118
No log 6.0 78 1.4818 0.4179 1.4818 1.2173
No log 6.1538 80 1.2862 0.4662 1.2862 1.1341
No log 6.3077 82 1.3360 0.4545 1.3360 1.1558
No log 6.4615 84 1.3945 0.3710 1.3945 1.1809
No log 6.6154 86 1.6762 0.2742 1.6762 1.2947
No log 6.7692 88 1.9397 0.1463 1.9397 1.3927
No log 6.9231 90 1.6699 0.2880 1.6699 1.2922
No log 7.0769 92 1.2812 0.4252 1.2812 1.1319
No log 7.2308 94 1.0981 0.5882 1.0981 1.0479
No log 7.3846 96 1.2965 0.4925 1.2965 1.1387
No log 7.5385 98 1.5465 0.3407 1.5465 1.2436
No log 7.6923 100 1.6509 0.2774 1.6509 1.2849
No log 7.8462 102 1.3620 0.4328 1.3620 1.1671
No log 8.0 104 1.1476 0.5606 1.1476 1.0713
No log 8.1538 106 1.2599 0.4355 1.2599 1.1225
No log 8.3077 108 1.2850 0.4333 1.2850 1.1336
No log 8.4615 110 1.3498 0.4746 1.3498 1.1618
No log 8.6154 112 1.3730 0.4237 1.3730 1.1718
No log 8.7692 114 1.2500 0.4874 1.2500 1.1180
No log 8.9231 116 1.1898 0.5246 1.1898 1.0908
No log 9.0769 118 1.2384 0.5246 1.2384 1.1128
No log 9.2308 120 1.2161 0.5645 1.2161 1.1028
No log 9.3846 122 1.1797 0.5692 1.1797 1.0862
No log 9.5385 124 1.1548 0.5606 1.1548 1.0746
No log 9.6923 126 1.0363 0.5588 1.0363 1.0180
No log 9.8462 128 1.0388 0.5649 1.0388 1.0192
No log 10.0 130 1.2591 0.5263 1.2591 1.1221
No log 10.1538 132 1.2740 0.5075 1.2740 1.1287
No log 10.3077 134 1.2702 0.5037 1.2702 1.1270
No log 10.4615 136 1.0334 0.5649 1.0334 1.0166
No log 10.6154 138 1.0205 0.6107 1.0205 1.0102
No log 10.7692 140 1.0959 0.6047 1.0959 1.0469
No log 10.9231 142 1.1777 0.5263 1.1777 1.0852
No log 11.0769 144 1.2653 0.4885 1.2653 1.1249
No log 11.2308 146 1.4523 0.3731 1.4523 1.2051
No log 11.3846 148 1.5188 0.3066 1.5188 1.2324
No log 11.5385 150 1.2696 0.5152 1.2696 1.1268
No log 11.6923 152 1.0161 0.5758 1.0161 1.0080
No log 11.8462 154 0.9505 0.6119 0.9505 0.9749
No log 12.0 156 1.0154 0.6260 1.0154 1.0077
No log 12.1538 158 1.0923 0.5954 1.0923 1.0451
No log 12.3077 160 1.0244 0.6212 1.0244 1.0121
No log 12.4615 162 1.0371 0.6015 1.0371 1.0184
No log 12.6154 164 0.9552 0.6569 0.9552 0.9774
No log 12.7692 166 0.8909 0.6714 0.8909 0.9439
No log 12.9231 168 0.9448 0.6370 0.9448 0.9720
No log 13.0769 170 1.0428 0.6260 1.0428 1.0212
No log 13.2308 172 1.0664 0.6 1.0664 1.0326
No log 13.3846 174 1.0311 0.6357 1.0311 1.0154
No log 13.5385 176 0.9768 0.6202 0.9768 0.9884
No log 13.6923 178 0.9365 0.6466 0.9365 0.9677
No log 13.8462 180 0.9935 0.6466 0.9935 0.9967
No log 14.0 182 1.0994 0.5564 1.0994 1.0485
No log 14.1538 184 1.0568 0.5649 1.0568 1.0280
No log 14.3077 186 0.9570 0.6667 0.9570 0.9783
No log 14.4615 188 0.9477 0.6769 0.9477 0.9735
No log 14.6154 190 0.9780 0.6406 0.9780 0.9889
No log 14.7692 192 0.9710 0.6406 0.9710 0.9854
No log 14.9231 194 0.9185 0.6406 0.9185 0.9584
No log 15.0769 196 0.8883 0.6769 0.8883 0.9425
No log 15.2308 198 0.9613 0.6212 0.9613 0.9804
No log 15.3846 200 1.1531 0.5714 1.1531 1.0738
No log 15.5385 202 1.2054 0.5522 1.2054 1.0979
No log 15.6923 204 0.9776 0.6212 0.9776 0.9888
No log 15.8462 206 0.8711 0.6522 0.8711 0.9333
No log 16.0 208 0.8977 0.6269 0.8977 0.9474
No log 16.1538 210 0.9442 0.6119 0.9442 0.9717
No log 16.3077 212 1.1178 0.5385 1.1178 1.0573
No log 16.4615 214 1.4654 0.3969 1.4654 1.2105
No log 16.6154 216 1.5398 0.3731 1.5398 1.2409
No log 16.7692 218 1.3944 0.4341 1.3944 1.1808
No log 16.9231 220 1.1912 0.5806 1.1912 1.0914
No log 17.0769 222 1.1486 0.4874 1.1486 1.0717
No log 17.2308 224 1.1676 0.4576 1.1676 1.0805
No log 17.3846 226 1.2121 0.5167 1.2121 1.1010
No log 17.5385 228 1.1961 0.5455 1.1961 1.0937
No log 17.6923 230 1.1805 0.5691 1.1805 1.0865
No log 17.8462 232 1.1639 0.5873 1.1639 1.0789
No log 18.0 234 1.2193 0.5538 1.2193 1.1042
No log 18.1538 236 1.2416 0.4962 1.2416 1.1143
No log 18.3077 238 1.1364 0.5909 1.1364 1.0660
No log 18.4615 240 1.0141 0.6212 1.0141 1.0070
No log 18.6154 242 0.8960 0.6515 0.8960 0.9466
No log 18.7692 244 0.8859 0.6515 0.8859 0.9412
No log 18.9231 246 0.9256 0.6047 0.9256 0.9621
No log 19.0769 248 1.0612 0.5938 1.0612 1.0302
No log 19.2308 250 1.1211 0.5625 1.1211 1.0588
No log 19.3846 252 1.0641 0.5669 1.0641 1.0315
No log 19.5385 254 0.9766 0.6202 0.9766 0.9883
No log 19.6923 256 0.9448 0.6462 0.9448 0.9720
No log 19.8462 258 0.9012 0.6512 0.9012 0.9493
No log 20.0 260 0.9145 0.6357 0.9145 0.9563
No log 20.1538 262 0.9881 0.6308 0.9881 0.9940
No log 20.3077 264 1.1198 0.5344 1.1198 1.0582
No log 20.4615 266 1.1364 0.5344 1.1364 1.0660
No log 20.6154 268 1.2359 0.5344 1.2359 1.1117
No log 20.7692 270 1.3071 0.4885 1.3071 1.1433
No log 20.9231 272 1.2890 0.5077 1.2890 1.1353
No log 21.0769 274 1.1916 0.5625 1.1916 1.0916
No log 21.2308 276 1.1413 0.6142 1.1413 1.0683
No log 21.3846 278 1.0797 0.6202 1.0797 1.0391
No log 21.5385 280 1.0495 0.6308 1.0495 1.0245
No log 21.6923 282 1.0436 0.6357 1.0436 1.0215
No log 21.8462 284 0.9794 0.6308 0.9794 0.9896
No log 22.0 286 0.9409 0.6357 0.9409 0.9700
No log 22.1538 288 0.9067 0.625 0.9067 0.9522
No log 22.3077 290 0.9269 0.6190 0.9269 0.9628
No log 22.4615 292 0.9859 0.6240 0.9859 0.9929
No log 22.6154 294 1.0478 0.608 1.0478 1.0236
No log 22.7692 296 1.0605 0.5738 1.0605 1.0298
No log 22.9231 298 1.0589 0.5620 1.0589 1.0290
No log 23.0769 300 1.0394 0.5738 1.0394 1.0195
No log 23.2308 302 0.9915 0.6094 0.9915 0.9957
No log 23.3846 304 0.9623 0.6094 0.9623 0.9810
No log 23.5385 306 0.9544 0.6190 0.9544 0.9769
No log 23.6923 308 0.9565 0.625 0.9565 0.9780
No log 23.8462 310 1.0263 0.5827 1.0263 1.0131
No log 24.0 312 1.2389 0.512 1.2389 1.1130
No log 24.1538 314 1.2492 0.5 1.2492 1.1177
No log 24.3077 316 1.0436 0.6107 1.0436 1.0216
No log 24.4615 318 0.8992 0.6154 0.8992 0.9483
No log 24.6154 320 0.8712 0.6047 0.8712 0.9334
No log 24.7692 322 0.8814 0.6154 0.8814 0.9388
No log 24.9231 324 0.9713 0.6357 0.9713 0.9855
No log 25.0769 326 1.0263 0.6094 1.0263 1.0130
No log 25.2308 328 0.9837 0.6357 0.9837 0.9918
No log 25.3846 330 0.9141 0.6202 0.9141 0.9561
No log 25.5385 332 0.8900 0.625 0.8900 0.9434
No log 25.6923 334 0.9046 0.6190 0.9046 0.9511
No log 25.8462 336 0.9370 0.624 0.9370 0.9680
No log 26.0 338 1.0346 0.6349 1.0346 1.0172
No log 26.1538 340 1.1204 0.5938 1.1204 1.0585
No log 26.3077 342 1.1279 0.6047 1.1279 1.0620
No log 26.4615 344 1.0385 0.6406 1.0385 1.0191
No log 26.6154 346 0.9787 0.6400 0.9787 0.9893
No log 26.7692 348 0.9680 0.6299 0.9680 0.9839
No log 26.9231 350 0.9458 0.6357 0.9458 0.9725
No log 27.0769 352 0.9622 0.6357 0.9622 0.9809
No log 27.2308 354 0.9928 0.6357 0.9928 0.9964
No log 27.3846 356 0.9931 0.6357 0.9931 0.9965
No log 27.5385 358 1.0063 0.6357 1.0063 1.0031
No log 27.6923 360 0.9920 0.6357 0.9920 0.9960
No log 27.8462 362 0.9958 0.6357 0.9958 0.9979
No log 28.0 364 1.0336 0.6357 1.0336 1.0167
No log 28.1538 366 1.0675 0.625 1.0675 1.0332
No log 28.3077 368 1.0993 0.625 1.0993 1.0485
No log 28.4615 370 1.0503 0.6357 1.0503 1.0248
No log 28.6154 372 0.9920 0.6308 0.9920 0.9960
No log 28.7692 374 1.0021 0.6308 1.0021 1.0011
No log 28.9231 376 1.0056 0.6308 1.0056 1.0028
No log 29.0769 378 0.9996 0.6308 0.9996 0.9998
No log 29.2308 380 0.9658 0.6047 0.9658 0.9828
No log 29.3846 382 0.9367 0.6032 0.9367 0.9678
No log 29.5385 384 0.9140 0.6308 0.9140 0.9561
No log 29.6923 386 0.9492 0.6357 0.9492 0.9743
No log 29.8462 388 0.9934 0.6357 0.9934 0.9967
No log 30.0 390 1.0133 0.6308 1.0133 1.0066
No log 30.1538 392 1.0141 0.6308 1.0140 1.0070
No log 30.3077 394 0.9652 0.6357 0.9652 0.9825
No log 30.4615 396 0.9676 0.6357 0.9676 0.9837
No log 30.6154 398 0.9469 0.6308 0.9469 0.9731
No log 30.7692 400 0.9429 0.6308 0.9429 0.9710
No log 30.9231 402 0.9832 0.6308 0.9832 0.9916
No log 31.0769 404 1.0503 0.6308 1.0503 1.0248
No log 31.2308 406 1.1537 0.5802 1.1537 1.0741
No log 31.3846 408 1.2028 0.5191 1.2028 1.0967
No log 31.5385 410 1.1059 0.6308 1.1059 1.0516
No log 31.6923 412 0.9944 0.6260 0.9944 0.9972
No log 31.8462 414 0.9568 0.6357 0.9568 0.9782
No log 32.0 416 0.9450 0.6406 0.9450 0.9721
No log 32.1538 418 0.9901 0.6406 0.9901 0.9951
No log 32.3077 420 1.0393 0.6142 1.0393 1.0194
No log 32.4615 422 1.0532 0.6142 1.0532 1.0263
No log 32.6154 424 1.0875 0.6094 1.0875 1.0429
No log 32.7692 426 1.0560 0.6047 1.0560 1.0276
No log 32.9231 428 1.0803 0.6047 1.0803 1.0394
No log 33.0769 430 1.0997 0.6154 1.0997 1.0487
No log 33.2308 432 1.0684 0.6047 1.0684 1.0336
No log 33.3846 434 0.9969 0.6357 0.9969 0.9984
No log 33.5385 436 0.9823 0.5806 0.9823 0.9911
No log 33.6923 438 1.0075 0.56 1.0075 1.0038
No log 33.8462 440 1.0125 0.56 1.0125 1.0062
No log 34.0 442 1.0106 0.5620 1.0106 1.0053
No log 34.1538 444 1.0375 0.6299 1.0375 1.0186
No log 34.3077 446 1.0578 0.6406 1.0578 1.0285
No log 34.4615 448 1.0305 0.6406 1.0305 1.0151
No log 34.6154 450 1.0036 0.608 1.0036 1.0018
No log 34.7692 452 0.9968 0.6512 0.9968 0.9984
No log 34.9231 454 1.0464 0.6357 1.0464 1.0229
No log 35.0769 456 1.0925 0.6047 1.0925 1.0452
No log 35.2308 458 1.1627 0.6047 1.1627 1.0783
No log 35.3846 460 1.2130 0.6 1.2130 1.1014
No log 35.5385 462 1.1700 0.6047 1.1700 1.0817
No log 35.6923 464 1.0775 0.6047 1.0775 1.0380
No log 35.8462 466 0.9984 0.6357 0.9984 0.9992
No log 36.0 468 0.9561 0.6462 0.9561 0.9778
No log 36.1538 470 0.9766 0.5938 0.9766 0.9882
No log 36.3077 472 1.0218 0.5873 1.0218 1.0108
No log 36.4615 474 1.0432 0.6406 1.0432 1.0214
No log 36.6154 476 1.0819 0.6357 1.0819 1.0401
No log 36.7692 478 1.1114 0.6142 1.1114 1.0542
No log 36.9231 480 1.1168 0.6142 1.1168 1.0568
No log 37.0769 482 1.0837 0.6190 1.0837 1.0410
No log 37.2308 484 1.0513 0.608 1.0513 1.0253
No log 37.3846 486 1.0264 0.6299 1.0264 1.0131
No log 37.5385 488 1.0411 0.6357 1.0411 1.0203
No log 37.6923 490 1.1081 0.6094 1.1081 1.0527
No log 37.8462 492 1.1698 0.5512 1.1698 1.0816
No log 38.0 494 1.1797 0.5397 1.1797 1.0861
No log 38.1538 496 1.1269 0.6094 1.1269 1.0616
No log 38.3077 498 1.0682 0.6094 1.0682 1.0336
0.2694 38.4615 500 1.0572 0.625 1.0572 1.0282
0.2694 38.6154 502 1.0819 0.625 1.0819 1.0402
0.2694 38.7692 504 1.1057 0.5984 1.1057 1.0515
0.2694 38.9231 506 1.1214 0.6094 1.1214 1.0590
0.2694 39.0769 508 1.1595 0.6094 1.1595 1.0768
0.2694 39.2308 510 1.1799 0.5781 1.1799 1.0862

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k3_task1_organization

Finetuned
(4019)
this model