ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k10_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7580
  • Qwk: 0.6571
  • Mse: 0.7580
  • Rmse: 0.8707

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0417 2 6.9338 -0.0058 6.9338 2.6332
No log 0.0833 4 4.8169 0.0823 4.8169 2.1947
No log 0.125 6 4.7240 -0.0845 4.7240 2.1735
No log 0.1667 8 3.0901 -0.0120 3.0901 1.7579
No log 0.2083 10 2.1405 0.1129 2.1405 1.4630
No log 0.25 12 1.9724 0.1429 1.9724 1.4044
No log 0.2917 14 1.8892 0.1509 1.8892 1.3745
No log 0.3333 16 1.8319 0.1538 1.8319 1.3535
No log 0.375 18 1.8422 0.1509 1.8422 1.3573
No log 0.4167 20 1.8038 0.1495 1.8038 1.3430
No log 0.4583 22 1.8284 0.2783 1.8284 1.3522
No log 0.5 24 2.0006 0.2188 2.0006 1.4144
No log 0.5417 26 2.3467 0.2029 2.3467 1.5319
No log 0.5833 28 2.0743 0.1654 2.0743 1.4402
No log 0.625 30 1.6918 0.2881 1.6918 1.3007
No log 0.6667 32 1.8086 0.2927 1.8086 1.3448
No log 0.7083 34 1.8562 0.25 1.8562 1.3624
No log 0.75 36 1.6794 0.2353 1.6794 1.2959
No log 0.7917 38 1.4983 0.3000 1.4983 1.2240
No log 0.8333 40 1.7190 0.3939 1.7190 1.3111
No log 0.875 42 2.0270 0.2590 2.0270 1.4237
No log 0.9167 44 2.2363 0.1972 2.2363 1.4954
No log 0.9583 46 2.1540 0.2286 2.1540 1.4677
No log 1.0 48 1.9657 0.2628 1.9657 1.4020
No log 1.0417 50 1.8555 0.3134 1.8555 1.3622
No log 1.0833 52 1.8058 0.3788 1.8058 1.3438
No log 1.125 54 1.7536 0.3664 1.7536 1.3242
No log 1.1667 56 1.6380 0.3411 1.6380 1.2798
No log 1.2083 58 1.6961 0.3692 1.6961 1.3024
No log 1.25 60 1.7979 0.3284 1.7979 1.3408
No log 1.2917 62 2.3074 0.2177 2.3074 1.5190
No log 1.3333 64 2.1084 0.2384 2.1084 1.4520
No log 1.375 66 1.5984 0.4 1.5984 1.2643
No log 1.4167 68 1.8386 0.2628 1.8386 1.3559
No log 1.4583 70 1.8308 0.2963 1.8308 1.3531
No log 1.5 72 1.5294 0.4 1.5294 1.2367
No log 1.5417 74 1.2290 0.4964 1.2290 1.1086
No log 1.5833 76 1.5969 0.4720 1.5969 1.2637
No log 1.625 78 1.9367 0.4583 1.9367 1.3916
No log 1.6667 80 1.9995 0.4767 1.9995 1.4140
No log 1.7083 82 2.0895 0.4044 2.0895 1.4455
No log 1.75 84 1.7491 0.4557 1.7491 1.3225
No log 1.7917 86 1.4832 0.4539 1.4832 1.2178
No log 1.8333 88 1.4339 0.4493 1.4339 1.1975
No log 1.875 90 1.4687 0.4173 1.4687 1.2119
No log 1.9167 92 1.3651 0.4776 1.3651 1.1684
No log 1.9583 94 1.1450 0.5312 1.1450 1.0701
No log 2.0 96 1.0694 0.528 1.0694 1.0341
No log 2.0417 98 1.0643 0.5 1.0643 1.0316
No log 2.0833 100 1.0731 0.5469 1.0731 1.0359
No log 2.125 102 1.1067 0.5538 1.1067 1.0520
No log 2.1667 104 1.2224 0.5224 1.2224 1.1056
No log 2.2083 106 1.3325 0.5547 1.3325 1.1543
No log 2.25 108 1.3594 0.5255 1.3594 1.1659
No log 2.2917 110 1.2461 0.5426 1.2461 1.1163
No log 2.3333 112 1.1649 0.5426 1.1649 1.0793
No log 2.375 114 1.2825 0.4925 1.2825 1.1325
No log 2.4167 116 1.4325 0.5180 1.4325 1.1969
No log 2.4583 118 1.6123 0.4706 1.6123 1.2698
No log 2.5 120 1.5230 0.4658 1.5230 1.2341
No log 2.5417 122 1.3743 0.5248 1.3743 1.1723
No log 2.5833 124 1.3488 0.4964 1.3488 1.1614
No log 2.625 126 1.2541 0.6015 1.2541 1.1199
No log 2.6667 128 1.1842 0.5564 1.1842 1.0882
No log 2.7083 130 1.2023 0.5263 1.2023 1.0965
No log 2.75 132 1.2201 0.5263 1.2201 1.1046
No log 2.7917 134 1.1986 0.5373 1.1986 1.0948
No log 2.8333 136 1.1908 0.5496 1.1908 1.0912
No log 2.875 138 1.1711 0.5263 1.1711 1.0822
No log 2.9167 140 1.4487 0.4828 1.4487 1.2036
No log 2.9583 142 2.0705 0.3473 2.0705 1.4389
No log 3.0 144 2.0432 0.3596 2.0432 1.4294
No log 3.0417 146 1.6651 0.5062 1.6651 1.2904
No log 3.0833 148 1.3299 0.6093 1.3299 1.1532
No log 3.125 150 1.1404 0.5455 1.1404 1.0679
No log 3.1667 152 1.0087 0.5231 1.0087 1.0043
No log 3.2083 154 0.9787 0.5426 0.9787 0.9893
No log 3.25 156 0.9179 0.6569 0.9179 0.9581
No log 3.2917 158 0.8988 0.6165 0.8988 0.9481
No log 3.3333 160 0.9403 0.5926 0.9403 0.9697
No log 3.375 162 0.9890 0.5942 0.9890 0.9945
No log 3.4167 164 0.9504 0.5899 0.9504 0.9749
No log 3.4583 166 1.0165 0.5797 1.0165 1.0082
No log 3.5 168 1.0268 0.5970 1.0268 1.0133
No log 3.5417 170 1.0656 0.5957 1.0656 1.0323
No log 3.5833 172 1.2401 0.6329 1.2401 1.1136
No log 3.625 174 1.2818 0.5786 1.2818 1.1322
No log 3.6667 176 1.1922 0.5816 1.1922 1.0919
No log 3.7083 178 1.0368 0.5882 1.0368 1.0182
No log 3.75 180 1.0446 0.5846 1.0446 1.0221
No log 3.7917 182 1.2155 0.512 1.2155 1.1025
No log 3.8333 184 1.2322 0.4762 1.2322 1.1100
No log 3.875 186 1.1016 0.5781 1.1016 1.0496
No log 3.9167 188 0.9983 0.6154 0.9983 0.9992
No log 3.9583 190 1.0376 0.5781 1.0376 1.0186
No log 4.0 192 1.1018 0.5797 1.1018 1.0497
No log 4.0417 194 1.1067 0.5693 1.1067 1.0520
No log 4.0833 196 1.0292 0.6316 1.0292 1.0145
No log 4.125 198 0.9604 0.5954 0.9604 0.9800
No log 4.1667 200 0.9175 0.6316 0.9175 0.9579
No log 4.2083 202 0.9059 0.6418 0.9059 0.9518
No log 4.25 204 0.8937 0.6165 0.8937 0.9454
No log 4.2917 206 0.9313 0.6043 0.9313 0.9650
No log 4.3333 208 0.9872 0.5957 0.9872 0.9936
No log 4.375 210 1.0694 0.6234 1.0694 1.0341
No log 4.4167 212 1.0282 0.6323 1.0282 1.0140
No log 4.4583 214 0.9978 0.5986 0.9978 0.9989
No log 4.5 216 1.0196 0.6621 1.0196 1.0098
No log 4.5417 218 0.9798 0.6622 0.9798 0.9899
No log 4.5833 220 0.9048 0.6351 0.9048 0.9512
No log 4.625 222 0.9075 0.6122 0.9075 0.9526
No log 4.6667 224 0.9418 0.6383 0.9418 0.9704
No log 4.7083 226 1.0362 0.6056 1.0362 1.0179
No log 4.75 228 1.1230 0.6207 1.1230 1.0597
No log 4.7917 230 1.0562 0.6187 1.0562 1.0277
No log 4.8333 232 0.9362 0.5758 0.9362 0.9676
No log 4.875 234 0.9128 0.6667 0.9128 0.9554
No log 4.9167 236 0.9073 0.7059 0.9073 0.9525
No log 4.9583 238 0.8737 0.6861 0.8737 0.9347
No log 5.0 240 0.8989 0.6176 0.8989 0.9481
No log 5.0417 242 1.0006 0.5926 1.0006 1.0003
No log 5.0833 244 1.1183 0.5797 1.1183 1.0575
No log 5.125 246 1.0705 0.5957 1.0705 1.0346
No log 5.1667 248 1.0382 0.6351 1.0382 1.0189
No log 5.2083 250 0.8836 0.64 0.8836 0.9400
No log 5.25 252 0.7511 0.7347 0.7511 0.8667
No log 5.2917 254 0.7770 0.7413 0.7770 0.8815
No log 5.3333 256 0.8374 0.7162 0.8374 0.9151
No log 5.375 258 0.7934 0.6812 0.7934 0.8907
No log 5.4167 260 0.7717 0.7286 0.7717 0.8785
No log 5.4583 262 0.9368 0.6014 0.9368 0.9679
No log 5.5 264 1.1097 0.6369 1.1097 1.0534
No log 5.5417 266 1.0583 0.6460 1.0583 1.0288
No log 5.5833 268 0.9100 0.6093 0.9100 0.9539
No log 5.625 270 0.8395 0.6800 0.8395 0.9163
No log 5.6667 272 0.8506 0.6711 0.8506 0.9223
No log 5.7083 274 0.8567 0.6875 0.8567 0.9256
No log 5.75 276 0.8855 0.6624 0.8855 0.9410
No log 5.7917 278 0.7916 0.6923 0.7916 0.8897
No log 5.8333 280 0.7723 0.6846 0.7723 0.8788
No log 5.875 282 0.7382 0.6950 0.7382 0.8592
No log 5.9167 284 0.7075 0.7234 0.7075 0.8411
No log 5.9583 286 0.6931 0.7234 0.6931 0.8325
No log 6.0 288 0.6966 0.7092 0.6966 0.8346
No log 6.0417 290 0.7271 0.7034 0.7271 0.8527
No log 6.0833 292 0.8854 0.6582 0.8854 0.9410
No log 6.125 294 1.0394 0.6585 1.0394 1.0195
No log 6.1667 296 0.9479 0.6541 0.9479 0.9736
No log 6.2083 298 0.7249 0.7075 0.7249 0.8514
No log 6.25 300 0.6917 0.7338 0.6917 0.8317
No log 6.2917 302 0.7595 0.7059 0.7595 0.8715
No log 6.3333 304 0.7714 0.7059 0.7714 0.8783
No log 6.375 306 0.7577 0.6857 0.7577 0.8705
No log 6.4167 308 0.9051 0.6099 0.9051 0.9514
No log 6.4583 310 0.9909 0.5541 0.9909 0.9955
No log 6.5 312 0.9484 0.5850 0.9484 0.9739
No log 6.5417 314 0.8680 0.6443 0.8680 0.9317
No log 6.5833 316 0.8356 0.6713 0.8356 0.9141
No log 6.625 318 0.8193 0.6761 0.8193 0.9051
No log 6.6667 320 0.8066 0.6667 0.8066 0.8981
No log 6.7083 322 0.8021 0.6667 0.8021 0.8956
No log 6.75 324 0.8383 0.6761 0.8383 0.9156
No log 6.7917 326 0.9418 0.5890 0.9418 0.9705
No log 6.8333 328 0.9671 0.5906 0.9671 0.9834
No log 6.875 330 0.9312 0.6081 0.9312 0.9650
No log 6.9167 332 0.9191 0.6369 0.9191 0.9587
No log 6.9583 334 0.9374 0.6329 0.9374 0.9682
No log 7.0 336 0.9014 0.6323 0.9014 0.9494
No log 7.0417 338 0.8367 0.6974 0.8367 0.9147
No log 7.0833 340 0.8471 0.6667 0.8471 0.9204
No log 7.125 342 0.8403 0.6714 0.8403 0.9167
No log 7.1667 344 0.8092 0.6519 0.8092 0.8996
No log 7.2083 346 0.7934 0.6912 0.7934 0.8907
No log 7.25 348 0.7887 0.6912 0.7887 0.8881
No log 7.2917 350 0.8834 0.6377 0.8834 0.9399
No log 7.3333 352 1.0851 0.5890 1.0851 1.0417
No log 7.375 354 1.0841 0.6081 1.0841 1.0412
No log 7.4167 356 0.9015 0.6111 0.9015 0.9495
No log 7.4583 358 0.7806 0.6418 0.7806 0.8835
No log 7.5 360 0.7912 0.6519 0.7912 0.8895
No log 7.5417 362 0.8102 0.6667 0.8102 0.9001
No log 7.5833 364 0.7880 0.6809 0.7880 0.8877
No log 7.625 366 0.7720 0.6763 0.7720 0.8786
No log 7.6667 368 0.7926 0.6324 0.7926 0.8903
No log 7.7083 370 0.8544 0.6222 0.8544 0.9243
No log 7.75 372 0.9611 0.6710 0.9611 0.9803
No log 7.7917 374 0.9771 0.6460 0.9771 0.9885
No log 7.8333 376 0.9485 0.6707 0.9485 0.9739
No log 7.875 378 1.0035 0.6705 1.0035 1.0017
No log 7.9167 380 0.9525 0.6857 0.9525 0.9760
No log 7.9583 382 0.8487 0.6581 0.8487 0.9213
No log 8.0 384 0.8437 0.6579 0.8437 0.9185
No log 8.0417 386 0.8470 0.6622 0.8470 0.9203
No log 8.0833 388 0.8630 0.6620 0.8630 0.9290
No log 8.125 390 0.8521 0.6475 0.8521 0.9231
No log 8.1667 392 0.8409 0.6423 0.8409 0.9170
No log 8.2083 394 0.8963 0.6475 0.8963 0.9467
No log 8.25 396 1.0796 0.6294 1.0796 1.0390
No log 8.2917 398 1.1768 0.6415 1.1768 1.0848
No log 8.3333 400 1.0163 0.6887 1.0163 1.0081
No log 8.375 402 0.7810 0.6713 0.7810 0.8837
No log 8.4167 404 0.6657 0.7050 0.6657 0.8159
No log 8.4583 406 0.7659 0.7586 0.7659 0.8752
No log 8.5 408 0.8339 0.7042 0.8339 0.9132
No log 8.5417 410 0.7764 0.7246 0.7764 0.8811
No log 8.5833 412 0.6792 0.7206 0.6792 0.8241
No log 8.625 414 0.7024 0.6897 0.7024 0.8381
No log 8.6667 416 0.8823 0.6871 0.8823 0.9393
No log 8.7083 418 0.9712 0.6460 0.9712 0.9855
No log 8.75 420 0.9444 0.6584 0.9444 0.9718
No log 8.7917 422 0.8348 0.6797 0.8348 0.9136
No log 8.8333 424 0.8017 0.6986 0.8017 0.8954
No log 8.875 426 0.7947 0.6809 0.7947 0.8914
No log 8.9167 428 0.7778 0.7 0.7778 0.8819
No log 8.9583 430 0.7745 0.6765 0.7745 0.8801
No log 9.0 432 0.7666 0.6718 0.7666 0.8756
No log 9.0417 434 0.7651 0.6615 0.7651 0.8747
No log 9.0833 436 0.7273 0.6716 0.7273 0.8528
No log 9.125 438 0.6853 0.6912 0.6853 0.8279
No log 9.1667 440 0.6823 0.6993 0.6823 0.8260
No log 9.2083 442 0.7372 0.7114 0.7372 0.8586
No log 9.25 444 0.8797 0.7037 0.8797 0.9379
No log 9.2917 446 1.0444 0.6867 1.0444 1.0220
No log 9.3333 448 1.0286 0.6897 1.0286 1.0142
No log 9.375 450 0.8826 0.7 0.8826 0.9395
No log 9.4167 452 0.7436 0.7114 0.7436 0.8623
No log 9.4583 454 0.6942 0.6993 0.6942 0.8332
No log 9.5 456 0.6815 0.7194 0.6815 0.8255
No log 9.5417 458 0.7222 0.7259 0.7222 0.8499
No log 9.5833 460 0.7255 0.7259 0.7255 0.8517
No log 9.625 462 0.7492 0.7007 0.7492 0.8656
No log 9.6667 464 0.7751 0.6667 0.7751 0.8804
No log 9.7083 466 0.8147 0.6331 0.8147 0.9026
No log 9.75 468 0.8387 0.6383 0.8387 0.9158
No log 9.7917 470 0.8614 0.6479 0.8614 0.9281
No log 9.8333 472 0.8089 0.6479 0.8089 0.8994
No log 9.875 474 0.7260 0.6714 0.7260 0.8521
No log 9.9167 476 0.7183 0.6620 0.7183 0.8475
No log 9.9583 478 0.6961 0.6815 0.6961 0.8343
No log 10.0 480 0.6805 0.7194 0.6805 0.8249
No log 10.0417 482 0.6895 0.7194 0.6895 0.8303
No log 10.0833 484 0.7307 0.6815 0.7307 0.8548
No log 10.125 486 0.8319 0.6165 0.8319 0.9121
No log 10.1667 488 0.9783 0.6197 0.9783 0.9891
No log 10.2083 490 0.9726 0.6197 0.9726 0.9862
No log 10.25 492 0.8742 0.6383 0.8742 0.9350
No log 10.2917 494 0.7835 0.6522 0.7835 0.8852
No log 10.3333 496 0.7264 0.7194 0.7264 0.8523
No log 10.375 498 0.7246 0.7376 0.7246 0.8512
0.3971 10.4167 500 0.7562 0.6861 0.7562 0.8696
0.3971 10.4583 502 0.8768 0.5882 0.8768 0.9364
0.3971 10.5 504 0.9793 0.6053 0.9793 0.9896
0.3971 10.5417 506 0.9770 0.6323 0.9770 0.9885
0.3971 10.5833 508 0.8596 0.6122 0.8596 0.9272
0.3971 10.625 510 0.7580 0.6571 0.7580 0.8707

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k10_task1_organization

Finetuned
(4019)
this model