ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k12_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7417
  • Qwk: 0.5002
  • Mse: 0.7417
  • Rmse: 0.8612

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.05 2 3.8986 0.0069 3.8986 1.9745
No log 0.1 4 2.2101 -0.0234 2.2101 1.4866
No log 0.15 6 1.3129 0.0 1.3129 1.1458
No log 0.2 8 1.1252 0.2175 1.1252 1.0608
No log 0.25 10 1.2713 0.0380 1.2713 1.1275
No log 0.3 12 1.1860 0.0380 1.1860 1.0891
No log 0.35 14 1.0427 0.3460 1.0427 1.0211
No log 0.4 16 1.0931 0.1352 1.0931 1.0455
No log 0.45 18 1.1453 0.0820 1.1453 1.0702
No log 0.5 20 0.9747 0.3014 0.9747 0.9873
No log 0.55 22 0.9461 0.3243 0.9461 0.9727
No log 0.6 24 0.9433 0.2991 0.9433 0.9713
No log 0.65 26 0.9494 0.2615 0.9494 0.9744
No log 0.7 28 1.0625 0.2903 1.0625 1.0308
No log 0.75 30 1.1884 0.2377 1.1884 1.0901
No log 0.8 32 1.3454 0.1905 1.3454 1.1599
No log 0.85 34 1.2914 0.1354 1.2914 1.1364
No log 0.9 36 1.2334 0.125 1.2334 1.1106
No log 0.95 38 1.2679 0.0380 1.2679 1.1260
No log 1.0 40 1.2594 0.0232 1.2594 1.1222
No log 1.05 42 1.1263 0.0850 1.1263 1.0613
No log 1.1 44 1.0128 0.2114 1.0128 1.0064
No log 1.15 46 1.0034 0.3200 1.0034 1.0017
No log 1.2 48 1.1248 0.2975 1.1248 1.0606
No log 1.25 50 1.1972 0.2729 1.1972 1.0942
No log 1.3 52 1.2813 0.2727 1.2813 1.1319
No log 1.35 54 1.0775 0.2281 1.0775 1.0380
No log 1.4 56 0.9822 0.2713 0.9822 0.9911
No log 1.45 58 0.9424 0.2888 0.9424 0.9708
No log 1.5 60 0.9912 0.3428 0.9912 0.9956
No log 1.55 62 1.0802 0.2590 1.0802 1.0393
No log 1.6 64 1.2054 0.2317 1.2054 1.0979
No log 1.65 66 1.0701 0.2448 1.0701 1.0344
No log 1.7 68 0.9172 0.3207 0.9172 0.9577
No log 1.75 70 0.8746 0.3982 0.8746 0.9352
No log 1.8 72 0.8400 0.3942 0.8400 0.9165
No log 1.85 74 0.8095 0.4416 0.8095 0.8997
No log 1.9 76 0.8363 0.4090 0.8363 0.9145
No log 1.95 78 0.8212 0.4228 0.8212 0.9062
No log 2.0 80 0.8142 0.3488 0.8142 0.9023
No log 2.05 82 0.9078 0.4044 0.9078 0.9528
No log 2.1 84 1.1007 0.2195 1.1007 1.0492
No log 2.15 86 1.0363 0.3844 1.0363 1.0180
No log 2.2 88 0.9949 0.2591 0.9949 0.9975
No log 2.25 90 1.0085 0.3059 1.0085 1.0042
No log 2.3 92 0.9314 0.2764 0.9314 0.9651
No log 2.35 94 0.8592 0.4027 0.8592 0.9270
No log 2.4 96 0.8740 0.3753 0.8740 0.9349
No log 2.45 98 0.9866 0.3631 0.9866 0.9933
No log 2.5 100 0.8922 0.4214 0.8922 0.9445
No log 2.55 102 0.8725 0.4584 0.8725 0.9341
No log 2.6 104 0.8846 0.4815 0.8846 0.9406
No log 2.65 106 1.0096 0.4043 1.0096 1.0048
No log 2.7 108 0.8884 0.4045 0.8884 0.9426
No log 2.75 110 0.7584 0.5175 0.7584 0.8709
No log 2.8 112 0.7003 0.6084 0.7003 0.8368
No log 2.85 114 0.8404 0.4377 0.8404 0.9167
No log 2.9 116 1.0889 0.4379 1.0889 1.0435
No log 2.95 118 1.0972 0.4087 1.0972 1.0475
No log 3.0 120 0.7927 0.4499 0.7927 0.8903
No log 3.05 122 0.7301 0.6035 0.7301 0.8545
No log 3.1 124 0.7440 0.5536 0.7440 0.8625
No log 3.15 126 0.7334 0.6164 0.7334 0.8564
No log 3.2 128 0.7751 0.4737 0.7751 0.8804
No log 3.25 130 0.8420 0.5470 0.8420 0.9176
No log 3.3 132 1.0204 0.4894 1.0204 1.0101
No log 3.35 134 0.9790 0.4802 0.9790 0.9894
No log 3.4 136 0.8101 0.4943 0.8101 0.9001
No log 3.45 138 0.7800 0.5267 0.7800 0.8832
No log 3.5 140 0.7829 0.5977 0.7829 0.8848
No log 3.55 142 0.8259 0.5571 0.8259 0.9088
No log 3.6 144 0.8143 0.5553 0.8143 0.9024
No log 3.65 146 0.8774 0.5045 0.8774 0.9367
No log 3.7 148 0.9675 0.3397 0.9675 0.9836
No log 3.75 150 0.8430 0.4884 0.8430 0.9182
No log 3.8 152 0.7413 0.5262 0.7413 0.8610
No log 3.85 154 0.7107 0.5407 0.7107 0.8430
No log 3.9 156 0.7885 0.5370 0.7885 0.8880
No log 3.95 158 0.7512 0.5173 0.7512 0.8667
No log 4.0 160 0.7311 0.5729 0.7311 0.8550
No log 4.05 162 0.7989 0.5342 0.7989 0.8938
No log 4.1 164 0.8495 0.5294 0.8495 0.9217
No log 4.15 166 0.9484 0.4890 0.9484 0.9739
No log 4.2 168 0.8906 0.5486 0.8906 0.9437
No log 4.25 170 0.7351 0.5722 0.7351 0.8574
No log 4.3 172 0.7567 0.5847 0.7567 0.8699
No log 4.35 174 0.7717 0.5196 0.7717 0.8785
No log 4.4 176 0.9796 0.4910 0.9796 0.9897
No log 4.45 178 0.9473 0.5621 0.9473 0.9733
No log 4.5 180 0.7753 0.5489 0.7753 0.8805
No log 4.55 182 0.7462 0.4203 0.7462 0.8638
No log 4.6 184 0.7578 0.3818 0.7578 0.8705
No log 4.65 186 0.7916 0.5089 0.7916 0.8897
No log 4.7 188 0.8445 0.5173 0.8445 0.9190
No log 4.75 190 0.8310 0.4503 0.8310 0.9116
No log 4.8 192 0.8237 0.3335 0.8237 0.9076
No log 4.85 194 0.8531 0.3603 0.8531 0.9236
No log 4.9 196 0.8969 0.3667 0.8969 0.9471
No log 4.95 198 0.9482 0.4603 0.9482 0.9738
No log 5.0 200 0.9268 0.4603 0.9268 0.9627
No log 5.05 202 0.8345 0.4822 0.8345 0.9135
No log 5.1 204 0.7540 0.5819 0.7540 0.8683
No log 5.15 206 0.7573 0.5527 0.7573 0.8702
No log 5.2 208 0.6901 0.6511 0.6901 0.8307
No log 5.25 210 0.7017 0.6371 0.7017 0.8377
No log 5.3 212 0.6701 0.6224 0.6701 0.8186
No log 5.35 214 0.6708 0.6301 0.6708 0.8190
No log 5.4 216 0.7253 0.6245 0.7253 0.8516
No log 5.45 218 0.7727 0.5224 0.7727 0.8790
No log 5.5 220 0.8488 0.5006 0.8488 0.9213
No log 5.55 222 0.7701 0.6198 0.7701 0.8775
No log 5.6 224 0.7079 0.6365 0.7079 0.8414
No log 5.65 226 0.7154 0.6076 0.7154 0.8458
No log 5.7 228 0.8088 0.5355 0.8088 0.8994
No log 5.75 230 0.8190 0.5355 0.8190 0.9050
No log 5.8 232 0.7571 0.5575 0.7571 0.8701
No log 5.85 234 0.7464 0.5594 0.7464 0.8639
No log 5.9 236 0.7538 0.5678 0.7538 0.8682
No log 5.95 238 0.7448 0.5786 0.7448 0.8630
No log 6.0 240 0.7282 0.5373 0.7282 0.8533
No log 6.05 242 0.7178 0.4927 0.7178 0.8472
No log 6.1 244 0.7442 0.5948 0.7442 0.8627
No log 6.15 246 0.9118 0.4480 0.9118 0.9549
No log 6.2 248 0.8806 0.4476 0.8806 0.9384
No log 6.25 250 0.7721 0.5266 0.7721 0.8787
No log 6.3 252 0.7063 0.6259 0.7063 0.8404
No log 6.35 254 0.7231 0.5430 0.7231 0.8503
No log 6.4 256 0.6965 0.6015 0.6965 0.8346
No log 6.45 258 0.7179 0.6083 0.7179 0.8473
No log 6.5 260 0.9199 0.5090 0.9199 0.9591
No log 6.55 262 0.9342 0.5283 0.9342 0.9665
No log 6.6 264 0.7634 0.5896 0.7634 0.8737
No log 6.65 266 0.7252 0.6078 0.7252 0.8516
No log 6.7 268 0.7537 0.5503 0.7537 0.8681
No log 6.75 270 0.7095 0.6005 0.7095 0.8423
No log 6.8 272 0.7303 0.5797 0.7303 0.8546
No log 6.85 274 0.7113 0.6324 0.7113 0.8434
No log 6.9 276 0.6933 0.6206 0.6933 0.8326
No log 6.95 278 0.7037 0.6584 0.7037 0.8389
No log 7.0 280 0.8008 0.4915 0.8008 0.8949
No log 7.05 282 0.7922 0.5114 0.7922 0.8900
No log 7.1 284 0.6913 0.6511 0.6913 0.8314
No log 7.15 286 0.6741 0.6610 0.6741 0.8210
No log 7.2 288 0.6744 0.6555 0.6744 0.8212
No log 7.25 290 0.6748 0.6288 0.6748 0.8215
No log 7.3 292 0.7343 0.5255 0.7343 0.8569
No log 7.35 294 0.8056 0.5131 0.8056 0.8975
No log 7.4 296 0.7573 0.4929 0.7573 0.8702
No log 7.45 298 0.7103 0.5928 0.7103 0.8428
No log 7.5 300 0.7178 0.5638 0.7178 0.8472
No log 7.55 302 0.7185 0.6262 0.7185 0.8477
No log 7.6 304 0.7414 0.5975 0.7414 0.8610
No log 7.65 306 0.7642 0.5810 0.7642 0.8742
No log 7.7 308 0.7451 0.6296 0.7451 0.8632
No log 7.75 310 0.7018 0.6118 0.7018 0.8377
No log 7.8 312 0.6953 0.5610 0.6953 0.8338
No log 7.85 314 0.7207 0.5117 0.7207 0.8490
No log 7.9 316 0.7767 0.5266 0.7767 0.8813
No log 7.95 318 0.7890 0.5356 0.7890 0.8883
No log 8.0 320 0.7450 0.5677 0.7450 0.8631
No log 8.05 322 0.7087 0.6296 0.7087 0.8418
No log 8.1 324 0.6772 0.6650 0.6772 0.8229
No log 8.15 326 0.6812 0.6231 0.6812 0.8253
No log 8.2 328 0.6651 0.6717 0.6651 0.8155
No log 8.25 330 0.6879 0.6326 0.6879 0.8294
No log 8.3 332 0.6424 0.6397 0.6424 0.8015
No log 8.35 334 0.6252 0.6712 0.6252 0.7907
No log 8.4 336 0.6391 0.6564 0.6391 0.7995
No log 8.45 338 0.6440 0.6564 0.6440 0.8025
No log 8.5 340 0.6516 0.6528 0.6516 0.8072
No log 8.55 342 0.6593 0.6528 0.6593 0.8120
No log 8.6 344 0.6854 0.6459 0.6854 0.8279
No log 8.65 346 0.7569 0.5788 0.7569 0.8700
No log 8.7 348 0.7336 0.6380 0.7336 0.8565
No log 8.75 350 0.7308 0.6128 0.7308 0.8549
No log 8.8 352 0.7367 0.5809 0.7367 0.8583
No log 8.85 354 0.7352 0.5223 0.7352 0.8574
No log 8.9 356 0.7568 0.4646 0.7568 0.8700
No log 8.95 358 0.8159 0.5032 0.8159 0.9033
No log 9.0 360 0.8637 0.5232 0.8637 0.9293
No log 9.05 362 0.8012 0.5165 0.8012 0.8951
No log 9.1 364 0.7551 0.5866 0.7551 0.8690
No log 9.15 366 0.7462 0.5567 0.7462 0.8639
No log 9.2 368 0.7506 0.5675 0.7506 0.8664
No log 9.25 370 0.7766 0.5869 0.7766 0.8812
No log 9.3 372 0.8911 0.5110 0.8911 0.9440
No log 9.35 374 0.8050 0.5537 0.8050 0.8972
No log 9.4 376 0.7276 0.6177 0.7276 0.8530
No log 9.45 378 0.7214 0.6133 0.7214 0.8494
No log 9.5 380 0.7146 0.5370 0.7146 0.8454
No log 9.55 382 0.7519 0.5595 0.7519 0.8671
No log 9.6 384 0.8334 0.5231 0.8334 0.9129
No log 9.65 386 0.8016 0.5150 0.8016 0.8953
No log 9.7 388 0.7401 0.5317 0.7401 0.8603
No log 9.75 390 0.7306 0.5234 0.7306 0.8547
No log 9.8 392 0.7428 0.5197 0.7428 0.8619
No log 9.85 394 0.8091 0.5255 0.8091 0.8995
No log 9.9 396 0.9101 0.5318 0.9101 0.9540
No log 9.95 398 0.9030 0.5318 0.9030 0.9503
No log 10.0 400 0.8855 0.5122 0.8855 0.9410
No log 10.05 402 0.8028 0.4920 0.8028 0.8960
No log 10.1 404 0.7060 0.5690 0.7060 0.8402
No log 10.15 406 0.6981 0.5910 0.6981 0.8355
No log 10.2 408 0.6958 0.5249 0.6958 0.8341
No log 10.25 410 0.7528 0.5658 0.7528 0.8677
No log 10.3 412 0.7797 0.5343 0.7797 0.8830
No log 10.35 414 0.7715 0.5343 0.7715 0.8783
No log 10.4 416 0.7134 0.5098 0.7134 0.8446
No log 10.45 418 0.6951 0.4898 0.6951 0.8338
No log 10.5 420 0.6958 0.4898 0.6958 0.8342
No log 10.55 422 0.7257 0.5383 0.7257 0.8519
No log 10.6 424 0.7440 0.5566 0.7440 0.8626
No log 10.65 426 0.7059 0.5821 0.7059 0.8402
No log 10.7 428 0.6868 0.5954 0.6868 0.8288
No log 10.75 430 0.6870 0.6464 0.6870 0.8289
No log 10.8 432 0.7112 0.6275 0.7112 0.8433
No log 10.85 434 0.7441 0.5875 0.7441 0.8626
No log 10.9 436 0.6983 0.6626 0.6983 0.8357
No log 10.95 438 0.6998 0.5909 0.6998 0.8365
No log 11.0 440 0.6880 0.6228 0.6880 0.8295
No log 11.05 442 0.6758 0.5656 0.6758 0.8220
No log 11.1 444 0.6816 0.5905 0.6816 0.8256
No log 11.15 446 0.7528 0.5549 0.7528 0.8676
No log 11.2 448 0.7661 0.5549 0.7661 0.8753
No log 11.25 450 0.6987 0.5798 0.6987 0.8359
No log 11.3 452 0.6758 0.6301 0.6758 0.8221
No log 11.35 454 0.6904 0.6305 0.6904 0.8309
No log 11.4 456 0.6979 0.5577 0.6979 0.8354
No log 11.45 458 0.7144 0.5658 0.7144 0.8452
No log 11.5 460 0.6543 0.6446 0.6543 0.8089
No log 11.55 462 0.6473 0.6017 0.6473 0.8046
No log 11.6 464 0.6876 0.5560 0.6876 0.8292
No log 11.65 466 0.6651 0.5665 0.6651 0.8155
No log 11.7 468 0.6222 0.6506 0.6222 0.7888
No log 11.75 470 0.6546 0.6035 0.6546 0.8091
No log 11.8 472 0.6596 0.6035 0.6596 0.8122
No log 11.85 474 0.6279 0.6518 0.6279 0.7924
No log 11.9 476 0.6515 0.5581 0.6515 0.8071
No log 11.95 478 0.6608 0.5581 0.6608 0.8129
No log 12.0 480 0.6327 0.6209 0.6327 0.7954
No log 12.05 482 0.7047 0.5243 0.7047 0.8395
No log 12.1 484 0.7745 0.5122 0.7745 0.8801
No log 12.15 486 0.7326 0.5342 0.7326 0.8559
No log 12.2 488 0.7003 0.5810 0.7003 0.8368
No log 12.25 490 0.6934 0.5221 0.6934 0.8327
No log 12.3 492 0.7079 0.4612 0.7079 0.8414
No log 12.35 494 0.7390 0.5070 0.7390 0.8597
No log 12.4 496 0.8471 0.5220 0.8471 0.9204
No log 12.45 498 0.9200 0.5 0.9200 0.9592
0.2821 12.5 500 0.8419 0.5019 0.8419 0.9176
0.2821 12.55 502 0.7355 0.4843 0.7355 0.8576
0.2821 12.6 504 0.7297 0.4983 0.7297 0.8542
0.2821 12.65 506 0.7694 0.4990 0.7694 0.8771
0.2821 12.7 508 0.7487 0.4983 0.7487 0.8653
0.2821 12.75 510 0.7417 0.5002 0.7417 0.8612

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k12_task5_organization

Finetuned
(4019)
this model