ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k8_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7458
  • Qwk: 0.5717
  • Mse: 0.7458
  • Rmse: 0.8636

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0714 2 3.9539 -0.0232 3.9539 1.9885
No log 0.1429 4 2.4245 -0.0512 2.4245 1.5571
No log 0.2143 6 1.4652 -0.0078 1.4652 1.2105
No log 0.2857 8 1.2644 0.1085 1.2644 1.1245
No log 0.3571 10 1.2405 0.1261 1.2405 1.1138
No log 0.4286 12 1.0768 0.3439 1.0768 1.0377
No log 0.5 14 1.3773 -0.0148 1.3773 1.1736
No log 0.5714 16 1.6018 0.0389 1.6018 1.2656
No log 0.6429 18 1.4021 -0.0328 1.4021 1.1841
No log 0.7143 20 1.0623 0.2341 1.0623 1.0307
No log 0.7857 22 0.9946 0.1601 0.9946 0.9973
No log 0.8571 24 0.9912 0.3025 0.9912 0.9956
No log 0.9286 26 0.9989 0.3730 0.9989 0.9994
No log 1.0 28 1.0475 0.2221 1.0475 1.0235
No log 1.0714 30 1.0726 0.1752 1.0726 1.0357
No log 1.1429 32 1.0677 0.1333 1.0677 1.0333
No log 1.2143 34 1.0794 0.1752 1.0794 1.0389
No log 1.2857 36 1.0397 0.1333 1.0397 1.0197
No log 1.3571 38 1.0907 0.2465 1.0907 1.0444
No log 1.4286 40 1.2004 0.1738 1.2004 1.0956
No log 1.5 42 1.0882 0.2343 1.0882 1.0432
No log 1.5714 44 0.9470 0.2316 0.9470 0.9732
No log 1.6429 46 0.9356 0.3528 0.9356 0.9673
No log 1.7143 48 0.9707 0.2503 0.9707 0.9853
No log 1.7857 50 1.0179 0.3028 1.0179 1.0089
No log 1.8571 52 1.0388 0.3107 1.0388 1.0192
No log 1.9286 54 1.0884 0.3056 1.0884 1.0432
No log 2.0 56 1.0058 0.3217 1.0058 1.0029
No log 2.0714 58 0.9055 0.2572 0.9055 0.9516
No log 2.1429 60 0.9619 0.4192 0.9619 0.9808
No log 2.2143 62 0.9481 0.3663 0.9481 0.9737
No log 2.2857 64 0.9204 0.2986 0.9204 0.9594
No log 2.3571 66 0.9143 0.3067 0.9143 0.9562
No log 2.4286 68 0.9968 0.3844 0.9968 0.9984
No log 2.5 70 1.0137 0.3809 1.0137 1.0068
No log 2.5714 72 0.9621 0.4435 0.9621 0.9809
No log 2.6429 74 1.1374 0.4578 1.1374 1.0665
No log 2.7143 76 1.1673 0.4592 1.1673 1.0804
No log 2.7857 78 1.0311 0.4744 1.0311 1.0154
No log 2.8571 80 0.9612 0.4833 0.9612 0.9804
No log 2.9286 82 1.0207 0.4136 1.0207 1.0103
No log 3.0 84 0.9844 0.4672 0.9844 0.9922
No log 3.0714 86 0.9249 0.4138 0.9249 0.9617
No log 3.1429 88 0.9089 0.4181 0.9089 0.9534
No log 3.2143 90 0.9463 0.4034 0.9463 0.9728
No log 3.2857 92 0.9287 0.4006 0.9287 0.9637
No log 3.3571 94 0.8903 0.3521 0.8903 0.9435
No log 3.4286 96 0.8859 0.4076 0.8859 0.9412
No log 3.5 98 0.8773 0.4770 0.8773 0.9367
No log 3.5714 100 0.8774 0.4766 0.8774 0.9367
No log 3.6429 102 0.9068 0.5086 0.9068 0.9523
No log 3.7143 104 0.9172 0.4458 0.9172 0.9577
No log 3.7857 106 0.9156 0.4747 0.9156 0.9569
No log 3.8571 108 0.9013 0.5290 0.9013 0.9494
No log 3.9286 110 1.0932 0.4547 1.0932 1.0456
No log 4.0 112 1.1948 0.3792 1.1948 1.0931
No log 4.0714 114 1.0372 0.5048 1.0372 1.0184
No log 4.1429 116 0.8619 0.5806 0.8619 0.9284
No log 4.2143 118 1.0057 0.4013 1.0057 1.0028
No log 4.2857 120 0.9879 0.4118 0.9879 0.9939
No log 4.3571 122 0.8434 0.3843 0.8434 0.9184
No log 4.4286 124 0.7727 0.5346 0.7727 0.8791
No log 4.5 126 0.7704 0.5746 0.7704 0.8777
No log 4.5714 128 0.8005 0.5412 0.8005 0.8947
No log 4.6429 130 0.7824 0.5859 0.7824 0.8845
No log 4.7143 132 0.8463 0.4480 0.8463 0.9199
No log 4.7857 134 0.8596 0.4590 0.8596 0.9271
No log 4.8571 136 0.7894 0.6046 0.7894 0.8885
No log 4.9286 138 0.8493 0.5518 0.8493 0.9216
No log 5.0 140 0.8850 0.5726 0.8850 0.9407
No log 5.0714 142 0.8480 0.5523 0.8480 0.9208
No log 5.1429 144 0.8924 0.4125 0.8924 0.9447
No log 5.2143 146 0.9788 0.4334 0.9788 0.9893
No log 5.2857 148 0.9531 0.3735 0.9531 0.9763
No log 5.3571 150 0.9534 0.4171 0.9534 0.9764
No log 5.4286 152 0.9423 0.4543 0.9423 0.9707
No log 5.5 154 0.9392 0.4181 0.9392 0.9691
No log 5.5714 156 0.9582 0.4386 0.9582 0.9789
No log 5.6429 158 0.8330 0.5093 0.8330 0.9127
No log 5.7143 160 0.8001 0.5188 0.8001 0.8945
No log 5.7857 162 0.8226 0.4962 0.8226 0.9070
No log 5.8571 164 0.7613 0.4923 0.7613 0.8725
No log 5.9286 166 0.7455 0.5361 0.7455 0.8634
No log 6.0 168 0.7505 0.5346 0.7505 0.8663
No log 6.0714 170 0.7688 0.5093 0.7688 0.8768
No log 6.1429 172 0.7453 0.5346 0.7453 0.8633
No log 6.2143 174 0.7305 0.5234 0.7305 0.8547
No log 6.2857 176 0.7501 0.5606 0.7501 0.8661
No log 6.3571 178 0.8473 0.5729 0.8473 0.9205
No log 6.4286 180 0.8441 0.5835 0.8441 0.9187
No log 6.5 182 0.7560 0.5261 0.7560 0.8695
No log 6.5714 184 0.8293 0.4388 0.8293 0.9107
No log 6.6429 186 0.8230 0.4388 0.8230 0.9072
No log 6.7143 188 0.7524 0.5361 0.7524 0.8674
No log 6.7857 190 0.8168 0.4529 0.8168 0.9038
No log 6.8571 192 0.8288 0.5462 0.8288 0.9104
No log 6.9286 194 0.7712 0.5435 0.7712 0.8782
No log 7.0 196 0.7368 0.6278 0.7368 0.8584
No log 7.0714 198 0.7158 0.5979 0.7158 0.8460
No log 7.1429 200 0.7137 0.5722 0.7137 0.8448
No log 7.2143 202 0.7145 0.5614 0.7145 0.8453
No log 7.2857 204 0.7051 0.5648 0.7051 0.8397
No log 7.3571 206 0.7168 0.5523 0.7168 0.8466
No log 7.4286 208 0.7190 0.5721 0.7190 0.8479
No log 7.5 210 0.7365 0.4853 0.7365 0.8582
No log 7.5714 212 0.7536 0.5482 0.7536 0.8681
No log 7.6429 214 0.7051 0.5304 0.7051 0.8397
No log 7.7143 216 0.7144 0.6393 0.7144 0.8452
No log 7.7857 218 0.8613 0.6043 0.8613 0.9281
No log 7.8571 220 0.9266 0.5647 0.9266 0.9626
No log 7.9286 222 0.8327 0.6263 0.8327 0.9125
No log 8.0 224 0.7742 0.5685 0.7742 0.8799
No log 8.0714 226 0.8119 0.4933 0.8119 0.9011
No log 8.1429 228 0.8067 0.5234 0.8067 0.8982
No log 8.2143 230 0.7272 0.5213 0.7272 0.8528
No log 8.2857 232 0.7204 0.5635 0.7204 0.8488
No log 8.3571 234 0.7335 0.4826 0.7335 0.8564
No log 8.4286 236 0.6998 0.5622 0.6998 0.8365
No log 8.5 238 0.7363 0.5197 0.7363 0.8581
No log 8.5714 240 0.7410 0.5410 0.7410 0.8608
No log 8.6429 242 0.6956 0.5463 0.6956 0.8340
No log 8.7143 244 0.7024 0.5724 0.7024 0.8381
No log 8.7857 246 0.7172 0.5703 0.7172 0.8469
No log 8.8571 248 0.7103 0.5463 0.7103 0.8428
No log 8.9286 250 0.7268 0.5522 0.7268 0.8525
No log 9.0 252 0.7213 0.4975 0.7213 0.8493
No log 9.0714 254 0.7445 0.5622 0.7445 0.8628
No log 9.1429 256 0.8301 0.4864 0.8301 0.9111
No log 9.2143 258 0.8473 0.5011 0.8473 0.9205
No log 9.2857 260 0.7936 0.4691 0.7936 0.8908
No log 9.3571 262 0.7883 0.4741 0.7883 0.8879
No log 9.4286 264 0.7922 0.4755 0.7922 0.8900
No log 9.5 266 0.7921 0.5343 0.7921 0.8900
No log 9.5714 268 0.7909 0.4546 0.7909 0.8893
No log 9.6429 270 0.7807 0.5913 0.7807 0.8836
No log 9.7143 272 0.7545 0.5701 0.7545 0.8686
No log 9.7857 274 0.7325 0.5820 0.7325 0.8558
No log 9.8571 276 0.7157 0.5614 0.7157 0.8460
No log 9.9286 278 0.7095 0.5614 0.7095 0.8423
No log 10.0 280 0.7154 0.5635 0.7154 0.8458
No log 10.0714 282 0.7110 0.5635 0.7110 0.8432
No log 10.1429 284 0.7113 0.5820 0.7113 0.8434
No log 10.2143 286 0.7147 0.5845 0.7147 0.8454
No log 10.2857 288 0.7176 0.5635 0.7176 0.8471
No log 10.3571 290 0.7268 0.5635 0.7268 0.8525
No log 10.4286 292 0.7386 0.5236 0.7386 0.8594
No log 10.5 294 0.7539 0.5635 0.7539 0.8683
No log 10.5714 296 0.8801 0.5607 0.8801 0.9381
No log 10.6429 298 0.8865 0.5719 0.8865 0.9415
No log 10.7143 300 0.7917 0.5530 0.7917 0.8898
No log 10.7857 302 0.7443 0.4661 0.7443 0.8627
No log 10.8571 304 0.7617 0.4982 0.7617 0.8727
No log 10.9286 306 0.7333 0.5024 0.7333 0.8563
No log 11.0 308 0.7521 0.5854 0.7521 0.8673
No log 11.0714 310 0.7683 0.5854 0.7683 0.8765
No log 11.1429 312 0.7578 0.5693 0.7578 0.8705
No log 11.2143 314 0.7746 0.6225 0.7746 0.8801
No log 11.2857 316 0.7581 0.6032 0.7581 0.8707
No log 11.3571 318 0.7335 0.5212 0.7335 0.8565
No log 11.4286 320 0.7286 0.5422 0.7286 0.8536
No log 11.5 322 0.7195 0.5438 0.7195 0.8482
No log 11.5714 324 0.7219 0.5199 0.7219 0.8496
No log 11.6429 326 0.7073 0.5199 0.7073 0.8410
No log 11.7143 328 0.6878 0.5439 0.6878 0.8293
No log 11.7857 330 0.6964 0.5712 0.6964 0.8345
No log 11.8571 332 0.7503 0.5247 0.7503 0.8662
No log 11.9286 334 0.7685 0.5490 0.7685 0.8767
No log 12.0 336 0.7036 0.5932 0.7036 0.8388
No log 12.0714 338 0.6811 0.5722 0.6811 0.8253
No log 12.1429 340 0.6707 0.5009 0.6707 0.8189
No log 12.2143 342 0.7255 0.6278 0.7255 0.8518
No log 12.2857 344 0.7405 0.5963 0.7405 0.8605
No log 12.3571 346 0.6915 0.6529 0.6915 0.8316
No log 12.4286 348 0.6680 0.5274 0.6680 0.8173
No log 12.5 350 0.6565 0.5402 0.6565 0.8103
No log 12.5714 352 0.6553 0.5480 0.6553 0.8095
No log 12.6429 354 0.6595 0.5690 0.6595 0.8121
No log 12.7143 356 0.6854 0.5192 0.6854 0.8279
No log 12.7857 358 0.7814 0.5235 0.7814 0.8840
No log 12.8571 360 0.8465 0.6047 0.8465 0.9201
No log 12.9286 362 0.7649 0.5458 0.7649 0.8746
No log 13.0 364 0.7157 0.5551 0.7157 0.8460
No log 13.0714 366 0.7350 0.6225 0.7350 0.8573
No log 13.1429 368 0.7369 0.6060 0.7369 0.8584
No log 13.2143 370 0.7297 0.5854 0.7297 0.8542
No log 13.2857 372 0.7206 0.5843 0.7206 0.8489
No log 13.3571 374 0.7221 0.5599 0.7221 0.8498
No log 13.4286 376 0.7122 0.5329 0.7122 0.8439
No log 13.5 378 0.7244 0.5522 0.7244 0.8511
No log 13.5714 380 0.7184 0.5316 0.7184 0.8476
No log 13.6429 382 0.7221 0.5662 0.7221 0.8498
No log 13.7143 384 0.7324 0.5662 0.7324 0.8558
No log 13.7857 386 0.7456 0.5832 0.7456 0.8635
No log 13.8571 388 0.7423 0.5933 0.7423 0.8616
No log 13.9286 390 0.7295 0.5446 0.7295 0.8541
No log 14.0 392 0.7272 0.5722 0.7272 0.8527
No log 14.0714 394 0.7361 0.5375 0.7361 0.8579
No log 14.1429 396 0.7628 0.5234 0.7628 0.8734
No log 14.2143 398 0.7860 0.4962 0.7860 0.8866
No log 14.2857 400 0.7892 0.5626 0.7892 0.8884
No log 14.3571 402 0.7570 0.5540 0.7570 0.8701
No log 14.4286 404 0.7503 0.5505 0.7503 0.8662
No log 14.5 406 0.7700 0.5190 0.7700 0.8775
No log 14.5714 408 0.8516 0.5440 0.8516 0.9228
No log 14.6429 410 0.8170 0.5339 0.8170 0.9039
No log 14.7143 412 0.7519 0.5412 0.7519 0.8671
No log 14.7857 414 0.7438 0.5419 0.7438 0.8624
No log 14.8571 416 0.7488 0.5419 0.7488 0.8653
No log 14.9286 418 0.7557 0.5626 0.7557 0.8693
No log 15.0 420 0.8785 0.5937 0.8785 0.9373
No log 15.0714 422 0.9946 0.4104 0.9946 0.9973
No log 15.1429 424 0.9539 0.4764 0.9539 0.9767
No log 15.2143 426 0.8330 0.4238 0.8330 0.9127
No log 15.2857 428 0.7528 0.5373 0.7528 0.8677
No log 15.3571 430 0.7866 0.5081 0.7866 0.8869
No log 15.4286 432 0.7878 0.5422 0.7878 0.8876
No log 15.5 434 0.7838 0.5480 0.7838 0.8853
No log 15.5714 436 0.8219 0.5300 0.8219 0.9066
No log 15.6429 438 0.8273 0.5483 0.8273 0.9095
No log 15.7143 440 0.8249 0.4889 0.8249 0.9082
No log 15.7857 442 0.8600 0.4020 0.8600 0.9274
No log 15.8571 444 0.8679 0.4335 0.8679 0.9316
No log 15.9286 446 0.8571 0.4191 0.8571 0.9258
No log 16.0 448 0.8249 0.5317 0.8249 0.9083
No log 16.0714 450 0.8002 0.4661 0.8002 0.8945
No log 16.1429 452 0.7953 0.4810 0.7953 0.8918
No log 16.2143 454 0.7870 0.5415 0.7870 0.8871
No log 16.2857 456 0.7727 0.5370 0.7727 0.8790
No log 16.3571 458 0.7664 0.5370 0.7664 0.8755
No log 16.4286 460 0.7703 0.5505 0.7703 0.8776
No log 16.5 462 0.7668 0.5575 0.7668 0.8756
No log 16.5714 464 0.7676 0.5557 0.7676 0.8761
No log 16.6429 466 0.7704 0.5226 0.7704 0.8777
No log 16.7143 468 0.7778 0.5919 0.7778 0.8819
No log 16.7857 470 0.8100 0.6009 0.8100 0.9000
No log 16.8571 472 0.8025 0.6194 0.8025 0.8958
No log 16.9286 474 0.7579 0.5989 0.7579 0.8706
No log 17.0 476 0.7576 0.4889 0.7576 0.8704
No log 17.0714 478 0.7779 0.5085 0.7779 0.8820
No log 17.1429 480 0.7596 0.5206 0.7596 0.8716
No log 17.2143 482 0.7468 0.6139 0.7468 0.8642
No log 17.2857 484 0.7938 0.6056 0.7938 0.8910
No log 17.3571 486 0.8428 0.5895 0.8428 0.9181
No log 17.4286 488 0.7893 0.6249 0.7893 0.8884
No log 17.5 490 0.7242 0.6035 0.7242 0.8510
No log 17.5714 492 0.7367 0.5422 0.7367 0.8583
No log 17.6429 494 0.7774 0.5482 0.7774 0.8817
No log 17.7143 496 0.7662 0.5494 0.7662 0.8753
No log 17.7857 498 0.7147 0.5763 0.7147 0.8454
0.2464 17.8571 500 0.7090 0.5964 0.7090 0.8420
0.2464 17.9286 502 0.7065 0.6185 0.7065 0.8406
0.2464 18.0 504 0.6787 0.5851 0.6787 0.8238
0.2464 18.0714 506 0.7194 0.5291 0.7194 0.8482
0.2464 18.1429 508 0.7559 0.6099 0.7559 0.8694
0.2464 18.2143 510 0.7145 0.6082 0.7145 0.8453
0.2464 18.2857 512 0.6888 0.5213 0.6888 0.8299
0.2464 18.3571 514 0.6965 0.5212 0.6965 0.8346
0.2464 18.4286 516 0.6946 0.5443 0.6946 0.8334
0.2464 18.5 518 0.7042 0.5692 0.7042 0.8392
0.2464 18.5714 520 0.7458 0.5717 0.7458 0.8636

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k8_task5_organization

Finetuned
(4019)
this model