ArabicNewSplits6_FineTuningAraBERTFreeze_run2_AugV5_k2_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7379
  • Qwk: 0.5348
  • Mse: 0.7379
  • Rmse: 0.8590

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.2857 2 6.3223 -0.0278 6.3223 2.5144
No log 0.5714 4 4.1469 -0.0186 4.1469 2.0364
No log 0.8571 6 2.8200 0.0177 2.8200 1.6793
No log 1.1429 8 1.9186 0.0707 1.9186 1.3851
No log 1.4286 10 1.3137 0.0592 1.3137 1.1461
No log 1.7143 12 0.9507 0.0659 0.9507 0.9751
No log 2.0 14 0.7585 0.2207 0.7585 0.8709
No log 2.2857 16 0.7038 0.1869 0.7038 0.8389
No log 2.5714 18 0.6913 0.2549 0.6913 0.8314
No log 2.8571 20 0.6618 0.2901 0.6618 0.8135
No log 3.1429 22 0.6364 0.2868 0.6364 0.7978
No log 3.4286 24 0.6239 0.3051 0.6239 0.7899
No log 3.7143 26 0.6150 0.3406 0.6150 0.7842
No log 4.0 28 0.6029 0.3557 0.6029 0.7765
No log 4.2857 30 0.6183 0.3996 0.6183 0.7863
No log 4.5714 32 0.6830 0.3149 0.6830 0.8264
No log 4.8571 34 0.6970 0.3060 0.6970 0.8349
No log 5.1429 36 0.6602 0.3378 0.6602 0.8125
No log 5.4286 38 0.6170 0.3754 0.6170 0.7855
No log 5.7143 40 0.5800 0.3792 0.5800 0.7616
No log 6.0 42 0.5489 0.3865 0.5489 0.7409
No log 6.2857 44 0.5404 0.3623 0.5404 0.7351
No log 6.5714 46 0.5360 0.3872 0.5360 0.7321
No log 6.8571 48 0.5217 0.4336 0.5217 0.7223
No log 7.1429 50 0.5200 0.3953 0.5200 0.7211
No log 7.4286 52 0.5215 0.4281 0.5215 0.7221
No log 7.7143 54 0.5290 0.4018 0.5290 0.7273
No log 8.0 56 0.5577 0.4090 0.5577 0.7468
No log 8.2857 58 0.5906 0.3984 0.5906 0.7685
No log 8.5714 60 0.5810 0.3869 0.5810 0.7622
No log 8.8571 62 0.5694 0.4103 0.5694 0.7546
No log 9.1429 64 0.5708 0.4678 0.5708 0.7555
No log 9.4286 66 0.6283 0.4776 0.6283 0.7927
No log 9.7143 68 0.6458 0.4339 0.6458 0.8036
No log 10.0 70 0.5933 0.4965 0.5933 0.7703
No log 10.2857 72 0.5484 0.5119 0.5484 0.7405
No log 10.5714 74 0.5170 0.4415 0.5170 0.7190
No log 10.8571 76 0.5120 0.4493 0.5120 0.7155
No log 11.1429 78 0.5126 0.4259 0.5126 0.7159
No log 11.4286 80 0.5250 0.4616 0.5250 0.7246
No log 11.7143 82 0.5615 0.4745 0.5615 0.7494
No log 12.0 84 0.5735 0.4761 0.5735 0.7573
No log 12.2857 86 0.5607 0.4565 0.5607 0.7488
No log 12.5714 88 0.5474 0.5301 0.5474 0.7399
No log 12.8571 90 0.5526 0.5011 0.5526 0.7433
No log 13.1429 92 0.5847 0.4915 0.5847 0.7647
No log 13.4286 94 0.6117 0.4756 0.6117 0.7821
No log 13.7143 96 0.6097 0.5162 0.6097 0.7808
No log 14.0 98 0.5819 0.5457 0.5819 0.7628
No log 14.2857 100 0.5611 0.5241 0.5611 0.7491
No log 14.5714 102 0.5599 0.5314 0.5599 0.7483
No log 14.8571 104 0.5717 0.5082 0.5717 0.7561
No log 15.1429 106 0.6010 0.5352 0.6010 0.7752
No log 15.4286 108 0.6486 0.5422 0.6486 0.8054
No log 15.7143 110 0.6618 0.5346 0.6618 0.8135
No log 16.0 112 0.6345 0.5653 0.6345 0.7965
No log 16.2857 114 0.6011 0.5095 0.6011 0.7753
No log 16.5714 116 0.5931 0.5423 0.5931 0.7701
No log 16.8571 118 0.5900 0.5324 0.5900 0.7681
No log 17.1429 120 0.5979 0.5280 0.5979 0.7733
No log 17.4286 122 0.6282 0.5256 0.6282 0.7926
No log 17.7143 124 0.6509 0.5413 0.6509 0.8068
No log 18.0 126 0.6735 0.5582 0.6735 0.8206
No log 18.2857 128 0.6847 0.5512 0.6847 0.8275
No log 18.5714 130 0.6574 0.5468 0.6574 0.8108
No log 18.8571 132 0.6431 0.4980 0.6431 0.8020
No log 19.1429 134 0.6654 0.5427 0.6654 0.8157
No log 19.4286 136 0.7020 0.5608 0.7020 0.8378
No log 19.7143 138 0.7317 0.5294 0.7317 0.8554
No log 20.0 140 0.7061 0.5499 0.7061 0.8403
No log 20.2857 142 0.6601 0.5054 0.6601 0.8125
No log 20.5714 144 0.6402 0.4968 0.6402 0.8002
No log 20.8571 146 0.6443 0.5061 0.6443 0.8027
No log 21.1429 148 0.6630 0.4932 0.6630 0.8143
No log 21.4286 150 0.7029 0.5457 0.7029 0.8384
No log 21.7143 152 0.7248 0.5554 0.7248 0.8513
No log 22.0 154 0.7156 0.5378 0.7156 0.8459
No log 22.2857 156 0.7005 0.5077 0.7005 0.8370
No log 22.5714 158 0.7069 0.5457 0.7069 0.8408
No log 22.8571 160 0.7177 0.5391 0.7177 0.8472
No log 23.1429 162 0.7354 0.5309 0.7354 0.8575
No log 23.4286 164 0.7485 0.5155 0.7485 0.8652
No log 23.7143 166 0.7447 0.5126 0.7447 0.8629
No log 24.0 168 0.7260 0.5577 0.7260 0.8521
No log 24.2857 170 0.6998 0.5678 0.6998 0.8365
No log 24.5714 172 0.7011 0.4975 0.7011 0.8373
No log 24.8571 174 0.7264 0.5514 0.7264 0.8523
No log 25.1429 176 0.7611 0.5655 0.7611 0.8724
No log 25.4286 178 0.7825 0.5812 0.7825 0.8846
No log 25.7143 180 0.7783 0.5363 0.7783 0.8822
No log 26.0 182 0.7610 0.5100 0.7610 0.8723
No log 26.2857 184 0.7604 0.5225 0.7604 0.8720
No log 26.5714 186 0.7765 0.5116 0.7765 0.8812
No log 26.8571 188 0.7785 0.5034 0.7785 0.8823
No log 27.1429 190 0.7762 0.5340 0.7762 0.8810
No log 27.4286 192 0.7800 0.5427 0.7800 0.8832
No log 27.7143 194 0.7866 0.5342 0.7866 0.8869
No log 28.0 196 0.7878 0.5329 0.7878 0.8876
No log 28.2857 198 0.7877 0.5427 0.7877 0.8875
No log 28.5714 200 0.7863 0.5199 0.7863 0.8867
No log 28.8571 202 0.7802 0.5244 0.7802 0.8833
No log 29.1429 204 0.7624 0.5341 0.7624 0.8731
No log 29.4286 206 0.7574 0.5341 0.7574 0.8703
No log 29.7143 208 0.7610 0.5739 0.7610 0.8724
No log 30.0 210 0.7881 0.5655 0.7881 0.8878
No log 30.2857 212 0.7990 0.5590 0.7990 0.8939
No log 30.5714 214 0.7955 0.5398 0.7955 0.8919
No log 30.8571 216 0.7967 0.5230 0.7967 0.8926
No log 31.1429 218 0.8047 0.5230 0.8047 0.8970
No log 31.4286 220 0.8077 0.5161 0.8077 0.8987
No log 31.7143 222 0.8093 0.4985 0.8093 0.8996
No log 32.0 224 0.8044 0.5149 0.8044 0.8969
No log 32.2857 226 0.8010 0.5110 0.8010 0.8950
No log 32.5714 228 0.8017 0.5048 0.8017 0.8954
No log 32.8571 230 0.8020 0.5138 0.8020 0.8956
No log 33.1429 232 0.8074 0.5138 0.8074 0.8986
No log 33.4286 234 0.8000 0.5258 0.8000 0.8944
No log 33.7143 236 0.7845 0.5354 0.7845 0.8857
No log 34.0 238 0.7821 0.5343 0.7821 0.8844
No log 34.2857 240 0.7831 0.5436 0.7831 0.8849
No log 34.5714 242 0.7752 0.5252 0.7752 0.8804
No log 34.8571 244 0.7833 0.5354 0.7833 0.8851
No log 35.1429 246 0.7988 0.5197 0.7988 0.8937
No log 35.4286 248 0.8122 0.5183 0.8122 0.9012
No log 35.7143 250 0.8184 0.5080 0.8184 0.9047
No log 36.0 252 0.8302 0.5080 0.8302 0.9111
No log 36.2857 254 0.8414 0.5127 0.8414 0.9173
No log 36.5714 256 0.8512 0.5122 0.8512 0.9226
No log 36.8571 258 0.8519 0.5074 0.8519 0.9230
No log 37.1429 260 0.8511 0.5060 0.8511 0.9226
No log 37.4286 262 0.8368 0.5012 0.8368 0.9148
No log 37.7143 264 0.8220 0.4980 0.8220 0.9067
No log 38.0 266 0.8077 0.5186 0.8077 0.8987
No log 38.2857 268 0.7989 0.5144 0.7989 0.8938
No log 38.5714 270 0.7941 0.5283 0.7941 0.8911
No log 38.8571 272 0.7946 0.5176 0.7946 0.8914
No log 39.1429 274 0.7952 0.5026 0.7952 0.8917
No log 39.4286 276 0.7938 0.5237 0.7938 0.8910
No log 39.7143 278 0.7971 0.5138 0.7971 0.8928
No log 40.0 280 0.8026 0.5153 0.8026 0.8959
No log 40.2857 282 0.8054 0.5012 0.8054 0.8974
No log 40.5714 284 0.8206 0.5052 0.8206 0.9059
No log 40.8571 286 0.8305 0.5103 0.8305 0.9113
No log 41.1429 288 0.8274 0.5018 0.8274 0.9096
No log 41.4286 290 0.8243 0.4987 0.8243 0.9079
No log 41.7143 292 0.8259 0.5186 0.8259 0.9088
No log 42.0 294 0.8317 0.5246 0.8317 0.9120
No log 42.2857 296 0.8346 0.5104 0.8346 0.9135
No log 42.5714 298 0.8364 0.5104 0.8364 0.9145
No log 42.8571 300 0.8387 0.5090 0.8387 0.9158
No log 43.1429 302 0.8555 0.5250 0.8555 0.9249
No log 43.4286 304 0.8595 0.5182 0.8595 0.9271
No log 43.7143 306 0.8462 0.5351 0.8462 0.9199
No log 44.0 308 0.8195 0.5122 0.8195 0.9052
No log 44.2857 310 0.7924 0.4952 0.7924 0.8901
No log 44.5714 312 0.7853 0.5070 0.7853 0.8862
No log 44.8571 314 0.7832 0.4966 0.7832 0.8850
No log 45.1429 316 0.7883 0.5 0.7883 0.8879
No log 45.4286 318 0.7978 0.4993 0.7978 0.8932
No log 45.7143 320 0.8078 0.5118 0.8078 0.8988
No log 46.0 322 0.8157 0.5023 0.8157 0.9032
No log 46.2857 324 0.8245 0.5245 0.8245 0.9080
No log 46.5714 326 0.8324 0.5216 0.8324 0.9123
No log 46.8571 328 0.8459 0.5042 0.8459 0.9197
No log 47.1429 330 0.8525 0.4890 0.8525 0.9233
No log 47.4286 332 0.8452 0.5026 0.8452 0.9194
No log 47.7143 334 0.8355 0.4881 0.8355 0.9140
No log 48.0 336 0.8332 0.4917 0.8332 0.9128
No log 48.2857 338 0.8491 0.5060 0.8491 0.9214
No log 48.5714 340 0.8708 0.5284 0.8708 0.9332
No log 48.8571 342 0.8643 0.5307 0.8643 0.9297
No log 49.1429 344 0.8444 0.5116 0.8444 0.9189
No log 49.4286 346 0.8302 0.4978 0.8302 0.9112
No log 49.7143 348 0.8164 0.4938 0.8164 0.9035
No log 50.0 350 0.8044 0.4884 0.8044 0.8969
No log 50.2857 352 0.7977 0.4946 0.7977 0.8932
No log 50.5714 354 0.7988 0.4946 0.7988 0.8937
No log 50.8571 356 0.8099 0.4920 0.8099 0.8999
No log 51.1429 358 0.8318 0.5262 0.8318 0.9121
No log 51.4286 360 0.8538 0.5237 0.8538 0.9240
No log 51.7143 362 0.8498 0.5208 0.8498 0.9219
No log 52.0 364 0.8338 0.5161 0.8338 0.9131
No log 52.2857 366 0.8179 0.4957 0.8179 0.9044
No log 52.5714 368 0.7939 0.4923 0.7939 0.8910
No log 52.8571 370 0.7833 0.5279 0.7833 0.8851
No log 53.1429 372 0.7862 0.5251 0.7862 0.8867
No log 53.4286 374 0.7891 0.5289 0.7891 0.8883
No log 53.7143 376 0.7947 0.5047 0.7947 0.8915
No log 54.0 378 0.8059 0.5135 0.8059 0.8977
No log 54.2857 380 0.8056 0.5075 0.8056 0.8976
No log 54.5714 382 0.7955 0.5027 0.7955 0.8919
No log 54.8571 384 0.7891 0.5080 0.7891 0.8883
No log 55.1429 386 0.7817 0.5088 0.7817 0.8841
No log 55.4286 388 0.7844 0.4958 0.7844 0.8856
No log 55.7143 390 0.7855 0.5174 0.7855 0.8863
No log 56.0 392 0.7858 0.4957 0.7858 0.8865
No log 56.2857 394 0.7867 0.5075 0.7867 0.8870
No log 56.5714 396 0.7816 0.5075 0.7816 0.8841
No log 56.8571 398 0.7862 0.5027 0.7862 0.8867
No log 57.1429 400 0.7895 0.5034 0.7895 0.8886
No log 57.4286 402 0.7927 0.5082 0.7927 0.8903
No log 57.7143 404 0.7984 0.5053 0.7984 0.8935
No log 58.0 406 0.8086 0.5088 0.8086 0.8992
No log 58.2857 408 0.8143 0.5072 0.8143 0.9024
No log 58.5714 410 0.8196 0.5217 0.8196 0.9053
No log 58.8571 412 0.8369 0.5304 0.8369 0.9148
No log 59.1429 414 0.8580 0.5277 0.8580 0.9263
No log 59.4286 416 0.8552 0.5258 0.8552 0.9248
No log 59.7143 418 0.8381 0.5238 0.8381 0.9155
No log 60.0 420 0.8102 0.5208 0.8102 0.9001
No log 60.2857 422 0.7893 0.4979 0.7893 0.8884
No log 60.5714 424 0.7812 0.5163 0.7812 0.8839
No log 60.8571 426 0.7744 0.5329 0.7744 0.8800
No log 61.1429 428 0.7655 0.5253 0.7655 0.8749
No log 61.4286 430 0.7681 0.5232 0.7681 0.8764
No log 61.7143 432 0.7773 0.5204 0.7773 0.8816
No log 62.0 434 0.7767 0.5204 0.7767 0.8813
No log 62.2857 436 0.7667 0.5286 0.7667 0.8756
No log 62.5714 438 0.7610 0.5375 0.7610 0.8724
No log 62.8571 440 0.7570 0.5254 0.7570 0.8700
No log 63.1429 442 0.7523 0.5306 0.7523 0.8674
No log 63.4286 444 0.7500 0.5306 0.7500 0.8660
No log 63.7143 446 0.7486 0.5450 0.7486 0.8652
No log 64.0 448 0.7470 0.5450 0.7470 0.8643
No log 64.2857 450 0.7477 0.5279 0.7477 0.8647
No log 64.5714 452 0.7517 0.5485 0.7517 0.8670
No log 64.8571 454 0.7655 0.5396 0.7655 0.8749
No log 65.1429 456 0.7725 0.5396 0.7725 0.8789
No log 65.4286 458 0.7724 0.5346 0.7724 0.8789
No log 65.7143 460 0.7731 0.5346 0.7731 0.8793
No log 66.0 462 0.7695 0.5422 0.7695 0.8772
No log 66.2857 464 0.7655 0.5124 0.7655 0.8749
No log 66.5714 466 0.7630 0.5124 0.7630 0.8735
No log 66.8571 468 0.7616 0.4972 0.7616 0.8727
No log 67.1429 470 0.7632 0.4972 0.7632 0.8736
No log 67.4286 472 0.7714 0.5173 0.7714 0.8783
No log 67.7143 474 0.7813 0.5194 0.7813 0.8839
No log 68.0 476 0.7860 0.5185 0.7860 0.8866
No log 68.2857 478 0.7898 0.5307 0.7898 0.8887
No log 68.5714 480 0.7814 0.5362 0.7814 0.8840
No log 68.8571 482 0.7634 0.5222 0.7634 0.8737
No log 69.1429 484 0.7565 0.5173 0.7565 0.8698
No log 69.4286 486 0.7546 0.5173 0.7546 0.8687
No log 69.7143 488 0.7500 0.5351 0.7500 0.8660
No log 70.0 490 0.7488 0.5351 0.7488 0.8653
No log 70.2857 492 0.7493 0.5275 0.7493 0.8656
No log 70.5714 494 0.7498 0.5275 0.7498 0.8659
No log 70.8571 496 0.7548 0.5318 0.7548 0.8688
No log 71.1429 498 0.7617 0.5545 0.7617 0.8727
0.5217 71.4286 500 0.7702 0.5520 0.7702 0.8776
0.5217 71.7143 502 0.7700 0.5692 0.7700 0.8775
0.5217 72.0 504 0.7567 0.5655 0.7567 0.8699
0.5217 72.2857 506 0.7380 0.5660 0.7380 0.8591
0.5217 72.5714 508 0.7307 0.5242 0.7307 0.8548
0.5217 72.8571 510 0.7290 0.5254 0.7290 0.8538
0.5217 73.1429 512 0.7294 0.5254 0.7294 0.8540
0.5217 73.4286 514 0.7315 0.5299 0.7315 0.8553
0.5217 73.7143 516 0.7379 0.5348 0.7379 0.8590

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.1+cu121
  • Datasets 3.2.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERTFreeze_run2_AugV5_k2_task2_organization

Finetuned
(4023)
this model