ArabicNewSplits6_FineTuningAraBERTFreeze_run3_AugV5_k5_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7561
  • Qwk: 0.5557
  • Mse: 0.7561
  • Rmse: 0.8696

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1538 2 6.3810 -0.0278 6.3810 2.5261
No log 0.3077 4 4.3012 -0.0269 4.3012 2.0739
No log 0.4615 6 3.0577 0.0124 3.0577 1.7486
No log 0.6154 8 2.1623 0.0108 2.1623 1.4705
No log 0.7692 10 1.5637 -0.0098 1.5637 1.2505
No log 0.9231 12 1.2846 -0.0441 1.2846 1.1334
No log 1.0769 14 0.9397 0.1197 0.9397 0.9694
No log 1.2308 16 0.7299 0.1170 0.7299 0.8543
No log 1.3846 18 0.6524 0.2893 0.6524 0.8077
No log 1.5385 20 0.6328 0.3575 0.6328 0.7955
No log 1.6923 22 0.6360 0.3700 0.6360 0.7975
No log 1.8462 24 0.6593 0.3199 0.6593 0.8120
No log 2.0 26 0.6481 0.3320 0.6481 0.8051
No log 2.1538 28 0.6343 0.3611 0.6343 0.7964
No log 2.3077 30 0.6101 0.4079 0.6101 0.7811
No log 2.4615 32 0.6114 0.3542 0.6114 0.7819
No log 2.6154 34 0.6621 0.3094 0.6621 0.8137
No log 2.7692 36 0.7892 0.2059 0.7892 0.8884
No log 2.9231 38 0.9167 0.1174 0.9167 0.9574
No log 3.0769 40 0.9043 0.1811 0.9043 0.9509
No log 3.2308 42 0.8662 0.1988 0.8662 0.9307
No log 3.3846 44 0.9458 0.2315 0.9458 0.9725
No log 3.5385 46 0.9742 0.2739 0.9742 0.9870
No log 3.6923 48 0.8779 0.2381 0.8779 0.9370
No log 3.8462 50 0.7315 0.2593 0.7315 0.8553
No log 4.0 52 0.6227 0.2894 0.6227 0.7891
No log 4.1538 54 0.5763 0.3871 0.5763 0.7591
No log 4.3077 56 0.5694 0.3964 0.5694 0.7546
No log 4.4615 58 0.5776 0.4111 0.5776 0.7600
No log 4.6154 60 0.5981 0.3628 0.5981 0.7734
No log 4.7692 62 0.6236 0.3494 0.6236 0.7897
No log 4.9231 64 0.6652 0.3799 0.6652 0.8156
No log 5.0769 66 0.6912 0.3207 0.6912 0.8314
No log 5.2308 68 0.6922 0.3711 0.6922 0.8320
No log 5.3846 70 0.6520 0.4119 0.6520 0.8075
No log 5.5385 72 0.5984 0.4241 0.5984 0.7735
No log 5.6923 74 0.5833 0.4399 0.5833 0.7637
No log 5.8462 76 0.5640 0.4620 0.5640 0.7510
No log 6.0 78 0.5549 0.4482 0.5549 0.7449
No log 6.1538 80 0.5499 0.4695 0.5499 0.7416
No log 6.3077 82 0.5602 0.4714 0.5602 0.7484
No log 6.4615 84 0.5539 0.4885 0.5539 0.7443
No log 6.6154 86 0.5781 0.4639 0.5781 0.7604
No log 6.7692 88 0.5891 0.4639 0.5891 0.7675
No log 6.9231 90 0.5493 0.5004 0.5493 0.7412
No log 7.0769 92 0.5142 0.5287 0.5142 0.7171
No log 7.2308 94 0.5042 0.5230 0.5042 0.7100
No log 7.3846 96 0.5040 0.5315 0.5040 0.7099
No log 7.5385 98 0.5117 0.5516 0.5117 0.7154
No log 7.6923 100 0.5150 0.5518 0.5150 0.7176
No log 7.8462 102 0.5182 0.5589 0.5182 0.7198
No log 8.0 104 0.5435 0.5469 0.5435 0.7372
No log 8.1538 106 0.6016 0.5253 0.6016 0.7756
No log 8.3077 108 0.6418 0.5212 0.6418 0.8011
No log 8.4615 110 0.6141 0.5253 0.6141 0.7837
No log 8.6154 112 0.5486 0.5606 0.5486 0.7407
No log 8.7692 114 0.5263 0.5571 0.5263 0.7254
No log 8.9231 116 0.5199 0.5602 0.5199 0.7210
No log 9.0769 118 0.5141 0.5436 0.5141 0.7170
No log 9.2308 120 0.5158 0.5334 0.5158 0.7182
No log 9.3846 122 0.5165 0.5585 0.5165 0.7187
No log 9.5385 124 0.5550 0.5659 0.5550 0.7450
No log 9.6923 126 0.6120 0.4246 0.6120 0.7823
No log 9.8462 128 0.6165 0.4246 0.6165 0.7852
No log 10.0 130 0.5525 0.5556 0.5525 0.7433
No log 10.1538 132 0.5217 0.5436 0.5217 0.7223
No log 10.3077 134 0.5410 0.4821 0.5410 0.7355
No log 10.4615 136 0.5469 0.4944 0.5469 0.7395
No log 10.6154 138 0.5350 0.4974 0.5350 0.7314
No log 10.7692 140 0.5353 0.5621 0.5353 0.7317
No log 10.9231 142 0.5989 0.5228 0.5989 0.7739
No log 11.0769 144 0.6934 0.4478 0.6934 0.8327
No log 11.2308 146 0.7029 0.4478 0.7029 0.8384
No log 11.3846 148 0.6348 0.4596 0.6348 0.7968
No log 11.5385 150 0.5601 0.5704 0.5601 0.7484
No log 11.6923 152 0.5536 0.5019 0.5536 0.7440
No log 11.8462 154 0.5734 0.4739 0.5734 0.7572
No log 12.0 156 0.5737 0.5241 0.5737 0.7574
No log 12.1538 158 0.5609 0.5450 0.5609 0.7489
No log 12.3077 160 0.5769 0.5639 0.5769 0.7596
No log 12.4615 162 0.6247 0.5189 0.6247 0.7904
No log 12.6154 164 0.6202 0.5258 0.6202 0.7876
No log 12.7692 166 0.5949 0.5589 0.5949 0.7713
No log 12.9231 168 0.5878 0.5346 0.5878 0.7667
No log 13.0769 170 0.5925 0.5611 0.5925 0.7697
No log 13.2308 172 0.5971 0.5975 0.5971 0.7727
No log 13.3846 174 0.5985 0.6034 0.5985 0.7736
No log 13.5385 176 0.6016 0.5969 0.6016 0.7756
No log 13.6923 178 0.6014 0.5767 0.6014 0.7755
No log 13.8462 180 0.5998 0.6008 0.5998 0.7744
No log 14.0 182 0.6053 0.5925 0.6053 0.7780
No log 14.1538 184 0.6087 0.5925 0.6087 0.7802
No log 14.3077 186 0.6062 0.5773 0.6062 0.7786
No log 14.4615 188 0.6204 0.5859 0.6204 0.7876
No log 14.6154 190 0.6455 0.5461 0.6455 0.8034
No log 14.7692 192 0.6622 0.5433 0.6622 0.8137
No log 14.9231 194 0.6532 0.5106 0.6532 0.8082
No log 15.0769 196 0.6540 0.5348 0.6540 0.8087
No log 15.2308 198 0.6568 0.5276 0.6568 0.8105
No log 15.3846 200 0.6616 0.5722 0.6616 0.8134
No log 15.5385 202 0.6744 0.5647 0.6744 0.8212
No log 15.6923 204 0.6866 0.5421 0.6866 0.8286
No log 15.8462 206 0.6938 0.5673 0.6938 0.8329
No log 16.0 208 0.7194 0.5783 0.7194 0.8482
No log 16.1538 210 0.7367 0.5518 0.7367 0.8583
No log 16.3077 212 0.7319 0.5519 0.7319 0.8555
No log 16.4615 214 0.6944 0.5547 0.6944 0.8333
No log 16.6154 216 0.6864 0.5292 0.6864 0.8285
No log 16.7692 218 0.6943 0.5462 0.6943 0.8332
No log 16.9231 220 0.6772 0.5348 0.6772 0.8229
No log 17.0769 222 0.6867 0.5313 0.6867 0.8286
No log 17.2308 224 0.6923 0.5222 0.6923 0.8320
No log 17.3846 226 0.6801 0.5325 0.6801 0.8247
No log 17.5385 228 0.6760 0.5267 0.6760 0.8222
No log 17.6923 230 0.6782 0.5405 0.6782 0.8235
No log 17.8462 232 0.7003 0.5480 0.7003 0.8369
No log 18.0 234 0.6956 0.5561 0.6956 0.8341
No log 18.1538 236 0.6825 0.5672 0.6825 0.8261
No log 18.3077 238 0.6754 0.5410 0.6754 0.8219
No log 18.4615 240 0.6796 0.5550 0.6796 0.8244
No log 18.6154 242 0.7063 0.5717 0.7063 0.8404
No log 18.7692 244 0.7601 0.5454 0.7601 0.8718
No log 18.9231 246 0.7717 0.5416 0.7717 0.8785
No log 19.0769 248 0.7347 0.5558 0.7347 0.8572
No log 19.2308 250 0.7059 0.5297 0.7059 0.8402
No log 19.3846 252 0.7005 0.5334 0.7005 0.8370
No log 19.5385 254 0.6972 0.5232 0.6972 0.8350
No log 19.6923 256 0.6978 0.5465 0.6978 0.8353
No log 19.8462 258 0.6987 0.5428 0.6987 0.8359
No log 20.0 260 0.7097 0.5692 0.7097 0.8425
No log 20.1538 262 0.7192 0.5467 0.7192 0.8480
No log 20.3077 264 0.7277 0.5283 0.7277 0.8531
No log 20.4615 266 0.7376 0.5247 0.7376 0.8589
No log 20.6154 268 0.7435 0.5280 0.7435 0.8623
No log 20.7692 270 0.7432 0.5281 0.7432 0.8621
No log 20.9231 272 0.7313 0.5434 0.7313 0.8551
No log 21.0769 274 0.7349 0.54 0.7349 0.8572
No log 21.2308 276 0.7312 0.54 0.7312 0.8551
No log 21.3846 278 0.7186 0.5348 0.7186 0.8477
No log 21.5385 280 0.7179 0.5741 0.7179 0.8473
No log 21.6923 282 0.7061 0.5750 0.7061 0.8403
No log 21.8462 284 0.7273 0.5494 0.7273 0.8528
No log 22.0 286 0.7849 0.5530 0.7849 0.8860
No log 22.1538 288 0.7909 0.5271 0.7909 0.8893
No log 22.3077 290 0.7512 0.5467 0.7512 0.8667
No log 22.4615 292 0.7047 0.5303 0.7047 0.8394
No log 22.6154 294 0.6900 0.5588 0.6900 0.8307
No log 22.7692 296 0.7026 0.5620 0.7026 0.8382
No log 22.9231 298 0.7195 0.5706 0.7195 0.8482
No log 23.0769 300 0.7167 0.5790 0.7167 0.8466
No log 23.2308 302 0.7298 0.5325 0.7298 0.8543
No log 23.3846 304 0.7583 0.5208 0.7583 0.8708
No log 23.5385 306 0.7796 0.5221 0.7796 0.8829
No log 23.6923 308 0.7579 0.5056 0.7579 0.8706
No log 23.8462 310 0.7116 0.5206 0.7116 0.8436
No log 24.0 312 0.6882 0.5187 0.6882 0.8296
No log 24.1538 314 0.6741 0.5405 0.6741 0.8210
No log 24.3077 316 0.6715 0.5595 0.6715 0.8194
No log 24.4615 318 0.6712 0.5890 0.6712 0.8193
No log 24.6154 320 0.6583 0.5669 0.6583 0.8114
No log 24.7692 322 0.6623 0.5317 0.6623 0.8138
No log 24.9231 324 0.6795 0.5506 0.6795 0.8243
No log 25.0769 326 0.6759 0.5519 0.6759 0.8221
No log 25.2308 328 0.6681 0.5614 0.6681 0.8174
No log 25.3846 330 0.6633 0.5591 0.6633 0.8144
No log 25.5385 332 0.6799 0.5821 0.6799 0.8245
No log 25.6923 334 0.6945 0.5327 0.6945 0.8334
No log 25.8462 336 0.6874 0.5761 0.6874 0.8291
No log 26.0 338 0.6933 0.5449 0.6933 0.8326
No log 26.1538 340 0.7222 0.5493 0.7222 0.8498
No log 26.3077 342 0.7168 0.5442 0.7168 0.8466
No log 26.4615 344 0.6999 0.5548 0.6999 0.8366
No log 26.6154 346 0.6975 0.5703 0.6975 0.8352
No log 26.7692 348 0.6920 0.5405 0.6920 0.8318
No log 26.9231 350 0.6887 0.5496 0.6887 0.8299
No log 27.0769 352 0.6839 0.5717 0.6839 0.8270
No log 27.2308 354 0.6927 0.5773 0.6927 0.8323
No log 27.3846 356 0.7152 0.5603 0.7152 0.8457
No log 27.5385 358 0.7543 0.5261 0.7543 0.8685
No log 27.6923 360 0.7579 0.5169 0.7579 0.8706
No log 27.8462 362 0.7457 0.5352 0.7457 0.8635
No log 28.0 364 0.7419 0.5364 0.7419 0.8613
No log 28.1538 366 0.7425 0.5494 0.7425 0.8617
No log 28.3077 368 0.7504 0.5290 0.7504 0.8663
No log 28.4615 370 0.7576 0.5272 0.7576 0.8704
No log 28.6154 372 0.7627 0.5201 0.7627 0.8733
No log 28.7692 374 0.7794 0.5360 0.7794 0.8829
No log 28.9231 376 0.7771 0.5281 0.7771 0.8815
No log 29.0769 378 0.7567 0.5311 0.7567 0.8699
No log 29.2308 380 0.7501 0.5314 0.7501 0.8661
No log 29.3846 382 0.7436 0.5220 0.7436 0.8623
No log 29.5385 384 0.7363 0.5314 0.7363 0.8581
No log 29.6923 386 0.7414 0.5561 0.7414 0.8610
No log 29.8462 388 0.7436 0.5329 0.7436 0.8624
No log 30.0 390 0.7444 0.5297 0.7444 0.8628
No log 30.1538 392 0.7566 0.5388 0.7566 0.8698
No log 30.3077 394 0.7658 0.5374 0.7658 0.8751
No log 30.4615 396 0.7825 0.5164 0.7825 0.8846
No log 30.6154 398 0.7766 0.5374 0.7766 0.8812
No log 30.7692 400 0.7827 0.5297 0.7827 0.8847
No log 30.9231 402 0.7808 0.5493 0.7808 0.8836
No log 31.0769 404 0.7916 0.5333 0.7916 0.8897
No log 31.2308 406 0.8150 0.5055 0.8150 0.9028
No log 31.3846 408 0.8213 0.5361 0.8213 0.9063
No log 31.5385 410 0.7996 0.5192 0.7996 0.8942
No log 31.6923 412 0.7901 0.5409 0.7901 0.8889
No log 31.8462 414 0.7846 0.5288 0.7846 0.8858
No log 32.0 416 0.7849 0.5536 0.7849 0.8860
No log 32.1538 418 0.7986 0.5201 0.7986 0.8936
No log 32.3077 420 0.8213 0.5653 0.8213 0.9062
No log 32.4615 422 0.8141 0.5653 0.8141 0.9023
No log 32.6154 424 0.7836 0.5398 0.7836 0.8852
No log 32.7692 426 0.7591 0.5779 0.7591 0.8713
No log 32.9231 428 0.7572 0.5732 0.7572 0.8702
No log 33.0769 430 0.7802 0.5416 0.7802 0.8833
No log 33.2308 432 0.8037 0.5333 0.8037 0.8965
No log 33.3846 434 0.8187 0.5267 0.8187 0.9048
No log 33.5385 436 0.7945 0.5430 0.7945 0.8914
No log 33.6923 438 0.7569 0.5736 0.7569 0.8700
No log 33.8462 440 0.7381 0.5305 0.7381 0.8591
No log 34.0 442 0.7356 0.5708 0.7356 0.8577
No log 34.1538 444 0.7469 0.5534 0.7469 0.8642
No log 34.3077 446 0.7526 0.5534 0.7526 0.8675
No log 34.4615 448 0.7536 0.5623 0.7536 0.8681
No log 34.6154 450 0.7670 0.5509 0.7670 0.8758
No log 34.7692 452 0.7588 0.5258 0.7588 0.8711
No log 34.9231 454 0.7609 0.5441 0.7609 0.8723
No log 35.0769 456 0.7698 0.5388 0.7698 0.8774
No log 35.2308 458 0.7753 0.5429 0.7753 0.8805
No log 35.3846 460 0.7756 0.5100 0.7756 0.8807
No log 35.5385 462 0.7626 0.5820 0.7626 0.8733
No log 35.6923 464 0.7301 0.5762 0.7301 0.8545
No log 35.8462 466 0.7105 0.5715 0.7105 0.8429
No log 36.0 468 0.6933 0.5615 0.6933 0.8326
No log 36.1538 470 0.6913 0.5296 0.6913 0.8314
No log 36.3077 472 0.6819 0.5402 0.6819 0.8258
No log 36.4615 474 0.6791 0.5694 0.6791 0.8241
No log 36.6154 476 0.6960 0.5635 0.6960 0.8342
No log 36.7692 478 0.7343 0.5505 0.7343 0.8569
No log 36.9231 480 0.7534 0.5517 0.7534 0.8680
No log 37.0769 482 0.7452 0.5180 0.7452 0.8632
No log 37.2308 484 0.7438 0.5113 0.7438 0.8625
No log 37.3846 486 0.7474 0.4940 0.7474 0.8645
No log 37.5385 488 0.7435 0.5133 0.7435 0.8622
No log 37.6923 490 0.7539 0.5103 0.7539 0.8683
No log 37.8462 492 0.7832 0.5131 0.7832 0.8850
No log 38.0 494 0.7734 0.5225 0.7734 0.8794
No log 38.1538 496 0.7414 0.5440 0.7414 0.8611
No log 38.3077 498 0.7082 0.5598 0.7082 0.8416
0.5566 38.4615 500 0.6908 0.5734 0.6908 0.8311
0.5566 38.6154 502 0.6938 0.5592 0.6938 0.8330
0.5566 38.7692 504 0.7070 0.5423 0.7070 0.8408
0.5566 38.9231 506 0.7207 0.5395 0.7207 0.8489
0.5566 39.0769 508 0.7366 0.5600 0.7366 0.8583
0.5566 39.2308 510 0.7561 0.5557 0.7561 0.8696

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.1+cu121
  • Datasets 3.2.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERTFreeze_run3_AugV5_k5_task2_organization

Finetuned
(4023)
this model