ArabicNewSplits6_FineTuningAraBERTFreeze_run2_AugV5_k15_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6820
  • Qwk: 0.4819
  • Mse: 0.6820
  • Rmse: 0.8259

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0526 2 6.4968 -0.0278 6.4968 2.5489
No log 0.1053 4 4.4906 -0.0328 4.4906 2.1191
No log 0.1579 6 3.1975 -0.0105 3.1975 1.7882
No log 0.2105 8 2.3650 -0.0187 2.3650 1.5378
No log 0.2632 10 1.8921 -0.0394 1.8921 1.3755
No log 0.3158 12 1.7177 -0.0665 1.7177 1.3106
No log 0.3684 14 1.5085 -0.0691 1.5085 1.2282
No log 0.4211 16 1.5939 -0.0988 1.5939 1.2625
No log 0.4737 18 1.4571 -0.0908 1.4571 1.2071
No log 0.5263 20 1.3552 -0.0236 1.3552 1.1641
No log 0.5789 22 1.0656 0.0049 1.0656 1.0323
No log 0.6316 24 0.8282 0.0482 0.8282 0.9100
No log 0.6842 26 0.7647 0.2010 0.7647 0.8745
No log 0.7368 28 0.7350 0.1972 0.7350 0.8574
No log 0.7895 30 0.7154 0.2292 0.7154 0.8458
No log 0.8421 32 0.7510 0.1764 0.7510 0.8666
No log 0.8947 34 0.8633 0.1515 0.8633 0.9291
No log 0.9474 36 0.9880 0.0800 0.9880 0.9940
No log 1.0 38 1.0078 0.0900 1.0078 1.0039
No log 1.0526 40 0.9182 0.1727 0.9182 0.9582
No log 1.1053 42 0.7724 0.2210 0.7724 0.8789
No log 1.1579 44 0.6919 0.2524 0.6919 0.8318
No log 1.2105 46 0.6612 0.2939 0.6612 0.8131
No log 1.2632 48 0.6668 0.2812 0.6668 0.8166
No log 1.3158 50 0.6970 0.2137 0.6970 0.8349
No log 1.3684 52 0.7246 0.2396 0.7246 0.8512
No log 1.4211 54 0.7415 0.2134 0.7415 0.8611
No log 1.4737 56 0.8050 0.2539 0.8050 0.8972
No log 1.5263 58 0.8722 0.3122 0.8722 0.9339
No log 1.5789 60 0.9897 0.2665 0.9897 0.9948
No log 1.6316 62 1.0380 0.2426 1.0380 1.0188
No log 1.6842 64 0.9020 0.2737 0.9020 0.9498
No log 1.7368 66 0.7537 0.3332 0.7537 0.8682
No log 1.7895 68 0.6702 0.3912 0.6702 0.8187
No log 1.8421 70 0.6298 0.3681 0.6298 0.7936
No log 1.8947 72 0.6456 0.3833 0.6456 0.8035
No log 1.9474 74 0.7088 0.3714 0.7088 0.8419
No log 2.0 76 0.7695 0.3813 0.7695 0.8772
No log 2.0526 78 0.8231 0.4050 0.8231 0.9073
No log 2.1053 80 0.8233 0.4055 0.8233 0.9074
No log 2.1579 82 0.7863 0.4214 0.7863 0.8867
No log 2.2105 84 0.7123 0.4133 0.7123 0.8440
No log 2.2632 86 0.6911 0.4484 0.6911 0.8313
No log 2.3158 88 0.6827 0.4295 0.6827 0.8262
No log 2.3684 90 0.7457 0.4212 0.7457 0.8636
No log 2.4211 92 0.8325 0.3688 0.8325 0.9124
No log 2.4737 94 0.9415 0.3377 0.9415 0.9703
No log 2.5263 96 0.9484 0.3356 0.9484 0.9739
No log 2.5789 98 0.8384 0.3994 0.8384 0.9156
No log 2.6316 100 0.7044 0.4339 0.7044 0.8393
No log 2.6842 102 0.6184 0.4825 0.6184 0.7864
No log 2.7368 104 0.6095 0.4957 0.6095 0.7807
No log 2.7895 106 0.6281 0.5215 0.6281 0.7926
No log 2.8421 108 0.6347 0.4572 0.6347 0.7967
No log 2.8947 110 0.6697 0.4811 0.6697 0.8183
No log 2.9474 112 0.6992 0.4549 0.6992 0.8362
No log 3.0 114 0.6532 0.5031 0.6532 0.8082
No log 3.0526 116 0.6484 0.5202 0.6484 0.8052
No log 3.1053 118 0.6555 0.5193 0.6555 0.8096
No log 3.1579 120 0.6638 0.4997 0.6638 0.8147
No log 3.2105 122 0.6646 0.4934 0.6646 0.8152
No log 3.2632 124 0.7030 0.4852 0.7030 0.8384
No log 3.3158 126 0.7498 0.4881 0.7498 0.8659
No log 3.3684 128 0.7894 0.4505 0.7894 0.8885
No log 3.4211 130 0.7284 0.4973 0.7284 0.8535
No log 3.4737 132 0.6303 0.5340 0.6303 0.7939
No log 3.5263 134 0.5844 0.5574 0.5844 0.7645
No log 3.5789 136 0.5734 0.5682 0.5734 0.7572
No log 3.6316 138 0.6137 0.5097 0.6137 0.7834
No log 3.6842 140 0.6318 0.5337 0.6318 0.7949
No log 3.7368 142 0.6264 0.5238 0.6264 0.7915
No log 3.7895 144 0.6103 0.5424 0.6103 0.7812
No log 3.8421 146 0.5814 0.5505 0.5814 0.7625
No log 3.8947 148 0.5616 0.5868 0.5616 0.7494
No log 3.9474 150 0.5849 0.5592 0.5849 0.7648
No log 4.0 152 0.6196 0.5147 0.6196 0.7872
No log 4.0526 154 0.6161 0.5416 0.6161 0.7849
No log 4.1053 156 0.6505 0.4938 0.6505 0.8066
No log 4.1579 158 0.6571 0.4892 0.6571 0.8106
No log 4.2105 160 0.6811 0.4769 0.6811 0.8253
No log 4.2632 162 0.6331 0.5193 0.6331 0.7957
No log 4.3158 164 0.5733 0.5369 0.5733 0.7572
No log 4.3684 166 0.5381 0.5486 0.5381 0.7335
No log 4.4211 168 0.5273 0.5085 0.5273 0.7261
No log 4.4737 170 0.5201 0.5553 0.5201 0.7212
No log 4.5263 172 0.5170 0.5015 0.5170 0.7191
No log 4.5789 174 0.5244 0.5625 0.5244 0.7241
No log 4.6316 176 0.5355 0.5684 0.5355 0.7318
No log 4.6842 178 0.5622 0.5385 0.5622 0.7498
No log 4.7368 180 0.5936 0.5540 0.5936 0.7704
No log 4.7895 182 0.6092 0.5539 0.6092 0.7805
No log 4.8421 184 0.6109 0.5686 0.6109 0.7816
No log 4.8947 186 0.6351 0.5263 0.6351 0.7969
No log 4.9474 188 0.6731 0.4786 0.6731 0.8204
No log 5.0 190 0.7240 0.4449 0.7240 0.8509
No log 5.0526 192 0.7032 0.4872 0.7032 0.8386
No log 5.1053 194 0.7019 0.4690 0.7019 0.8378
No log 5.1579 196 0.6927 0.4680 0.6927 0.8323
No log 5.2105 198 0.6560 0.5071 0.6560 0.8099
No log 5.2632 200 0.6345 0.4926 0.6345 0.7966
No log 5.3158 202 0.6378 0.4946 0.6378 0.7986
No log 5.3684 204 0.6469 0.5229 0.6469 0.8043
No log 5.4211 206 0.6612 0.4753 0.6612 0.8132
No log 5.4737 208 0.6817 0.4911 0.6817 0.8256
No log 5.5263 210 0.6963 0.5142 0.6963 0.8344
No log 5.5789 212 0.6924 0.5276 0.6924 0.8321
No log 5.6316 214 0.6389 0.4908 0.6389 0.7993
No log 5.6842 216 0.5988 0.4790 0.5988 0.7738
No log 5.7368 218 0.5947 0.4933 0.5947 0.7712
No log 5.7895 220 0.6077 0.4587 0.6077 0.7796
No log 5.8421 222 0.6113 0.4202 0.6113 0.7819
No log 5.8947 224 0.5921 0.4666 0.5921 0.7695
No log 5.9474 226 0.5897 0.4823 0.5897 0.7679
No log 6.0 228 0.6057 0.5208 0.6057 0.7783
No log 6.0526 230 0.6151 0.4987 0.6151 0.7843
No log 6.1053 232 0.6181 0.5133 0.6181 0.7862
No log 6.1579 234 0.6333 0.5205 0.6333 0.7958
No log 6.2105 236 0.6724 0.5408 0.6724 0.8200
No log 6.2632 238 0.6647 0.5186 0.6647 0.8153
No log 6.3158 240 0.6556 0.5295 0.6556 0.8097
No log 6.3684 242 0.6440 0.5064 0.6440 0.8025
No log 6.4211 244 0.6484 0.5236 0.6484 0.8052
No log 6.4737 246 0.6633 0.5300 0.6633 0.8144
No log 6.5263 248 0.6724 0.5264 0.6724 0.8200
No log 6.5789 250 0.6670 0.5577 0.6670 0.8167
No log 6.6316 252 0.6621 0.5438 0.6621 0.8137
No log 6.6842 254 0.6611 0.5314 0.6611 0.8131
No log 6.7368 256 0.6541 0.5521 0.6541 0.8088
No log 6.7895 258 0.6463 0.5430 0.6463 0.8039
No log 6.8421 260 0.6408 0.5674 0.6408 0.8005
No log 6.8947 262 0.6478 0.5593 0.6478 0.8049
No log 6.9474 264 0.6607 0.5124 0.6607 0.8128
No log 7.0 266 0.6542 0.5217 0.6542 0.8088
No log 7.0526 268 0.6453 0.5136 0.6453 0.8033
No log 7.1053 270 0.6519 0.5220 0.6519 0.8074
No log 7.1579 272 0.6554 0.5166 0.6554 0.8096
No log 7.2105 274 0.6524 0.5043 0.6524 0.8077
No log 7.2632 276 0.6414 0.5324 0.6414 0.8009
No log 7.3158 278 0.6400 0.5550 0.6400 0.8000
No log 7.3684 280 0.6375 0.5220 0.6375 0.7984
No log 7.4211 282 0.6315 0.5470 0.6315 0.7947
No log 7.4737 284 0.6418 0.5535 0.6418 0.8012
No log 7.5263 286 0.6718 0.5260 0.6718 0.8196
No log 7.5789 288 0.7020 0.5211 0.7020 0.8379
No log 7.6316 290 0.7047 0.5241 0.7047 0.8394
No log 7.6842 292 0.6869 0.5620 0.6869 0.8288
No log 7.7368 294 0.6930 0.5932 0.6930 0.8325
No log 7.7895 296 0.7027 0.5888 0.7027 0.8383
No log 7.8421 298 0.7055 0.5546 0.7055 0.8400
No log 7.8947 300 0.7126 0.5279 0.7126 0.8442
No log 7.9474 302 0.6961 0.5163 0.6961 0.8343
No log 8.0 304 0.6758 0.5124 0.6758 0.8221
No log 8.0526 306 0.6700 0.5433 0.6700 0.8185
No log 8.1053 308 0.6806 0.5759 0.6806 0.8250
No log 8.1579 310 0.6880 0.5478 0.6880 0.8294
No log 8.2105 312 0.6962 0.5225 0.6962 0.8344
No log 8.2632 314 0.6842 0.5504 0.6842 0.8272
No log 8.3158 316 0.6658 0.5490 0.6658 0.8159
No log 8.3684 318 0.6608 0.4939 0.6608 0.8129
No log 8.4211 320 0.6552 0.4878 0.6552 0.8094
No log 8.4737 322 0.6429 0.4585 0.6429 0.8018
No log 8.5263 324 0.6461 0.5136 0.6461 0.8038
No log 8.5789 326 0.6515 0.5112 0.6515 0.8072
No log 8.6316 328 0.6574 0.5119 0.6574 0.8108
No log 8.6842 330 0.6579 0.5278 0.6579 0.8111
No log 8.7368 332 0.6631 0.5244 0.6631 0.8143
No log 8.7895 334 0.6671 0.5114 0.6671 0.8168
No log 8.8421 336 0.6419 0.5009 0.6419 0.8012
No log 8.8947 338 0.6269 0.5214 0.6269 0.7918
No log 8.9474 340 0.6513 0.5054 0.6513 0.8070
No log 9.0 342 0.6749 0.4994 0.6749 0.8215
No log 9.0526 344 0.6814 0.4994 0.6814 0.8255
No log 9.1053 346 0.6538 0.4773 0.6538 0.8086
No log 9.1579 348 0.6315 0.4962 0.6315 0.7947
No log 9.2105 350 0.6520 0.4606 0.6520 0.8074
No log 9.2632 352 0.6596 0.4692 0.6596 0.8121
No log 9.3158 354 0.6455 0.5041 0.6455 0.8034
No log 9.3684 356 0.6326 0.4988 0.6326 0.7953
No log 9.4211 358 0.6220 0.5148 0.6220 0.7887
No log 9.4737 360 0.6304 0.5006 0.6304 0.7940
No log 9.5263 362 0.6378 0.5174 0.6378 0.7986
No log 9.5789 364 0.6474 0.5344 0.6474 0.8046
No log 9.6316 366 0.6601 0.5344 0.6601 0.8124
No log 9.6842 368 0.6833 0.4690 0.6833 0.8266
No log 9.7368 370 0.6954 0.4703 0.6954 0.8339
No log 9.7895 372 0.6922 0.5267 0.6922 0.8320
No log 9.8421 374 0.6979 0.5176 0.6979 0.8354
No log 9.8947 376 0.7086 0.5234 0.7086 0.8418
No log 9.9474 378 0.7141 0.5082 0.7141 0.8450
No log 10.0 380 0.7191 0.4955 0.7191 0.8480
No log 10.0526 382 0.7347 0.5328 0.7347 0.8571
No log 10.1053 384 0.7125 0.5484 0.7125 0.8441
No log 10.1579 386 0.6688 0.5197 0.6688 0.8178
No log 10.2105 388 0.6430 0.5453 0.6430 0.8019
No log 10.2632 390 0.6487 0.5533 0.6487 0.8054
No log 10.3158 392 0.6712 0.5390 0.6712 0.8193
No log 10.3684 394 0.6812 0.5423 0.6812 0.8254
No log 10.4211 396 0.6993 0.5083 0.6993 0.8363
No log 10.4737 398 0.7100 0.5355 0.7100 0.8426
No log 10.5263 400 0.7171 0.5316 0.7171 0.8468
No log 10.5789 402 0.7177 0.5227 0.7177 0.8472
No log 10.6316 404 0.7275 0.5098 0.7275 0.8529
No log 10.6842 406 0.7304 0.5096 0.7304 0.8546
No log 10.7368 408 0.7220 0.5216 0.7220 0.8497
No log 10.7895 410 0.7170 0.4934 0.7170 0.8468
No log 10.8421 412 0.7259 0.4464 0.7259 0.8520
No log 10.8947 414 0.7124 0.4412 0.7124 0.8440
No log 10.9474 416 0.6842 0.5109 0.6842 0.8272
No log 11.0 418 0.6792 0.5030 0.6792 0.8242
No log 11.0526 420 0.6618 0.5097 0.6618 0.8135
No log 11.1053 422 0.6554 0.5097 0.6554 0.8096
No log 11.1579 424 0.6658 0.5087 0.6658 0.8159
No log 11.2105 426 0.6756 0.5689 0.6756 0.8219
No log 11.2632 428 0.6796 0.5292 0.6796 0.8244
No log 11.3158 430 0.6825 0.5163 0.6825 0.8261
No log 11.3684 432 0.6943 0.5280 0.6943 0.8333
No log 11.4211 434 0.7113 0.5228 0.7113 0.8434
No log 11.4737 436 0.7064 0.5219 0.7064 0.8405
No log 11.5263 438 0.6760 0.5246 0.6760 0.8222
No log 11.5789 440 0.6841 0.5171 0.6841 0.8271
No log 11.6316 442 0.7137 0.5420 0.7137 0.8448
No log 11.6842 444 0.6958 0.5230 0.6958 0.8342
No log 11.7368 446 0.6566 0.5135 0.6566 0.8103
No log 11.7895 448 0.6484 0.5430 0.6484 0.8052
No log 11.8421 450 0.6659 0.5198 0.6659 0.8160
No log 11.8947 452 0.6750 0.5238 0.6750 0.8216
No log 11.9474 454 0.6592 0.5087 0.6592 0.8119
No log 12.0 456 0.6501 0.5038 0.6501 0.8063
No log 12.0526 458 0.6373 0.5112 0.6373 0.7983
No log 12.1053 460 0.6372 0.5391 0.6372 0.7982
No log 12.1579 462 0.6601 0.4991 0.6601 0.8125
No log 12.2105 464 0.7270 0.5257 0.7270 0.8526
No log 12.2632 466 0.7690 0.5038 0.7690 0.8769
No log 12.3158 468 0.7601 0.5092 0.7601 0.8719
No log 12.3684 470 0.7050 0.5233 0.7050 0.8397
No log 12.4211 472 0.6961 0.4761 0.6961 0.8343
No log 12.4737 474 0.7288 0.5562 0.7288 0.8537
No log 12.5263 476 0.7419 0.5150 0.7419 0.8613
No log 12.5789 478 0.7581 0.4930 0.7581 0.8707
No log 12.6316 480 0.7556 0.5042 0.7556 0.8692
No log 12.6842 482 0.7168 0.5124 0.7168 0.8466
No log 12.7368 484 0.6506 0.54 0.6506 0.8066
No log 12.7895 486 0.6272 0.5555 0.6272 0.7920
No log 12.8421 488 0.6503 0.5417 0.6503 0.8064
No log 12.8947 490 0.6488 0.5604 0.6488 0.8055
No log 12.9474 492 0.6361 0.5446 0.6361 0.7975
No log 13.0 494 0.6281 0.4884 0.6281 0.7925
No log 13.0526 496 0.6454 0.4985 0.6454 0.8034
No log 13.1053 498 0.6756 0.5329 0.6756 0.8219
0.546 13.1579 500 0.6922 0.5321 0.6922 0.8320
0.546 13.2105 502 0.7105 0.5195 0.7105 0.8429
0.546 13.2632 504 0.7157 0.5295 0.7157 0.8460
0.546 13.3158 506 0.6978 0.5364 0.6978 0.8354
0.546 13.3684 508 0.6925 0.5519 0.6925 0.8321
0.546 13.4211 510 0.6836 0.4823 0.6836 0.8268
0.546 13.4737 512 0.6753 0.4837 0.6753 0.8218
0.546 13.5263 514 0.6701 0.4767 0.6701 0.8186
0.546 13.5789 516 0.6749 0.4761 0.6749 0.8215
0.546 13.6316 518 0.6820 0.4819 0.6820 0.8259

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.1+cu121
  • Datasets 3.2.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERTFreeze_run2_AugV5_k15_task2_organization

Finetuned
(4023)
this model