ArabicNewSplits5_FineTuningAraBERT_run2_AugV5_k6_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7040
  • Qwk: 0.5140
  • Mse: 0.7040
  • Rmse: 0.8391

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.05 2 4.2345 -0.0066 4.2345 2.0578
No log 0.1 4 2.5189 0.0111 2.5189 1.5871
No log 0.15 6 1.6177 -0.0076 1.6177 1.2719
No log 0.2 8 1.3707 0.0353 1.3707 1.1708
No log 0.25 10 1.0898 0.0787 1.0898 1.0439
No log 0.3 12 0.9221 0.0386 0.9221 0.9603
No log 0.35 14 0.8610 0.0745 0.8610 0.9279
No log 0.4 16 0.8883 0.0470 0.8883 0.9425
No log 0.45 18 1.0904 0.0969 1.0904 1.0442
No log 0.5 20 1.3596 0.1194 1.3596 1.1660
No log 0.55 22 1.4447 0.1047 1.4447 1.2020
No log 0.6 24 1.5277 0.0644 1.5277 1.2360
No log 0.65 26 1.3045 0.0572 1.3045 1.1422
No log 0.7 28 0.9030 0.1094 0.9030 0.9502
No log 0.75 30 0.7592 0.1736 0.7592 0.8713
No log 0.8 32 0.6924 0.2742 0.6924 0.8321
No log 0.85 34 0.6722 0.3162 0.6722 0.8199
No log 0.9 36 0.6666 0.3420 0.6666 0.8165
No log 0.95 38 0.6837 0.2565 0.6837 0.8268
No log 1.0 40 0.7820 0.2947 0.7820 0.8843
No log 1.05 42 0.9433 0.2560 0.9433 0.9712
No log 1.1 44 0.9602 0.2774 0.9602 0.9799
No log 1.15 46 0.9694 0.2799 0.9694 0.9846
No log 1.2 48 1.0663 0.2469 1.0663 1.0326
No log 1.25 50 1.0964 0.2282 1.0964 1.0471
No log 1.3 52 0.9938 0.2660 0.9938 0.9969
No log 1.35 54 0.8477 0.1881 0.8477 0.9207
No log 1.4 56 0.7796 0.1488 0.7796 0.8830
No log 1.45 58 0.6956 0.2961 0.6956 0.8340
No log 1.5 60 0.7317 0.4141 0.7317 0.8554
No log 1.55 62 1.0914 0.3114 1.0914 1.0447
No log 1.6 64 1.1584 0.3099 1.1584 1.0763
No log 1.65 66 1.0617 0.2851 1.0617 1.0304
No log 1.7 68 0.9695 0.2888 0.9695 0.9846
No log 1.75 70 0.7529 0.3814 0.7529 0.8677
No log 1.8 72 0.6535 0.3595 0.6535 0.8084
No log 1.85 74 0.6282 0.4313 0.6282 0.7926
No log 1.9 76 0.6050 0.4561 0.6050 0.7778
No log 1.95 78 0.6171 0.4340 0.6171 0.7856
No log 2.0 80 0.6705 0.3796 0.6705 0.8188
No log 2.05 82 0.7847 0.3405 0.7847 0.8858
No log 2.1 84 0.8992 0.3295 0.8992 0.9483
No log 2.15 86 0.8629 0.3375 0.8629 0.9289
No log 2.2 88 0.8449 0.3495 0.8449 0.9192
No log 2.25 90 0.8443 0.3502 0.8443 0.9188
No log 2.3 92 0.8470 0.3665 0.8470 0.9203
No log 2.35 94 0.8616 0.3970 0.8616 0.9282
No log 2.4 96 0.7459 0.3883 0.7459 0.8636
No log 2.45 98 0.6125 0.4661 0.6125 0.7826
No log 2.5 100 0.6291 0.4532 0.6291 0.7931
No log 2.55 102 0.7107 0.3735 0.7107 0.8430
No log 2.6 104 0.7318 0.2592 0.7318 0.8554
No log 2.65 106 0.7158 0.3686 0.7158 0.8460
No log 2.7 108 0.7374 0.4176 0.7374 0.8587
No log 2.75 110 0.7211 0.4420 0.7211 0.8492
No log 2.8 112 0.6437 0.4404 0.6437 0.8023
No log 2.85 114 0.6052 0.4507 0.6052 0.7780
No log 2.9 116 0.6063 0.4019 0.6063 0.7786
No log 2.95 118 0.6032 0.4382 0.6032 0.7767
No log 3.0 120 0.6639 0.4039 0.6639 0.8148
No log 3.05 122 0.7689 0.4192 0.7689 0.8768
No log 3.1 124 0.8349 0.4214 0.8349 0.9137
No log 3.15 126 0.8470 0.3929 0.8470 0.9203
No log 3.2 128 0.7725 0.4378 0.7725 0.8789
No log 3.25 130 0.6931 0.4831 0.6931 0.8325
No log 3.3 132 0.6742 0.4972 0.6742 0.8211
No log 3.35 134 0.7779 0.4903 0.7779 0.8820
No log 3.4 136 0.8525 0.4511 0.8525 0.9233
No log 3.45 138 0.7620 0.5572 0.7620 0.8729
No log 3.5 140 0.7525 0.5214 0.7525 0.8675
No log 3.55 142 0.8434 0.5182 0.8434 0.9184
No log 3.6 144 0.8466 0.4781 0.8466 0.9201
No log 3.65 146 0.7391 0.5526 0.7391 0.8597
No log 3.7 148 0.6688 0.5250 0.6688 0.8178
No log 3.75 150 0.6461 0.5063 0.6461 0.8038
No log 3.8 152 0.6446 0.5265 0.6446 0.8029
No log 3.85 154 0.6671 0.5144 0.6671 0.8168
No log 3.9 156 0.7272 0.5118 0.7272 0.8527
No log 3.95 158 0.8237 0.4533 0.8237 0.9076
No log 4.0 160 0.8398 0.4536 0.8398 0.9164
No log 4.05 162 0.7687 0.4693 0.7687 0.8767
No log 4.1 164 0.7068 0.4597 0.7068 0.8407
No log 4.15 166 0.6993 0.4833 0.6993 0.8362
No log 4.2 168 0.7247 0.4615 0.7247 0.8513
No log 4.25 170 0.7349 0.4664 0.7349 0.8572
No log 4.3 172 0.7385 0.4517 0.7385 0.8594
No log 4.35 174 0.7651 0.4631 0.7651 0.8747
No log 4.4 176 0.8042 0.4552 0.8042 0.8968
No log 4.45 178 0.8348 0.4530 0.8348 0.9137
No log 4.5 180 0.8073 0.4471 0.8073 0.8985
No log 4.55 182 0.7482 0.3690 0.7482 0.8650
No log 4.6 184 0.7305 0.4068 0.7305 0.8547
No log 4.65 186 0.7179 0.4209 0.7179 0.8473
No log 4.7 188 0.6999 0.4145 0.6999 0.8366
No log 4.75 190 0.7194 0.4197 0.7194 0.8482
No log 4.8 192 0.7754 0.4683 0.7754 0.8806
No log 4.85 194 0.7579 0.4596 0.7579 0.8706
No log 4.9 196 0.7361 0.4731 0.7361 0.8580
No log 4.95 198 0.7114 0.4445 0.7114 0.8434
No log 5.0 200 0.7258 0.4451 0.7258 0.8519
No log 5.05 202 0.7285 0.4659 0.7285 0.8535
No log 5.1 204 0.7233 0.5105 0.7233 0.8504
No log 5.15 206 0.7113 0.5183 0.7113 0.8434
No log 5.2 208 0.6961 0.5133 0.6961 0.8344
No log 5.25 210 0.6927 0.5115 0.6927 0.8323
No log 5.3 212 0.7031 0.5486 0.7031 0.8385
No log 5.35 214 0.7343 0.5217 0.7343 0.8569
No log 5.4 216 0.7441 0.5455 0.7441 0.8626
No log 5.45 218 0.7450 0.5178 0.7450 0.8631
No log 5.5 220 0.7611 0.5119 0.7611 0.8724
No log 5.55 222 0.7736 0.5203 0.7736 0.8796
No log 5.6 224 0.7669 0.5178 0.7669 0.8757
No log 5.65 226 0.7578 0.5078 0.7578 0.8705
No log 5.7 228 0.7693 0.4898 0.7693 0.8771
No log 5.75 230 0.7724 0.4898 0.7724 0.8789
No log 5.8 232 0.7409 0.5286 0.7409 0.8608
No log 5.85 234 0.7300 0.5461 0.7300 0.8544
No log 5.9 236 0.7157 0.5461 0.7157 0.8460
No log 5.95 238 0.7136 0.5403 0.7136 0.8447
No log 6.0 240 0.7053 0.5446 0.7053 0.8398
No log 6.05 242 0.7031 0.5299 0.7031 0.8385
No log 6.1 244 0.7179 0.5463 0.7179 0.8473
No log 6.15 246 0.7380 0.5332 0.7380 0.8591
No log 6.2 248 0.7582 0.5364 0.7582 0.8707
No log 6.25 250 0.7779 0.5269 0.7779 0.8820
No log 6.3 252 0.7880 0.5188 0.7880 0.8877
No log 6.35 254 0.8092 0.5216 0.8092 0.8995
No log 6.4 256 0.8443 0.4855 0.8443 0.9189
No log 6.45 258 0.8210 0.5115 0.8210 0.9061
No log 6.5 260 0.7678 0.5232 0.7678 0.8762
No log 6.55 262 0.7504 0.5187 0.7504 0.8662
No log 6.6 264 0.7440 0.5292 0.7440 0.8626
No log 6.65 266 0.7554 0.5296 0.7554 0.8692
No log 6.7 268 0.7656 0.5284 0.7656 0.8750
No log 6.75 270 0.7740 0.5253 0.7740 0.8797
No log 6.8 272 0.7853 0.5227 0.7853 0.8862
No log 6.85 274 0.7857 0.5282 0.7857 0.8864
No log 6.9 276 0.7609 0.5293 0.7609 0.8723
No log 6.95 278 0.7320 0.5292 0.7320 0.8556
No log 7.0 280 0.7209 0.5303 0.7209 0.8491
No log 7.05 282 0.7243 0.5136 0.7243 0.8511
No log 7.1 284 0.7159 0.5316 0.7159 0.8461
No log 7.15 286 0.6996 0.5403 0.6996 0.8364
No log 7.2 288 0.6906 0.5095 0.6906 0.8310
No log 7.25 290 0.6813 0.5477 0.6813 0.8254
No log 7.3 292 0.6796 0.5397 0.6796 0.8244
No log 7.35 294 0.6852 0.5320 0.6852 0.8277
No log 7.4 296 0.6998 0.5148 0.6998 0.8366
No log 7.45 298 0.7300 0.5254 0.7300 0.8544
No log 7.5 300 0.7663 0.5068 0.7663 0.8754
No log 7.55 302 0.7856 0.5089 0.7856 0.8863
No log 7.6 304 0.7638 0.4947 0.7638 0.8740
No log 7.65 306 0.7243 0.5192 0.7243 0.8511
No log 7.7 308 0.7081 0.5184 0.7081 0.8415
No log 7.75 310 0.7144 0.5170 0.7144 0.8452
No log 7.8 312 0.7192 0.5306 0.7192 0.8481
No log 7.85 314 0.7188 0.5258 0.7188 0.8478
No log 7.9 316 0.7209 0.5184 0.7209 0.8491
No log 7.95 318 0.7242 0.5120 0.7242 0.8510
No log 8.0 320 0.7298 0.5087 0.7298 0.8543
No log 8.05 322 0.7250 0.5080 0.7250 0.8515
No log 8.1 324 0.7300 0.5235 0.7300 0.8544
No log 8.15 326 0.7247 0.5230 0.7247 0.8513
No log 8.2 328 0.7123 0.5127 0.7123 0.8440
No log 8.25 330 0.7096 0.5137 0.7096 0.8424
No log 8.3 332 0.7048 0.5049 0.7048 0.8395
No log 8.35 334 0.6978 0.5044 0.6978 0.8353
No log 8.4 336 0.6960 0.5119 0.6960 0.8343
No log 8.45 338 0.6967 0.5281 0.6967 0.8347
No log 8.5 340 0.7015 0.5434 0.7015 0.8375
No log 8.55 342 0.7071 0.5348 0.7071 0.8409
No log 8.6 344 0.7082 0.5348 0.7082 0.8415
No log 8.65 346 0.7065 0.5274 0.7065 0.8405
No log 8.7 348 0.7043 0.5052 0.7043 0.8393
No log 8.75 350 0.7112 0.5092 0.7112 0.8433
No log 8.8 352 0.7238 0.5073 0.7238 0.8508
No log 8.85 354 0.7338 0.5031 0.7338 0.8566
No log 8.9 356 0.7311 0.5031 0.7311 0.8551
No log 8.95 358 0.7180 0.5176 0.7180 0.8474
No log 9.0 360 0.7074 0.5092 0.7074 0.8411
No log 9.05 362 0.7016 0.5092 0.7016 0.8376
No log 9.1 364 0.6953 0.5167 0.6953 0.8338
No log 9.15 366 0.6910 0.5109 0.6910 0.8313
No log 9.2 368 0.6893 0.5194 0.6893 0.8302
No log 9.25 370 0.6905 0.5194 0.6905 0.8309
No log 9.3 372 0.6933 0.5253 0.6933 0.8326
No log 9.35 374 0.6956 0.5109 0.6956 0.8340
No log 9.4 376 0.6966 0.5109 0.6966 0.8346
No log 9.45 378 0.6974 0.5109 0.6974 0.8351
No log 9.5 380 0.6991 0.5167 0.6991 0.8361
No log 9.55 382 0.7032 0.5130 0.7032 0.8386
No log 9.6 384 0.7079 0.5123 0.7079 0.8413
No log 9.65 386 0.7105 0.5123 0.7105 0.8429
No log 9.7 388 0.7119 0.4983 0.7119 0.8437
No log 9.75 390 0.7114 0.4983 0.7114 0.8434
No log 9.8 392 0.7094 0.5123 0.7094 0.8423
No log 9.85 394 0.7069 0.5123 0.7069 0.8408
No log 9.9 396 0.7049 0.5130 0.7049 0.8396
No log 9.95 398 0.7043 0.5130 0.7043 0.8392
No log 10.0 400 0.7040 0.5140 0.7040 0.8391

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits5_FineTuningAraBERT_run2_AugV5_k6_task2_organization

Finetuned
(4023)
this model