ArabicNewSplits6_FineTuningAraBERTFreeze_run1_AugV5_k17_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7327
  • Qwk: 0.5310
  • Mse: 0.7327
  • Rmse: 0.8560

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0476 2 6.4129 -0.0278 6.4129 2.5324
No log 0.0952 4 4.3705 -0.0299 4.3705 2.0906
No log 0.1429 6 3.1054 0.0108 3.1054 1.7622
No log 0.1905 8 2.1835 -0.0015 2.1835 1.4777
No log 0.2381 10 1.5484 -0.0143 1.5484 1.2444
No log 0.2857 12 1.2081 0.0103 1.2081 1.0991
No log 0.3333 14 1.1184 0.0113 1.1184 1.0575
No log 0.3810 16 1.0768 0.0616 1.0768 1.0377
No log 0.4286 18 1.0894 0.0161 1.0894 1.0437
No log 0.4762 20 1.1070 0.0723 1.1070 1.0521
No log 0.5238 22 0.9790 0.0831 0.9790 0.9895
No log 0.5714 24 0.9453 0.1022 0.9453 0.9723
No log 0.6190 26 0.9456 0.0308 0.9456 0.9724
No log 0.6667 28 0.8824 0.1140 0.8824 0.9393
No log 0.7143 30 0.9057 0.1182 0.9057 0.9517
No log 0.7619 32 0.9325 0.0998 0.9325 0.9656
No log 0.8095 34 0.9847 0.1575 0.9847 0.9923
No log 0.8571 36 1.0482 0.0739 1.0482 1.0238
No log 0.9048 38 1.0332 0.1336 1.0332 1.0164
No log 0.9524 40 0.9029 0.2339 0.9029 0.9502
No log 1.0 42 0.8458 0.2552 0.8458 0.9197
No log 1.0476 44 0.8342 0.2368 0.8342 0.9133
No log 1.0952 46 0.8733 0.2388 0.8733 0.9345
No log 1.1429 48 0.8734 0.2885 0.8734 0.9346
No log 1.1905 50 0.9292 0.2745 0.9292 0.9639
No log 1.2381 52 0.9547 0.2733 0.9547 0.9771
No log 1.2857 54 0.9091 0.2565 0.9091 0.9535
No log 1.3333 56 0.9092 0.2862 0.9092 0.9535
No log 1.3810 58 0.9601 0.2667 0.9601 0.9799
No log 1.4286 60 0.9006 0.2938 0.9006 0.9490
No log 1.4762 62 0.7776 0.3863 0.7776 0.8818
No log 1.5238 64 0.6944 0.4165 0.6944 0.8333
No log 1.5714 66 0.6967 0.4215 0.6967 0.8347
No log 1.6190 68 0.7219 0.4388 0.7219 0.8496
No log 1.6667 70 0.7527 0.4142 0.7527 0.8676
No log 1.7143 72 0.8897 0.2280 0.8897 0.9433
No log 1.7619 74 0.8480 0.3325 0.8480 0.9209
No log 1.8095 76 0.7373 0.3994 0.7373 0.8586
No log 1.8571 78 0.6805 0.4343 0.6805 0.8249
No log 1.9048 80 0.6174 0.4808 0.6174 0.7858
No log 1.9524 82 0.5946 0.4340 0.5946 0.7711
No log 2.0 84 0.5780 0.4397 0.5780 0.7602
No log 2.0476 86 0.5763 0.4669 0.5763 0.7592
No log 2.0952 88 0.5790 0.4699 0.5790 0.7609
No log 2.1429 90 0.5817 0.4851 0.5817 0.7627
No log 2.1905 92 0.5830 0.4690 0.5830 0.7636
No log 2.2381 94 0.6195 0.4882 0.6195 0.7871
No log 2.2857 96 0.6633 0.4756 0.6633 0.8144
No log 2.3333 98 0.7332 0.4517 0.7332 0.8563
No log 2.3810 100 0.8118 0.4429 0.8118 0.9010
No log 2.4286 102 0.8412 0.4595 0.8412 0.9172
No log 2.4762 104 0.9248 0.4174 0.9248 0.9616
No log 2.5238 106 0.8876 0.4208 0.8876 0.9421
No log 2.5714 108 0.7461 0.4677 0.7461 0.8638
No log 2.6190 110 0.6157 0.5148 0.6157 0.7847
No log 2.6667 112 0.5442 0.5065 0.5442 0.7377
No log 2.7143 114 0.5332 0.4255 0.5332 0.7302
No log 2.7619 116 0.5292 0.4416 0.5292 0.7274
No log 2.8095 118 0.5282 0.4440 0.5282 0.7268
No log 2.8571 120 0.5559 0.5365 0.5559 0.7456
No log 2.9048 122 0.6226 0.5300 0.6226 0.7890
No log 2.9524 124 0.6845 0.5134 0.6845 0.8273
No log 3.0 126 0.6790 0.5264 0.6790 0.8240
No log 3.0476 128 0.6518 0.4943 0.6518 0.8073
No log 3.0952 130 0.6415 0.4994 0.6415 0.8010
No log 3.1429 132 0.6157 0.5603 0.6157 0.7847
No log 3.1905 134 0.6040 0.5603 0.6040 0.7772
No log 3.2381 136 0.6075 0.5271 0.6075 0.7794
No log 3.2857 138 0.6211 0.5406 0.6211 0.7881
No log 3.3333 140 0.6420 0.5623 0.6420 0.8013
No log 3.3810 142 0.6743 0.5142 0.6743 0.8212
No log 3.4286 144 0.6946 0.5020 0.6946 0.8334
No log 3.4762 146 0.7112 0.4955 0.7112 0.8433
No log 3.5238 148 0.7523 0.4596 0.7523 0.8674
No log 3.5714 150 0.7430 0.4578 0.7430 0.8619
No log 3.6190 152 0.6904 0.4756 0.6904 0.8309
No log 3.6667 154 0.7094 0.4954 0.7094 0.8423
No log 3.7143 156 0.7955 0.4455 0.7955 0.8919
No log 3.7619 158 0.8879 0.3584 0.8879 0.9423
No log 3.8095 160 0.9024 0.3916 0.9024 0.9499
No log 3.8571 162 0.9123 0.3638 0.9123 0.9551
No log 3.9048 164 0.8239 0.4456 0.8239 0.9077
No log 3.9524 166 0.8221 0.4326 0.8221 0.9067
No log 4.0 168 0.7992 0.4141 0.7992 0.8940
No log 4.0476 170 0.7525 0.4501 0.7525 0.8675
No log 4.0952 172 0.7221 0.4967 0.7221 0.8497
No log 4.1429 174 0.7307 0.5191 0.7307 0.8548
No log 4.1905 176 0.7447 0.5070 0.7447 0.8629
No log 4.2381 178 0.7217 0.5167 0.7217 0.8495
No log 4.2857 180 0.6936 0.5167 0.6936 0.8329
No log 4.3333 182 0.6779 0.5540 0.6779 0.8233
No log 4.3810 184 0.6865 0.5033 0.6865 0.8285
No log 4.4286 186 0.7023 0.4887 0.7023 0.8380
No log 4.4762 188 0.6898 0.4895 0.6898 0.8305
No log 4.5238 190 0.6645 0.4895 0.6645 0.8151
No log 4.5714 192 0.6520 0.4838 0.6520 0.8075
No log 4.6190 194 0.6545 0.4819 0.6545 0.8090
No log 4.6667 196 0.6529 0.4799 0.6529 0.8080
No log 4.7143 198 0.6572 0.4691 0.6572 0.8107
No log 4.7619 200 0.6689 0.5122 0.6689 0.8179
No log 4.8095 202 0.7142 0.5178 0.7142 0.8451
No log 4.8571 204 0.7472 0.5014 0.7472 0.8644
No log 4.9048 206 0.7439 0.5135 0.7439 0.8625
No log 4.9524 208 0.7256 0.4903 0.7256 0.8518
No log 5.0 210 0.7254 0.4997 0.7254 0.8517
No log 5.0476 212 0.7137 0.5291 0.7137 0.8448
No log 5.0952 214 0.7268 0.4912 0.7268 0.8525
No log 5.1429 216 0.7396 0.5277 0.7396 0.8600
No log 5.1905 218 0.7405 0.5114 0.7405 0.8605
No log 5.2381 220 0.7375 0.4949 0.7375 0.8588
No log 5.2857 222 0.7684 0.4952 0.7684 0.8766
No log 5.3333 224 0.8225 0.5112 0.8225 0.9069
No log 5.3810 226 0.8066 0.5103 0.8066 0.8981
No log 5.4286 228 0.7299 0.5372 0.7299 0.8544
No log 5.4762 230 0.6838 0.5219 0.6838 0.8269
No log 5.5238 232 0.6623 0.4835 0.6623 0.8138
No log 5.5714 234 0.6567 0.4608 0.6567 0.8104
No log 5.6190 236 0.6678 0.4524 0.6678 0.8172
No log 5.6667 238 0.6757 0.4773 0.6757 0.8220
No log 5.7143 240 0.6906 0.4725 0.6906 0.8310
No log 5.7619 242 0.7085 0.4518 0.7085 0.8417
No log 5.8095 244 0.7203 0.4663 0.7203 0.8487
No log 5.8571 246 0.7245 0.4755 0.7245 0.8512
No log 5.9048 248 0.7232 0.5011 0.7232 0.8504
No log 5.9524 250 0.7238 0.5030 0.7238 0.8508
No log 6.0 252 0.7368 0.5200 0.7368 0.8584
No log 6.0476 254 0.7267 0.4770 0.7267 0.8525
No log 6.0952 256 0.7272 0.4848 0.7272 0.8528
No log 6.1429 258 0.7162 0.4626 0.7162 0.8463
No log 6.1905 260 0.6970 0.4953 0.6970 0.8348
No log 6.2381 262 0.6790 0.5382 0.6790 0.8240
No log 6.2857 264 0.6677 0.5067 0.6677 0.8171
No log 6.3333 266 0.6625 0.5170 0.6625 0.8139
No log 6.3810 268 0.6590 0.5379 0.6590 0.8118
No log 6.4286 270 0.6559 0.5423 0.6559 0.8099
No log 6.4762 272 0.6546 0.5480 0.6546 0.8091
No log 6.5238 274 0.6718 0.5168 0.6718 0.8197
No log 6.5714 276 0.7072 0.5089 0.7072 0.8410
No log 6.6190 278 0.7165 0.5089 0.7165 0.8465
No log 6.6667 280 0.6958 0.5056 0.6958 0.8341
No log 6.7143 282 0.6946 0.4910 0.6946 0.8335
No log 6.7619 284 0.7128 0.5328 0.7128 0.8443
No log 6.8095 286 0.7408 0.5160 0.7408 0.8607
No log 6.8571 288 0.7767 0.5208 0.7767 0.8813
No log 6.9048 290 0.8149 0.5322 0.8149 0.9027
No log 6.9524 292 0.8145 0.5029 0.8145 0.9025
No log 7.0 294 0.7798 0.5041 0.7798 0.8831
No log 7.0476 296 0.7786 0.5321 0.7786 0.8824
No log 7.0952 298 0.7911 0.5193 0.7911 0.8894
No log 7.1429 300 0.7784 0.5188 0.7784 0.8823
No log 7.1905 302 0.7270 0.5246 0.7270 0.8527
No log 7.2381 304 0.6897 0.5385 0.6897 0.8305
No log 7.2857 306 0.6899 0.5541 0.6899 0.8306
No log 7.3333 308 0.6938 0.525 0.6938 0.8329
No log 7.3810 310 0.6960 0.5402 0.6960 0.8342
No log 7.4286 312 0.7022 0.5318 0.7022 0.8380
No log 7.4762 314 0.7034 0.5565 0.7034 0.8387
No log 7.5238 316 0.7113 0.5550 0.7113 0.8434
No log 7.5714 318 0.7258 0.5445 0.7258 0.8519
No log 7.6190 320 0.7280 0.5044 0.7280 0.8532
No log 7.6667 322 0.7376 0.4676 0.7376 0.8588
No log 7.7143 324 0.7380 0.4778 0.7380 0.8590
No log 7.7619 326 0.7314 0.5120 0.7314 0.8552
No log 7.8095 328 0.7223 0.5025 0.7223 0.8499
No log 7.8571 330 0.7119 0.4870 0.7119 0.8437
No log 7.9048 332 0.7018 0.5170 0.7018 0.8377
No log 7.9524 334 0.6997 0.4936 0.6997 0.8365
No log 8.0 336 0.7222 0.5161 0.7222 0.8498
No log 8.0476 338 0.7751 0.4808 0.7751 0.8804
No log 8.0952 340 0.7739 0.4870 0.7739 0.8797
No log 8.1429 342 0.7364 0.4800 0.7364 0.8582
No log 8.1905 344 0.7020 0.5119 0.7020 0.8379
No log 8.2381 346 0.6834 0.5262 0.6834 0.8267
No log 8.2857 348 0.6783 0.5184 0.6783 0.8236
No log 8.3333 350 0.6733 0.5374 0.6733 0.8206
No log 8.3810 352 0.6683 0.5303 0.6683 0.8175
No log 8.4286 354 0.6725 0.5149 0.6725 0.8201
No log 8.4762 356 0.6851 0.5222 0.6851 0.8277
No log 8.5238 358 0.7039 0.5227 0.7039 0.8390
No log 8.5714 360 0.7158 0.5442 0.7158 0.8460
No log 8.6190 362 0.7190 0.5524 0.7190 0.8479
No log 8.6667 364 0.7236 0.5105 0.7236 0.8506
No log 8.7143 366 0.7212 0.5206 0.7212 0.8492
No log 8.7619 368 0.7215 0.5230 0.7215 0.8494
No log 8.8095 370 0.7246 0.5132 0.7246 0.8512
No log 8.8571 372 0.7640 0.5120 0.7640 0.8741
No log 8.9048 374 0.7929 0.5206 0.7929 0.8904
No log 8.9524 376 0.7926 0.5265 0.7926 0.8903
No log 9.0 378 0.7697 0.5091 0.7697 0.8773
No log 9.0476 380 0.7424 0.5149 0.7424 0.8616
No log 9.0952 382 0.7268 0.5414 0.7268 0.8525
No log 9.1429 384 0.7208 0.5426 0.7208 0.8490
No log 9.1905 386 0.7174 0.5426 0.7174 0.8470
No log 9.2381 388 0.6918 0.5372 0.6918 0.8317
No log 9.2857 390 0.6549 0.5351 0.6549 0.8092
No log 9.3333 392 0.6526 0.5694 0.6526 0.8078
No log 9.3810 394 0.6733 0.5314 0.6733 0.8206
No log 9.4286 396 0.6807 0.5194 0.6807 0.8251
No log 9.4762 398 0.6887 0.5097 0.6887 0.8299
No log 9.5238 400 0.7188 0.5282 0.7188 0.8478
No log 9.5714 402 0.7890 0.5480 0.7890 0.8883
No log 9.6190 404 0.8264 0.5056 0.8264 0.9090
No log 9.6667 406 0.8018 0.5199 0.8018 0.8954
No log 9.7143 408 0.7518 0.5294 0.7518 0.8671
No log 9.7619 410 0.7306 0.5167 0.7306 0.8548
No log 9.8095 412 0.7298 0.4886 0.7298 0.8543
No log 9.8571 414 0.7382 0.5114 0.7382 0.8592
No log 9.9048 416 0.7426 0.5561 0.7426 0.8618
No log 9.9524 418 0.7460 0.5629 0.7460 0.8637
No log 10.0 420 0.7458 0.5332 0.7458 0.8636
No log 10.0476 422 0.7560 0.5629 0.7560 0.8695
No log 10.0952 424 0.7847 0.5601 0.7847 0.8858
No log 10.1429 426 0.8097 0.5289 0.8097 0.8998
No log 10.1905 428 0.7984 0.5300 0.7984 0.8935
No log 10.2381 430 0.7515 0.5668 0.7515 0.8669
No log 10.2857 432 0.7191 0.5563 0.7191 0.8480
No log 10.3333 434 0.7262 0.5077 0.7262 0.8522
No log 10.3810 436 0.7284 0.5355 0.7284 0.8535
No log 10.4286 438 0.7275 0.5482 0.7275 0.8529
No log 10.4762 440 0.7386 0.5560 0.7386 0.8594
No log 10.5238 442 0.7363 0.5299 0.7363 0.8581
No log 10.5714 444 0.7085 0.5520 0.7085 0.8417
No log 10.6190 446 0.6910 0.5383 0.6910 0.8313
No log 10.6667 448 0.6944 0.5673 0.6944 0.8333
No log 10.7143 450 0.7249 0.4836 0.7249 0.8514
No log 10.7619 452 0.7275 0.4495 0.7275 0.8529
No log 10.8095 454 0.7005 0.4781 0.7005 0.8370
No log 10.8571 456 0.6730 0.5652 0.6730 0.8203
No log 10.9048 458 0.6703 0.5618 0.6703 0.8187
No log 10.9524 460 0.6645 0.5395 0.6645 0.8152
No log 11.0 462 0.6661 0.5636 0.6661 0.8162
No log 11.0476 464 0.6746 0.5803 0.6746 0.8214
No log 11.0952 466 0.6832 0.5513 0.6832 0.8265
No log 11.1429 468 0.6702 0.5748 0.6702 0.8187
No log 11.1905 470 0.6694 0.5720 0.6694 0.8182
No log 11.2381 472 0.6780 0.5333 0.6780 0.8234
No log 11.2857 474 0.6899 0.5120 0.6899 0.8306
No log 11.3333 476 0.6942 0.5465 0.6942 0.8332
No log 11.3810 478 0.6967 0.5425 0.6967 0.8347
No log 11.4286 480 0.7108 0.5330 0.7108 0.8431
No log 11.4762 482 0.7232 0.5322 0.7232 0.8504
No log 11.5238 484 0.7305 0.5446 0.7305 0.8547
No log 11.5714 486 0.7283 0.5552 0.7283 0.8534
No log 11.6190 488 0.7189 0.5591 0.7189 0.8479
No log 11.6667 490 0.7151 0.5591 0.7151 0.8456
No log 11.7143 492 0.7163 0.5644 0.7163 0.8463
No log 11.7619 494 0.7212 0.5938 0.7212 0.8493
No log 11.8095 496 0.7214 0.5938 0.7214 0.8493
No log 11.8571 498 0.7270 0.5685 0.7270 0.8527
0.5114 11.9048 500 0.7269 0.5687 0.7269 0.8526
0.5114 11.9524 502 0.7331 0.5068 0.7331 0.8562
0.5114 12.0 504 0.7226 0.5252 0.7226 0.8501
0.5114 12.0476 506 0.7129 0.5244 0.7129 0.8444
0.5114 12.0952 508 0.7183 0.5206 0.7183 0.8475
0.5114 12.1429 510 0.7327 0.5310 0.7327 0.8560

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.1+cu121
  • Datasets 3.2.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERTFreeze_run1_AugV5_k17_task2_organization

Finetuned
(4024)
this model