ArabicNewSplits6_FineTuningAraBERTFreeze_run1_AugV5_k6_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7516
  • Qwk: 0.5185
  • Mse: 0.7516
  • Rmse: 0.8670

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.125 2 6.4267 -0.0218 6.4267 2.5351
No log 0.25 4 4.2604 -0.0201 4.2604 2.0641
No log 0.375 6 2.9315 -0.0051 2.9315 1.7122
No log 0.5 8 2.0387 0.0022 2.0387 1.4278
No log 0.625 10 1.4472 0.0181 1.4472 1.2030
No log 0.75 12 1.1906 0.0143 1.1906 1.0911
No log 0.875 14 1.0876 0.0501 1.0876 1.0429
No log 1.0 16 0.9912 0.0292 0.9912 0.9956
No log 1.125 18 0.8560 0.1177 0.8560 0.9252
No log 1.25 20 0.7199 0.2171 0.7199 0.8485
No log 1.375 22 0.7144 0.2775 0.7144 0.8452
No log 1.5 24 0.7139 0.2732 0.7139 0.8449
No log 1.625 26 0.7170 0.2689 0.7170 0.8468
No log 1.75 28 0.7246 0.2442 0.7246 0.8513
No log 1.875 30 0.7426 0.2213 0.7426 0.8617
No log 2.0 32 0.7251 0.2099 0.7251 0.8516
No log 2.125 34 0.7101 0.2674 0.7101 0.8427
No log 2.25 36 0.7549 0.3289 0.7549 0.8689
No log 2.375 38 0.7912 0.3150 0.7912 0.8895
No log 2.5 40 0.7508 0.3392 0.7508 0.8665
No log 2.625 42 0.6400 0.4372 0.6400 0.8000
No log 2.75 44 0.5891 0.4320 0.5891 0.7675
No log 2.875 46 0.5837 0.4235 0.5837 0.7640
No log 3.0 48 0.6033 0.3459 0.6033 0.7767
No log 3.125 50 0.6605 0.3036 0.6605 0.8127
No log 3.25 52 0.6663 0.3576 0.6663 0.8163
No log 3.375 54 0.6785 0.3793 0.6785 0.8237
No log 3.5 56 0.6884 0.4422 0.6884 0.8297
No log 3.625 58 0.6979 0.4619 0.6979 0.8354
No log 3.75 60 0.7081 0.4297 0.7081 0.8415
No log 3.875 62 0.6911 0.4318 0.6911 0.8313
No log 4.0 64 0.5996 0.4906 0.5996 0.7743
No log 4.125 66 0.5163 0.5 0.5163 0.7185
No log 4.25 68 0.5106 0.4461 0.5106 0.7145
No log 4.375 70 0.6031 0.3830 0.6031 0.7766
No log 4.5 72 0.6171 0.3377 0.6171 0.7856
No log 4.625 74 0.5908 0.3746 0.5908 0.7686
No log 4.75 76 0.5580 0.4495 0.5580 0.7470
No log 4.875 78 0.5455 0.5132 0.5455 0.7386
No log 5.0 80 0.5584 0.5123 0.5584 0.7473
No log 5.125 82 0.5848 0.5205 0.5848 0.7647
No log 5.25 84 0.6381 0.5160 0.6381 0.7988
No log 5.375 86 0.6336 0.5232 0.6336 0.7960
No log 5.5 88 0.6091 0.5377 0.6091 0.7805
No log 5.625 90 0.5904 0.5247 0.5904 0.7684
No log 5.75 92 0.5529 0.5212 0.5529 0.7436
No log 5.875 94 0.5524 0.5321 0.5524 0.7433
No log 6.0 96 0.5489 0.5212 0.5489 0.7409
No log 6.125 98 0.5090 0.5088 0.5090 0.7134
No log 6.25 100 0.5027 0.5536 0.5027 0.7090
No log 6.375 102 0.5137 0.5482 0.5137 0.7168
No log 6.5 104 0.5199 0.5410 0.5199 0.7210
No log 6.625 106 0.5346 0.5683 0.5346 0.7311
No log 6.75 108 0.5276 0.5645 0.5276 0.7263
No log 6.875 110 0.5316 0.5556 0.5315 0.7291
No log 7.0 112 0.5403 0.5574 0.5403 0.7351
No log 7.125 114 0.5634 0.5662 0.5634 0.7506
No log 7.25 116 0.5819 0.5437 0.5819 0.7628
No log 7.375 118 0.5998 0.5369 0.5998 0.7744
No log 7.5 120 0.6538 0.5181 0.6538 0.8086
No log 7.625 122 0.6922 0.5254 0.6922 0.8320
No log 7.75 124 0.6880 0.5224 0.6880 0.8295
No log 7.875 126 0.6155 0.5394 0.6155 0.7845
No log 8.0 128 0.5554 0.5606 0.5554 0.7452
No log 8.125 130 0.5438 0.5538 0.5438 0.7374
No log 8.25 132 0.5602 0.5332 0.5602 0.7485
No log 8.375 134 0.5545 0.5656 0.5545 0.7446
No log 8.5 136 0.5441 0.5554 0.5441 0.7377
No log 8.625 138 0.5506 0.5505 0.5506 0.7420
No log 8.75 140 0.6015 0.5592 0.6015 0.7756
No log 8.875 142 0.6260 0.5277 0.6260 0.7912
No log 9.0 144 0.6060 0.5638 0.6060 0.7785
No log 9.125 146 0.5820 0.5703 0.5820 0.7629
No log 9.25 148 0.5572 0.5672 0.5572 0.7464
No log 9.375 150 0.5470 0.5554 0.5470 0.7396
No log 9.5 152 0.5447 0.5554 0.5447 0.7380
No log 9.625 154 0.5519 0.5355 0.5519 0.7429
No log 9.75 156 0.5584 0.5489 0.5584 0.7473
No log 9.875 158 0.5799 0.5588 0.5799 0.7615
No log 10.0 160 0.6146 0.5654 0.6146 0.7839
No log 10.125 162 0.6211 0.5695 0.6211 0.7881
No log 10.25 164 0.6220 0.5741 0.6220 0.7887
No log 10.375 166 0.6190 0.5493 0.6190 0.7868
No log 10.5 168 0.5992 0.5538 0.5992 0.7741
No log 10.625 170 0.5937 0.5846 0.5937 0.7705
No log 10.75 172 0.6018 0.5837 0.6018 0.7758
No log 10.875 174 0.6167 0.5667 0.6167 0.7853
No log 11.0 176 0.6210 0.5604 0.6210 0.7880
No log 11.125 178 0.6088 0.5837 0.6088 0.7803
No log 11.25 180 0.5987 0.5477 0.5987 0.7738
No log 11.375 182 0.6052 0.5336 0.6052 0.7779
No log 11.5 184 0.6136 0.5147 0.6136 0.7833
No log 11.625 186 0.6244 0.5226 0.6244 0.7902
No log 11.75 188 0.6395 0.4828 0.6395 0.7997
No log 11.875 190 0.6648 0.5217 0.6648 0.8154
No log 12.0 192 0.6611 0.5118 0.6611 0.8131
No log 12.125 194 0.6376 0.5506 0.6376 0.7985
No log 12.25 196 0.6257 0.5523 0.6257 0.7910
No log 12.375 198 0.6347 0.5430 0.6347 0.7967
No log 12.5 200 0.6547 0.5479 0.6547 0.8091
No log 12.625 202 0.6959 0.5181 0.6959 0.8342
No log 12.75 204 0.7520 0.4933 0.7520 0.8672
No log 12.875 206 0.7360 0.4989 0.7360 0.8579
No log 13.0 208 0.6761 0.5098 0.6761 0.8222
No log 13.125 210 0.6396 0.5357 0.6396 0.7997
No log 13.25 212 0.6387 0.5693 0.6387 0.7992
No log 13.375 214 0.6413 0.5779 0.6413 0.8008
No log 13.5 216 0.6453 0.5658 0.6453 0.8033
No log 13.625 218 0.6576 0.5726 0.6576 0.8109
No log 13.75 220 0.7048 0.5206 0.7048 0.8395
No log 13.875 222 0.7529 0.5087 0.7529 0.8677
No log 14.0 224 0.7418 0.5010 0.7418 0.8613
No log 14.125 226 0.6879 0.5154 0.6879 0.8294
No log 14.25 228 0.6573 0.5693 0.6573 0.8107
No log 14.375 230 0.6579 0.5491 0.6579 0.8111
No log 14.5 232 0.6597 0.5355 0.6597 0.8122
No log 14.625 234 0.6618 0.5328 0.6618 0.8135
No log 14.75 236 0.6871 0.5116 0.6871 0.8289
No log 14.875 238 0.7023 0.5297 0.7023 0.8380
No log 15.0 240 0.6880 0.5884 0.6880 0.8294
No log 15.125 242 0.6775 0.5510 0.6775 0.8231
No log 15.25 244 0.6749 0.5389 0.6749 0.8215
No log 15.375 246 0.6747 0.5329 0.6747 0.8214
No log 15.5 248 0.6778 0.5203 0.6778 0.8233
No log 15.625 250 0.6688 0.5475 0.6688 0.8178
No log 15.75 252 0.6649 0.5276 0.6649 0.8154
No log 15.875 254 0.6744 0.5509 0.6744 0.8212
No log 16.0 256 0.6801 0.5178 0.6801 0.8247
No log 16.125 258 0.6617 0.5494 0.6617 0.8134
No log 16.25 260 0.6524 0.5312 0.6524 0.8077
No log 16.375 262 0.6534 0.54 0.6534 0.8083
No log 16.5 264 0.6609 0.4974 0.6609 0.8130
No log 16.625 266 0.6650 0.5533 0.6650 0.8155
No log 16.75 268 0.6698 0.5323 0.6698 0.8184
No log 16.875 270 0.6756 0.5370 0.6756 0.8220
No log 17.0 272 0.6862 0.4978 0.6862 0.8284
No log 17.125 274 0.6925 0.5110 0.6925 0.8321
No log 17.25 276 0.7063 0.5255 0.7063 0.8404
No log 17.375 278 0.7117 0.5333 0.7117 0.8436
No log 17.5 280 0.7237 0.5300 0.7237 0.8507
No log 17.625 282 0.7337 0.5352 0.7337 0.8565
No log 17.75 284 0.7411 0.5519 0.7411 0.8609
No log 17.875 286 0.7416 0.5466 0.7416 0.8612
No log 18.0 288 0.7428 0.5268 0.7428 0.8619
No log 18.125 290 0.7442 0.5426 0.7442 0.8627
No log 18.25 292 0.7380 0.5575 0.7380 0.8591
No log 18.375 294 0.7490 0.5452 0.7490 0.8654
No log 18.5 296 0.7480 0.54 0.7480 0.8649
No log 18.625 298 0.7409 0.5412 0.7409 0.8608
No log 18.75 300 0.7245 0.5437 0.7245 0.8512
No log 18.875 302 0.7186 0.5781 0.7186 0.8477
No log 19.0 304 0.7233 0.5672 0.7233 0.8505
No log 19.125 306 0.7233 0.5672 0.7233 0.8505
No log 19.25 308 0.7177 0.5832 0.7177 0.8472
No log 19.375 310 0.7330 0.5452 0.7330 0.8562
No log 19.5 312 0.7426 0.5260 0.7426 0.8618
No log 19.625 314 0.7511 0.4968 0.7511 0.8667
No log 19.75 316 0.7496 0.4933 0.7496 0.8658
No log 19.875 318 0.7371 0.5304 0.7371 0.8585
No log 20.0 320 0.7218 0.5750 0.7218 0.8496
No log 20.125 322 0.7118 0.5836 0.7118 0.8437
No log 20.25 324 0.7106 0.5672 0.7106 0.8430
No log 20.375 326 0.7006 0.5536 0.7006 0.8370
No log 20.5 328 0.6977 0.5659 0.6977 0.8353
No log 20.625 330 0.6954 0.5659 0.6954 0.8339
No log 20.75 332 0.6896 0.5605 0.6896 0.8304
No log 20.875 334 0.6931 0.5139 0.6931 0.8325
No log 21.0 336 0.7008 0.5146 0.7008 0.8371
No log 21.125 338 0.7030 0.5140 0.7030 0.8385
No log 21.25 340 0.6981 0.5140 0.6981 0.8355
No log 21.375 342 0.6928 0.5170 0.6928 0.8323
No log 21.5 344 0.6810 0.5292 0.6810 0.8252
No log 21.625 346 0.6814 0.5208 0.6814 0.8255
No log 21.75 348 0.6867 0.5548 0.6867 0.8287
No log 21.875 350 0.7016 0.5420 0.7016 0.8376
No log 22.0 352 0.7145 0.4986 0.7145 0.8453
No log 22.125 354 0.7404 0.5235 0.7404 0.8605
No log 22.25 356 0.7566 0.5090 0.7566 0.8698
No log 22.375 358 0.7696 0.5146 0.7696 0.8773
No log 22.5 360 0.7863 0.5140 0.7863 0.8867
No log 22.625 362 0.7784 0.5140 0.7784 0.8823
No log 22.75 364 0.7497 0.5074 0.7497 0.8659
No log 22.875 366 0.7421 0.5279 0.7421 0.8614
No log 23.0 368 0.7264 0.5236 0.7264 0.8523
No log 23.125 370 0.7229 0.5547 0.7229 0.8502
No log 23.25 372 0.7225 0.5437 0.7225 0.8500
No log 23.375 374 0.7283 0.5302 0.7283 0.8534
No log 23.5 376 0.7244 0.5033 0.7244 0.8511
No log 23.625 378 0.7332 0.5102 0.7332 0.8563
No log 23.75 380 0.7250 0.5155 0.7250 0.8515
No log 23.875 382 0.7286 0.5385 0.7286 0.8536
No log 24.0 384 0.7361 0.5482 0.7361 0.8579
No log 24.125 386 0.7425 0.5414 0.7425 0.8617
No log 24.25 388 0.7498 0.5639 0.7498 0.8659
No log 24.375 390 0.7336 0.5615 0.7336 0.8565
No log 24.5 392 0.7141 0.5604 0.7141 0.8451
No log 24.625 394 0.7160 0.5670 0.7160 0.8462
No log 24.75 396 0.7163 0.5448 0.7163 0.8464
No log 24.875 398 0.7103 0.5519 0.7103 0.8428
No log 25.0 400 0.6969 0.5478 0.6969 0.8348
No log 25.125 402 0.6936 0.5382 0.6936 0.8328
No log 25.25 404 0.6991 0.5495 0.6991 0.8361
No log 25.375 406 0.7149 0.5233 0.7149 0.8455
No log 25.5 408 0.7373 0.5387 0.7373 0.8586
No log 25.625 410 0.7521 0.5068 0.7521 0.8672
No log 25.75 412 0.7754 0.5380 0.7754 0.8805
No log 25.875 414 0.7813 0.5289 0.7813 0.8839
No log 26.0 416 0.7560 0.5338 0.7560 0.8695
No log 26.125 418 0.7379 0.5385 0.7379 0.8590
No log 26.25 420 0.7302 0.5826 0.7302 0.8545
No log 26.375 422 0.7306 0.5891 0.7306 0.8547
No log 26.5 424 0.7309 0.5445 0.7309 0.8549
No log 26.625 426 0.7550 0.5371 0.7550 0.8689
No log 26.75 428 0.7853 0.5174 0.7853 0.8862
No log 26.875 430 0.7989 0.5113 0.7989 0.8938
No log 27.0 432 0.8049 0.5261 0.8049 0.8972
No log 27.125 434 0.8020 0.5060 0.8020 0.8955
No log 27.25 436 0.8036 0.5067 0.8036 0.8964
No log 27.375 438 0.7937 0.5130 0.7937 0.8909
No log 27.5 440 0.7713 0.5050 0.7713 0.8783
No log 27.625 442 0.7477 0.5183 0.7477 0.8647
No log 27.75 444 0.7221 0.5266 0.7221 0.8497
No log 27.875 446 0.7085 0.5376 0.7085 0.8417
No log 28.0 448 0.7032 0.5373 0.7032 0.8386
No log 28.125 450 0.7121 0.5370 0.7121 0.8438
No log 28.25 452 0.7326 0.4864 0.7326 0.8559
No log 28.375 454 0.7439 0.4830 0.7439 0.8625
No log 28.5 456 0.7322 0.4805 0.7322 0.8557
No log 28.625 458 0.7127 0.5116 0.7127 0.8442
No log 28.75 460 0.6984 0.5157 0.6984 0.8357
No log 28.875 462 0.6991 0.5017 0.6991 0.8361
No log 29.0 464 0.7008 0.5099 0.7008 0.8371
No log 29.125 466 0.7135 0.5017 0.7135 0.8447
No log 29.25 468 0.7383 0.5668 0.7383 0.8593
No log 29.375 470 0.7735 0.5440 0.7735 0.8795
No log 29.5 472 0.7892 0.5438 0.7892 0.8883
No log 29.625 474 0.7711 0.5412 0.7711 0.8781
No log 29.75 476 0.7617 0.5311 0.7617 0.8727
No log 29.875 478 0.7569 0.5348 0.7569 0.8700
No log 30.0 480 0.7712 0.5362 0.7712 0.8782
No log 30.125 482 0.7903 0.5078 0.7903 0.8890
No log 30.25 484 0.7980 0.4973 0.7980 0.8933
No log 30.375 486 0.8008 0.5124 0.8008 0.8949
No log 30.5 488 0.8071 0.4935 0.8071 0.8984
No log 30.625 490 0.7983 0.4744 0.7983 0.8935
No log 30.75 492 0.7839 0.4749 0.7839 0.8854
No log 30.875 494 0.7546 0.4950 0.7546 0.8687
No log 31.0 496 0.7221 0.5022 0.7221 0.8498
No log 31.125 498 0.6994 0.5113 0.6994 0.8363
0.5278 31.25 500 0.6783 0.5239 0.6783 0.8236
0.5278 31.375 502 0.6682 0.5341 0.6682 0.8174
0.5278 31.5 504 0.6790 0.5214 0.6790 0.8240
0.5278 31.625 506 0.7003 0.5137 0.7003 0.8368
0.5278 31.75 508 0.7114 0.5179 0.7114 0.8434
0.5278 31.875 510 0.7093 0.5383 0.7093 0.8422
0.5278 32.0 512 0.7004 0.5643 0.7004 0.8369
0.5278 32.125 514 0.7079 0.5747 0.7079 0.8414
0.5278 32.25 516 0.7176 0.5750 0.7176 0.8471
0.5278 32.375 518 0.7380 0.5657 0.7380 0.8590
0.5278 32.5 520 0.7836 0.5531 0.7836 0.8852
0.5278 32.625 522 0.8031 0.5098 0.8031 0.8962
0.5278 32.75 524 0.7886 0.5307 0.7886 0.8880
0.5278 32.875 526 0.7621 0.5329 0.7621 0.8730
0.5278 33.0 528 0.7519 0.5347 0.7519 0.8671
0.5278 33.125 530 0.7512 0.5444 0.7512 0.8667
0.5278 33.25 532 0.7379 0.5409 0.7379 0.8590
0.5278 33.375 534 0.7222 0.5288 0.7222 0.8498
0.5278 33.5 536 0.7198 0.5317 0.7198 0.8484
0.5278 33.625 538 0.7317 0.5572 0.7317 0.8554
0.5278 33.75 540 0.7399 0.5572 0.7399 0.8602
0.5278 33.875 542 0.7481 0.4933 0.7481 0.8649
0.5278 34.0 544 0.7434 0.5144 0.7434 0.8622
0.5278 34.125 546 0.7470 0.5144 0.7470 0.8643
0.5278 34.25 548 0.7502 0.5134 0.7502 0.8661
0.5278 34.375 550 0.7516 0.5185 0.7516 0.8670

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.1+cu121
  • Datasets 3.2.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERTFreeze_run1_AugV5_k6_task2_organization

Finetuned
(4023)
this model