ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k16_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7395
  • Qwk: 0.5555
  • Mse: 0.7395
  • Rmse: 0.8599

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0385 2 3.8461 -0.0150 3.8461 1.9611
No log 0.0769 4 2.0861 0.1002 2.0861 1.4443
No log 0.1154 6 1.2587 0.0672 1.2587 1.1219
No log 0.1538 8 1.1006 0.3003 1.1006 1.0491
No log 0.1923 10 1.1331 0.1471 1.1331 1.0645
No log 0.2308 12 1.1765 0.0201 1.1765 1.0847
No log 0.2692 14 1.2198 0.0201 1.2198 1.1044
No log 0.3077 16 1.2113 0.0201 1.2113 1.1006
No log 0.3462 18 1.0048 0.2268 1.0048 1.0024
No log 0.3846 20 0.9699 0.3730 0.9699 0.9848
No log 0.4231 22 0.9598 0.3709 0.9598 0.9797
No log 0.4615 24 0.9873 0.3392 0.9873 0.9936
No log 0.5 26 1.0922 0.1725 1.0922 1.0451
No log 0.5385 28 1.1274 0.1389 1.1274 1.0618
No log 0.5769 30 1.0709 0.1361 1.0709 1.0349
No log 0.6154 32 1.0740 0.1509 1.0740 1.0363
No log 0.6538 34 1.0658 0.2246 1.0658 1.0324
No log 0.6923 36 1.1220 0.2004 1.1220 1.0592
No log 0.7308 38 1.0736 0.1953 1.0736 1.0362
No log 0.7692 40 1.0547 0.1211 1.0547 1.0270
No log 0.8077 42 1.1509 0.1296 1.1509 1.0728
No log 0.8462 44 1.5457 0.1598 1.5457 1.2432
No log 0.8846 46 1.9034 0.1187 1.9034 1.3797
No log 0.9231 48 1.8813 0.0535 1.8813 1.3716
No log 0.9615 50 1.6573 0.0030 1.6573 1.2874
No log 1.0 52 1.3133 0.0996 1.3133 1.1460
No log 1.0385 54 1.0435 0.1832 1.0435 1.0215
No log 1.0769 56 0.9472 0.3332 0.9472 0.9732
No log 1.1154 58 0.9093 0.375 0.9093 0.9536
No log 1.1538 60 1.0423 0.2827 1.0423 1.0209
No log 1.1923 62 1.1314 0.2294 1.1314 1.0637
No log 1.2308 64 1.0833 0.3090 1.0833 1.0408
No log 1.2692 66 0.9925 0.2465 0.9925 0.9963
No log 1.3077 68 0.9234 0.2841 0.9234 0.9609
No log 1.3462 70 0.9127 0.3310 0.9127 0.9554
No log 1.3846 72 0.9342 0.2465 0.9342 0.9665
No log 1.4231 74 0.9685 0.3187 0.9685 0.9841
No log 1.4615 76 0.9408 0.3181 0.9408 0.9699
No log 1.5 78 0.9103 0.3353 0.9103 0.9541
No log 1.5385 80 0.8648 0.375 0.8648 0.9299
No log 1.5769 82 0.8378 0.3876 0.8378 0.9153
No log 1.6154 84 0.8297 0.4110 0.8297 0.9109
No log 1.6538 86 0.8330 0.3795 0.8330 0.9127
No log 1.6923 88 0.8630 0.4198 0.8630 0.9290
No log 1.7308 90 0.8654 0.4466 0.8654 0.9302
No log 1.7692 92 0.8291 0.4359 0.8291 0.9106
No log 1.8077 94 0.7640 0.4630 0.7640 0.8740
No log 1.8462 96 0.7202 0.5866 0.7202 0.8487
No log 1.8846 98 0.7451 0.5291 0.7451 0.8632
No log 1.9231 100 0.8192 0.5400 0.8192 0.9051
No log 1.9615 102 0.7729 0.4519 0.7729 0.8791
No log 2.0 104 0.7798 0.4368 0.7798 0.8831
No log 2.0385 106 0.7724 0.5197 0.7724 0.8789
No log 2.0769 108 0.6877 0.5777 0.6877 0.8293
No log 2.1154 110 0.6803 0.6347 0.6803 0.8248
No log 2.1538 112 0.7311 0.5510 0.7311 0.8550
No log 2.1923 114 0.8115 0.5143 0.8115 0.9008
No log 2.2308 116 0.7970 0.5292 0.7970 0.8928
No log 2.2692 118 0.7368 0.5292 0.7368 0.8584
No log 2.3077 120 0.7281 0.5291 0.7281 0.8533
No log 2.3462 122 0.7857 0.4694 0.7857 0.8864
No log 2.3846 124 0.7174 0.5052 0.7174 0.8470
No log 2.4231 126 0.6796 0.5740 0.6796 0.8244
No log 2.4615 128 0.7188 0.5353 0.7188 0.8478
No log 2.5 130 0.9919 0.4878 0.9919 0.9959
No log 2.5385 132 1.0541 0.4773 1.0541 1.0267
No log 2.5769 134 0.7825 0.5052 0.7825 0.8846
No log 2.6154 136 0.7114 0.5300 0.7114 0.8435
No log 2.6538 138 0.7523 0.4780 0.7523 0.8674
No log 2.6923 140 0.7422 0.4156 0.7422 0.8615
No log 2.7308 142 0.8448 0.4568 0.8448 0.9191
No log 2.7692 144 0.9138 0.4681 0.9138 0.9559
No log 2.8077 146 0.8435 0.4496 0.8435 0.9184
No log 2.8462 148 0.7284 0.5552 0.7284 0.8535
No log 2.8846 150 0.7287 0.5891 0.7287 0.8537
No log 2.9231 152 0.8115 0.4127 0.8115 0.9009
No log 2.9615 154 0.9478 0.4167 0.9478 0.9735
No log 3.0 156 1.0743 0.4467 1.0743 1.0365
No log 3.0385 158 0.9624 0.4792 0.9624 0.9810
No log 3.0769 160 0.7687 0.4722 0.7687 0.8767
No log 3.1154 162 0.7445 0.5776 0.7445 0.8628
No log 3.1538 164 0.8504 0.4902 0.8504 0.9221
No log 3.1923 166 0.9118 0.4885 0.9118 0.9549
No log 3.2308 168 0.9488 0.4777 0.9488 0.9741
No log 3.2692 170 0.9089 0.4885 0.9089 0.9533
No log 3.3077 172 0.8478 0.4825 0.8478 0.9207
No log 3.3462 174 0.7354 0.5731 0.7354 0.8576
No log 3.3846 176 0.7394 0.5703 0.7394 0.8599
No log 3.4231 178 0.7617 0.5077 0.7617 0.8728
No log 3.4615 180 0.9536 0.4463 0.9536 0.9765
No log 3.5 182 0.9674 0.4463 0.9674 0.9836
No log 3.5385 184 0.8057 0.5006 0.8057 0.8976
No log 3.5769 186 0.7009 0.5671 0.7009 0.8372
No log 3.6154 188 0.7114 0.5327 0.7114 0.8435
No log 3.6538 190 0.6981 0.4789 0.6981 0.8355
No log 3.6923 192 0.7722 0.5383 0.7722 0.8787
No log 3.7308 194 0.9349 0.4877 0.9349 0.9669
No log 3.7692 196 0.8755 0.5220 0.8755 0.9357
No log 3.8077 198 0.7592 0.5213 0.7592 0.8713
No log 3.8462 200 0.7536 0.5971 0.7536 0.8681
No log 3.8846 202 0.7401 0.5971 0.7401 0.8603
No log 3.9231 204 0.7243 0.5407 0.7243 0.8511
No log 3.9615 206 0.7169 0.5510 0.7169 0.8467
No log 4.0 208 0.7051 0.5708 0.7051 0.8397
No log 4.0385 210 0.7429 0.5447 0.7429 0.8619
No log 4.0769 212 0.7121 0.5688 0.7121 0.8439
No log 4.1154 214 0.6920 0.6311 0.6920 0.8319
No log 4.1538 216 0.7040 0.6239 0.7040 0.8390
No log 4.1923 218 0.7055 0.6035 0.7055 0.8400
No log 4.2308 220 0.7299 0.5413 0.7299 0.8544
No log 4.2692 222 0.6955 0.5921 0.6955 0.8340
No log 4.3077 224 0.7170 0.5312 0.7170 0.8468
No log 4.3462 226 0.7529 0.4998 0.7529 0.8677
No log 4.3846 228 0.7069 0.6349 0.7069 0.8408
No log 4.4231 230 0.7509 0.5810 0.7509 0.8665
No log 4.4615 232 0.8805 0.4264 0.8805 0.9384
No log 4.5 234 0.8602 0.5115 0.8602 0.9275
No log 4.5385 236 0.7505 0.4824 0.7505 0.8663
No log 4.5769 238 0.7036 0.5127 0.7036 0.8388
No log 4.6154 240 0.7092 0.5261 0.7092 0.8422
No log 4.6538 242 0.7248 0.4473 0.7248 0.8514
No log 4.6923 244 0.7193 0.5534 0.7193 0.8481
No log 4.7308 246 0.7155 0.5730 0.7155 0.8459
No log 4.7692 248 0.6869 0.6464 0.6869 0.8288
No log 4.8077 250 0.6815 0.6359 0.6815 0.8255
No log 4.8462 252 0.6772 0.6774 0.6772 0.8229
No log 4.8846 254 0.6683 0.6430 0.6683 0.8175
No log 4.9231 256 0.6858 0.5666 0.6858 0.8281
No log 4.9615 258 0.7167 0.5219 0.7167 0.8466
No log 5.0 260 0.6864 0.5241 0.6864 0.8285
No log 5.0385 262 0.7157 0.5134 0.7157 0.8460
No log 5.0769 264 0.7389 0.5124 0.7389 0.8596
No log 5.1154 266 0.7057 0.5006 0.7057 0.8400
No log 5.1538 268 0.6583 0.5846 0.6583 0.8114
No log 5.1923 270 0.6507 0.6547 0.6507 0.8067
No log 5.2308 272 0.7290 0.5006 0.7290 0.8538
No log 5.2692 274 0.8245 0.4988 0.8245 0.9080
No log 5.3077 276 0.8793 0.4976 0.8793 0.9377
No log 5.3462 278 0.7557 0.5306 0.7557 0.8693
No log 5.3846 280 0.6753 0.5163 0.6753 0.8218
No log 5.4231 282 0.6316 0.5274 0.6316 0.7947
No log 5.4615 284 0.6423 0.5033 0.6423 0.8015
No log 5.5 286 0.6571 0.4847 0.6571 0.8106
No log 5.5385 288 0.6784 0.5056 0.6784 0.8237
No log 5.5769 290 0.6674 0.5877 0.6674 0.8169
No log 5.6154 292 0.7269 0.5345 0.7269 0.8526
No log 5.6538 294 0.7201 0.5547 0.7201 0.8486
No log 5.6923 296 0.6597 0.6606 0.6597 0.8122
No log 5.7308 298 0.6515 0.6683 0.6515 0.8071
No log 5.7692 300 0.6539 0.6394 0.6539 0.8086
No log 5.8077 302 0.6533 0.6850 0.6533 0.8083
No log 5.8462 304 0.6695 0.6232 0.6695 0.8182
No log 5.8846 306 0.6765 0.6344 0.6765 0.8225
No log 5.9231 308 0.7067 0.5629 0.7067 0.8407
No log 5.9615 310 0.7496 0.5477 0.7496 0.8658
No log 6.0 312 0.7424 0.5648 0.7424 0.8616
No log 6.0385 314 0.7907 0.5860 0.7907 0.8892
No log 6.0769 316 0.8670 0.5359 0.8670 0.9311
No log 6.1154 318 0.7814 0.5759 0.7814 0.8840
No log 6.1538 320 0.7016 0.5614 0.7016 0.8376
No log 6.1923 322 0.7004 0.4960 0.7004 0.8369
No log 6.2308 324 0.6860 0.4988 0.6860 0.8283
No log 6.2692 326 0.6867 0.4995 0.6867 0.8287
No log 6.3077 328 0.6738 0.5153 0.6738 0.8209
No log 6.3462 330 0.6993 0.5999 0.6993 0.8362
No log 6.3846 332 0.6888 0.5771 0.6888 0.8300
No log 6.4231 334 0.6910 0.5317 0.6910 0.8313
No log 6.4615 336 0.6985 0.5740 0.6985 0.8358
No log 6.5 338 0.7115 0.5677 0.7115 0.8435
No log 6.5385 340 0.7395 0.5140 0.7395 0.8599
No log 6.5769 342 0.7055 0.5150 0.7055 0.8400
No log 6.6154 344 0.6578 0.5121 0.6578 0.8111
No log 6.6538 346 0.6570 0.5188 0.6570 0.8106
No log 6.6923 348 0.6700 0.5759 0.6700 0.8185
No log 6.7308 350 0.6788 0.5759 0.6788 0.8239
No log 6.7692 352 0.6741 0.5534 0.6741 0.8210
No log 6.8077 354 0.6886 0.5279 0.6886 0.8298
No log 6.8462 356 0.6804 0.5210 0.6804 0.8249
No log 6.8846 358 0.6815 0.5565 0.6815 0.8255
No log 6.9231 360 0.6691 0.4947 0.6691 0.8180
No log 6.9615 362 0.7046 0.5565 0.7046 0.8394
No log 7.0 364 0.6880 0.5207 0.6880 0.8295
No log 7.0385 366 0.6841 0.5855 0.6841 0.8271
No log 7.0769 368 0.7425 0.5788 0.7425 0.8617
No log 7.1154 370 0.7421 0.5595 0.7421 0.8615
No log 7.1538 372 0.6984 0.5405 0.6984 0.8357
No log 7.1923 374 0.7348 0.5093 0.7348 0.8572
No log 7.2308 376 0.7343 0.4849 0.7343 0.8569
No log 7.2692 378 0.7234 0.5127 0.7234 0.8505
No log 7.3077 380 0.7612 0.4491 0.7612 0.8724
No log 7.3462 382 0.8132 0.4815 0.8132 0.9018
No log 7.3846 384 0.8203 0.4575 0.8203 0.9057
No log 7.4231 386 0.7543 0.4478 0.7543 0.8685
No log 7.4615 388 0.7237 0.4882 0.7237 0.8507
No log 7.5 390 0.7182 0.4745 0.7182 0.8475
No log 7.5385 392 0.7357 0.4725 0.7357 0.8577
No log 7.5769 394 0.7653 0.4812 0.7653 0.8748
No log 7.6154 396 0.7806 0.4812 0.7806 0.8835
No log 7.6538 398 0.7393 0.4599 0.7393 0.8598
No log 7.6923 400 0.7127 0.4745 0.7127 0.8442
No log 7.7308 402 0.7120 0.4995 0.7120 0.8438
No log 7.7692 404 0.7014 0.5010 0.7014 0.8375
No log 7.8077 406 0.7140 0.5010 0.7140 0.8450
No log 7.8462 408 0.7273 0.4378 0.7273 0.8528
No log 7.8846 410 0.7353 0.4378 0.7353 0.8575
No log 7.9231 412 0.7436 0.4378 0.7436 0.8623
No log 7.9615 414 0.7445 0.4353 0.7445 0.8629
No log 8.0 416 0.7520 0.4946 0.7520 0.8672
No log 8.0385 418 0.7651 0.4818 0.7651 0.8747
No log 8.0769 420 0.7640 0.4818 0.7640 0.8741
No log 8.1154 422 0.7387 0.5175 0.7387 0.8595
No log 8.1538 424 0.7269 0.5304 0.7269 0.8526
No log 8.1923 426 0.7082 0.5434 0.7082 0.8416
No log 8.2308 428 0.6996 0.5274 0.6996 0.8364
No log 8.2692 430 0.6986 0.4867 0.6986 0.8358
No log 8.3077 432 0.7275 0.4214 0.7275 0.8529
No log 8.3462 434 0.7243 0.4214 0.7243 0.8511
No log 8.3846 436 0.6842 0.4746 0.6842 0.8272
No log 8.4231 438 0.6626 0.5631 0.6626 0.8140
No log 8.4615 440 0.6660 0.5797 0.6660 0.8161
No log 8.5 442 0.7493 0.5033 0.7493 0.8656
No log 8.5385 444 0.7793 0.5128 0.7793 0.8828
No log 8.5769 446 0.6837 0.5729 0.6837 0.8268
No log 8.6154 448 0.6416 0.5988 0.6416 0.8010
No log 8.6538 450 0.6620 0.5890 0.6620 0.8137
No log 8.6923 452 0.6378 0.6241 0.6378 0.7986
No log 8.7308 454 0.6558 0.6102 0.6558 0.8098
No log 8.7692 456 0.6992 0.5607 0.6992 0.8362
No log 8.8077 458 0.6666 0.6187 0.6666 0.8165
No log 8.8462 460 0.6448 0.6409 0.6448 0.8030
No log 8.8846 462 0.6387 0.6028 0.6387 0.7992
No log 8.9231 464 0.6812 0.5062 0.6812 0.8254
No log 8.9615 466 0.7112 0.5666 0.7112 0.8434
No log 9.0 468 0.6855 0.5279 0.6855 0.8280
No log 9.0385 470 0.6509 0.5552 0.6509 0.8068
No log 9.0769 472 0.6509 0.5442 0.6509 0.8068
No log 9.1154 474 0.6579 0.4954 0.6579 0.8111
No log 9.1538 476 0.6567 0.5305 0.6567 0.8104
No log 9.1923 478 0.6514 0.4759 0.6514 0.8071
No log 9.2308 480 0.6545 0.5002 0.6545 0.8090
No log 9.2692 482 0.6575 0.5832 0.6575 0.8108
No log 9.3077 484 0.7036 0.5223 0.7036 0.8388
No log 9.3462 486 0.7738 0.5486 0.7738 0.8797
No log 9.3846 488 0.7100 0.5769 0.7100 0.8426
No log 9.4231 490 0.6142 0.7041 0.6142 0.7837
No log 9.4615 492 0.6153 0.6559 0.6153 0.7844
No log 9.5 494 0.6051 0.6653 0.6051 0.7779
No log 9.5385 496 0.6178 0.6772 0.6178 0.7860
No log 9.5769 498 0.7008 0.5995 0.7008 0.8371
0.2853 9.6154 500 0.6836 0.5877 0.6836 0.8268
0.2853 9.6538 502 0.6329 0.6013 0.6329 0.7955
0.2853 9.6923 504 0.6295 0.5450 0.6295 0.7934
0.2853 9.7308 506 0.6345 0.6198 0.6345 0.7966
0.2853 9.7692 508 0.6612 0.5666 0.6612 0.8131
0.2853 9.8077 510 0.6523 0.5777 0.6523 0.8077
0.2853 9.8462 512 0.5922 0.6291 0.5922 0.7695
0.2853 9.8846 514 0.5793 0.6400 0.5793 0.7611
0.2853 9.9231 516 0.5843 0.6400 0.5843 0.7644
0.2853 9.9615 518 0.5878 0.6252 0.5878 0.7667
0.2853 10.0 520 0.5938 0.6301 0.5938 0.7706
0.2853 10.0385 522 0.5997 0.6317 0.5997 0.7744
0.2853 10.0769 524 0.6146 0.6217 0.6146 0.7840
0.2853 10.1154 526 0.6165 0.6354 0.6165 0.7852
0.2853 10.1538 528 0.6253 0.6167 0.6253 0.7908
0.2853 10.1923 530 0.6490 0.5552 0.6490 0.8056
0.2853 10.2308 532 0.6490 0.5570 0.6490 0.8056
0.2853 10.2692 534 0.6604 0.5450 0.6604 0.8126
0.2853 10.3077 536 0.6876 0.5186 0.6876 0.8292
0.2853 10.3462 538 0.7395 0.5555 0.7395 0.8599

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k16_task5_organization

Finetuned
(4019)
this model