ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k14_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8333
  • Qwk: 0.4532
  • Mse: 0.8333
  • Rmse: 0.9128

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0465 2 4.5590 -0.0132 4.5590 2.1352
No log 0.0930 4 2.6215 0.0274 2.6215 1.6191
No log 0.1395 6 2.1014 -0.0634 2.1014 1.4496
No log 0.1860 8 1.5274 0.0682 1.5274 1.2359
No log 0.2326 10 1.3009 0.0050 1.3009 1.1406
No log 0.2791 12 1.3325 0.1472 1.3325 1.1543
No log 0.3256 14 1.4485 0.1106 1.4485 1.2035
No log 0.3721 16 1.2890 0.1136 1.2890 1.1353
No log 0.4186 18 1.2099 0.1753 1.2099 1.0999
No log 0.4651 20 1.1414 0.2628 1.1414 1.0684
No log 0.5116 22 1.1356 0.1995 1.1356 1.0657
No log 0.5581 24 1.1563 0.1812 1.1563 1.0753
No log 0.6047 26 1.1275 0.2520 1.1275 1.0618
No log 0.6512 28 1.1237 0.2416 1.1237 1.0600
No log 0.6977 30 1.3092 0.0290 1.3092 1.1442
No log 0.7442 32 1.6587 0.0972 1.6587 1.2879
No log 0.7907 34 1.4808 0.0878 1.4808 1.2169
No log 0.8372 36 1.0764 0.2378 1.0764 1.0375
No log 0.8837 38 1.1760 0.2119 1.1760 1.0844
No log 0.9302 40 1.2321 0.2454 1.2321 1.1100
No log 0.9767 42 1.1000 0.1343 1.1000 1.0488
No log 1.0233 44 0.9928 0.3356 0.9928 0.9964
No log 1.0698 46 1.3474 0.2112 1.3474 1.1608
No log 1.1163 48 1.6458 0.1758 1.6458 1.2829
No log 1.1628 50 1.4960 0.1364 1.4960 1.2231
No log 1.2093 52 1.2728 0.2067 1.2728 1.1282
No log 1.2558 54 0.9945 0.4557 0.9945 0.9972
No log 1.3023 56 0.9582 0.2351 0.9582 0.9789
No log 1.3488 58 1.1087 0.2781 1.1087 1.0530
No log 1.3953 60 1.1919 0.3261 1.1919 1.0917
No log 1.4419 62 0.9973 0.1902 0.9973 0.9987
No log 1.4884 64 0.8402 0.4036 0.8402 0.9166
No log 1.5349 66 0.9349 0.4379 0.9349 0.9669
No log 1.5814 68 0.8776 0.4379 0.8776 0.9368
No log 1.6279 70 0.7821 0.4295 0.7821 0.8844
No log 1.6744 72 1.3755 0.2540 1.3755 1.1728
No log 1.7209 74 1.4784 0.2672 1.4784 1.2159
No log 1.7674 76 0.9516 0.6015 0.9516 0.9755
No log 1.8140 78 0.9469 0.3636 0.9469 0.9731
No log 1.8605 80 1.2864 0.3304 1.2864 1.1342
No log 1.9070 82 1.1093 0.3989 1.1093 1.0532
No log 1.9535 84 0.8032 0.4846 0.8032 0.8962
No log 2.0 86 1.1049 0.4876 1.1049 1.0511
No log 2.0465 88 1.1338 0.4704 1.1338 1.0648
No log 2.0930 90 0.8739 0.5294 0.8739 0.9348
No log 2.1395 92 0.8654 0.4699 0.8654 0.9303
No log 2.1860 94 0.9298 0.4361 0.9298 0.9643
No log 2.2326 96 0.7870 0.5686 0.7870 0.8871
No log 2.2791 98 0.8074 0.5405 0.8074 0.8986
No log 2.3256 100 0.8089 0.5532 0.8089 0.8994
No log 2.3721 102 0.7256 0.5479 0.7256 0.8518
No log 2.4186 104 0.7107 0.5984 0.7107 0.8430
No log 2.4651 106 0.8181 0.6234 0.8181 0.9045
No log 2.5116 108 0.7202 0.5843 0.7202 0.8486
No log 2.5581 110 0.8258 0.5769 0.8258 0.9087
No log 2.6047 112 0.9466 0.5506 0.9466 0.9729
No log 2.6512 114 0.7999 0.5814 0.7999 0.8944
No log 2.6977 116 0.7498 0.5942 0.7498 0.8659
No log 2.7442 118 0.7850 0.5731 0.7850 0.8860
No log 2.7907 120 1.0037 0.5410 1.0037 1.0018
No log 2.8372 122 0.9996 0.5429 0.9996 0.9998
No log 2.8837 124 0.7828 0.5806 0.7828 0.8848
No log 2.9302 126 0.7633 0.5495 0.7633 0.8737
No log 2.9767 128 0.8617 0.5229 0.8617 0.9283
No log 3.0233 130 0.7744 0.4848 0.7744 0.8800
No log 3.0698 132 0.8337 0.4732 0.8337 0.9131
No log 3.1163 134 0.9678 0.5507 0.9678 0.9838
No log 3.1628 136 0.8448 0.4732 0.8448 0.9191
No log 3.2093 138 0.7728 0.4583 0.7728 0.8791
No log 3.2558 140 0.8106 0.5485 0.8106 0.9003
No log 3.3023 142 0.8223 0.5060 0.8223 0.9068
No log 3.3488 144 0.7818 0.5226 0.7818 0.8842
No log 3.3953 146 0.7699 0.5138 0.7699 0.8775
No log 3.4419 148 0.7832 0.5102 0.7832 0.8850
No log 3.4884 150 0.8076 0.5220 0.8076 0.8986
No log 3.5349 152 0.8112 0.4722 0.8112 0.9006
No log 3.5814 154 0.8938 0.5505 0.8938 0.9454
No log 3.6279 156 1.0939 0.5238 1.0939 1.0459
No log 3.6744 158 0.9808 0.4936 0.9808 0.9904
No log 3.7209 160 0.8102 0.5331 0.8102 0.9001
No log 3.7674 162 0.8869 0.5636 0.8869 0.9418
No log 3.8140 164 0.8774 0.5375 0.8774 0.9367
No log 3.8605 166 0.8192 0.4846 0.8192 0.9051
No log 3.9070 168 0.7868 0.4122 0.7868 0.8870
No log 3.9535 170 0.8091 0.3987 0.8091 0.8995
No log 4.0 172 0.8347 0.3819 0.8347 0.9136
No log 4.0465 174 0.8093 0.4767 0.8093 0.8996
No log 4.0930 176 0.8163 0.4914 0.8163 0.9035
No log 4.1395 178 0.9742 0.5648 0.9742 0.9870
No log 4.1860 180 1.0468 0.5659 1.0468 1.0231
No log 4.2326 182 0.9695 0.5873 0.9695 0.9846
No log 4.2791 184 0.9716 0.5509 0.9716 0.9857
No log 4.3256 186 0.8349 0.4638 0.8349 0.9137
No log 4.3721 188 0.7616 0.5114 0.7616 0.8727
No log 4.4186 190 0.7505 0.5846 0.7505 0.8663
No log 4.4651 192 0.7417 0.6084 0.7417 0.8612
No log 4.5116 194 0.7710 0.4886 0.7710 0.8781
No log 4.5581 196 0.9273 0.5224 0.9273 0.9630
No log 4.6047 198 1.0416 0.5272 1.0416 1.0206
No log 4.6512 200 0.9234 0.4404 0.9234 0.9609
No log 4.6977 202 0.7530 0.4609 0.7530 0.8678
No log 4.7442 204 0.7715 0.5618 0.7715 0.8783
No log 4.7907 206 0.7983 0.5766 0.7983 0.8935
No log 4.8372 208 0.7950 0.5520 0.7950 0.8916
No log 4.8837 210 0.7843 0.4977 0.7843 0.8856
No log 4.9302 212 0.7950 0.4563 0.7950 0.8916
No log 4.9767 214 0.8006 0.4847 0.8006 0.8947
No log 5.0233 216 0.9065 0.4375 0.9065 0.9521
No log 5.0698 218 1.0078 0.4761 1.0078 1.0039
No log 5.1163 220 0.9202 0.4166 0.9202 0.9593
No log 5.1628 222 0.8100 0.4628 0.8100 0.9000
No log 5.2093 224 0.7842 0.4822 0.7842 0.8855
No log 5.2558 226 0.7857 0.4760 0.7857 0.8864
No log 5.3023 228 0.8098 0.4568 0.8098 0.8999
No log 5.3488 230 0.8835 0.5593 0.8835 0.9399
No log 5.3953 232 0.9950 0.4739 0.9950 0.9975
No log 5.4419 234 0.9349 0.4613 0.9349 0.9669
No log 5.4884 236 0.8532 0.4803 0.8532 0.9237
No log 5.5349 238 0.8380 0.4760 0.8380 0.9154
No log 5.5814 240 0.8349 0.4952 0.8349 0.9137
No log 5.6279 242 0.8327 0.5481 0.8327 0.9125
No log 5.6744 244 0.8117 0.5153 0.8117 0.9009
No log 5.7209 246 0.8071 0.5364 0.8071 0.8984
No log 5.7674 248 0.8717 0.4236 0.8717 0.9336
No log 5.8140 250 0.9791 0.5334 0.9791 0.9895
No log 5.8605 252 0.9403 0.4781 0.9403 0.9697
No log 5.9070 254 0.8173 0.4644 0.8173 0.9041
No log 5.9535 256 0.8268 0.4961 0.8268 0.9093
No log 6.0 258 0.8140 0.5110 0.8140 0.9022
No log 6.0465 260 0.8094 0.5268 0.8094 0.8997
No log 6.0930 262 0.8651 0.4482 0.8651 0.9301
No log 6.1395 264 0.9314 0.4533 0.9314 0.9651
No log 6.1860 266 0.8569 0.4579 0.8569 0.9257
No log 6.2326 268 0.8128 0.3991 0.8128 0.9015
No log 6.2791 270 0.7904 0.4123 0.7904 0.8891
No log 6.3256 272 0.7729 0.4123 0.7729 0.8791
No log 6.3721 274 0.7554 0.4479 0.7554 0.8691
No log 6.4186 276 0.7530 0.4435 0.7530 0.8678
No log 6.4651 278 0.7455 0.4435 0.7455 0.8634
No log 6.5116 280 0.7123 0.4479 0.7123 0.8440
No log 6.5581 282 0.7234 0.6141 0.7234 0.8505
No log 6.6047 284 0.7094 0.4813 0.7094 0.8423
No log 6.6512 286 0.7494 0.4164 0.7494 0.8657
No log 6.6977 288 0.7424 0.4165 0.7424 0.8616
No log 6.7442 290 0.7204 0.4519 0.7204 0.8488
No log 6.7907 292 0.6946 0.5012 0.6946 0.8334
No log 6.8372 294 0.7291 0.5683 0.7291 0.8539
No log 6.8837 296 0.7313 0.5683 0.7313 0.8552
No log 6.9302 298 0.6992 0.5856 0.6992 0.8362
No log 6.9767 300 0.6965 0.5856 0.6965 0.8346
No log 7.0233 302 0.7546 0.5736 0.7546 0.8687
No log 7.0698 304 0.7731 0.5513 0.7731 0.8792
No log 7.1163 306 0.7121 0.4695 0.7121 0.8438
No log 7.1628 308 0.7421 0.6215 0.7421 0.8615
No log 7.2093 310 0.7716 0.6416 0.7716 0.8784
No log 7.2558 312 0.7379 0.6148 0.7379 0.8590
No log 7.3023 314 0.7207 0.5911 0.7207 0.8490
No log 7.3488 316 0.7345 0.6014 0.7345 0.8570
No log 7.3953 318 0.7209 0.6225 0.7209 0.8490
No log 7.4419 320 0.7244 0.5214 0.7244 0.8511
No log 7.4884 322 0.7451 0.5012 0.7451 0.8632
No log 7.5349 324 0.8307 0.5532 0.8307 0.9114
No log 7.5814 326 0.8960 0.5488 0.8960 0.9465
No log 7.6279 328 0.8082 0.5823 0.8082 0.8990
No log 7.6744 330 0.7493 0.5121 0.7493 0.8656
No log 7.7209 332 0.7293 0.5391 0.7293 0.8540
No log 7.7674 334 0.7349 0.5391 0.7349 0.8572
No log 7.8140 336 0.7474 0.5152 0.7474 0.8645
No log 7.8605 338 0.7738 0.5905 0.7738 0.8796
No log 7.9070 340 0.7837 0.5650 0.7837 0.8853
No log 7.9535 342 0.8463 0.5553 0.8463 0.9199
No log 8.0 344 0.8743 0.5531 0.8743 0.9351
No log 8.0465 346 0.8657 0.5637 0.8657 0.9304
No log 8.0930 348 0.8990 0.5614 0.8990 0.9482
No log 8.1395 350 0.9514 0.5830 0.9514 0.9754
No log 8.1860 352 0.8763 0.5532 0.8763 0.9361
No log 8.2326 354 0.7440 0.5073 0.7440 0.8626
No log 8.2791 356 0.7029 0.4993 0.7029 0.8384
No log 8.3256 358 0.7120 0.5714 0.7120 0.8438
No log 8.3721 360 0.7006 0.5061 0.7006 0.8370
No log 8.4186 362 0.7262 0.4876 0.7262 0.8522
No log 8.4651 364 0.7606 0.5283 0.7606 0.8721
No log 8.5116 366 0.8493 0.5487 0.8493 0.9216
No log 8.5581 368 0.8713 0.5423 0.8713 0.9335
No log 8.6047 370 0.8266 0.5218 0.8266 0.9092
No log 8.6512 372 0.8264 0.5326 0.8264 0.9091
No log 8.6977 374 0.8965 0.5334 0.8965 0.9469
No log 8.7442 376 0.9062 0.5507 0.9062 0.9519
No log 8.7907 378 0.8973 0.5485 0.8973 0.9472
No log 8.8372 380 0.8339 0.5694 0.8339 0.9132
No log 8.8837 382 0.8252 0.5694 0.8252 0.9084
No log 8.9302 384 0.7663 0.5041 0.7663 0.8754
No log 8.9767 386 0.7295 0.5185 0.7295 0.8541
No log 9.0233 388 0.7958 0.5531 0.7958 0.8921
No log 9.0698 390 0.8979 0.5425 0.8979 0.9476
No log 9.1163 392 0.8961 0.5485 0.8961 0.9466
No log 9.1628 394 0.7635 0.5 0.7635 0.8738
No log 9.2093 396 0.7039 0.5732 0.7039 0.8390
No log 9.2558 398 0.7505 0.5644 0.7505 0.8663
No log 9.3023 400 0.8339 0.5275 0.8339 0.9132
No log 9.3488 402 0.9207 0.4833 0.9207 0.9595
No log 9.3953 404 0.8111 0.4713 0.8111 0.9006
No log 9.4419 406 0.7517 0.6049 0.7517 0.8670
No log 9.4884 408 0.7582 0.5692 0.7582 0.8707
No log 9.5349 410 0.7346 0.5534 0.7346 0.8571
No log 9.5814 412 0.8267 0.5151 0.8267 0.9093
No log 9.6279 414 0.8977 0.4468 0.8977 0.9475
No log 9.6744 416 0.7648 0.5993 0.7648 0.8745
No log 9.7209 418 0.6682 0.5831 0.6682 0.8174
No log 9.7674 420 0.7966 0.5465 0.7966 0.8925
No log 9.8140 422 0.8522 0.5365 0.8522 0.9232
No log 9.8605 424 0.7566 0.5531 0.7566 0.8698
No log 9.9070 426 0.6739 0.5957 0.6739 0.8209
No log 9.9535 428 0.6697 0.6215 0.6697 0.8184
No log 10.0 430 0.6458 0.6843 0.6458 0.8036
No log 10.0465 432 0.6430 0.6902 0.6430 0.8019
No log 10.0930 434 0.6658 0.6369 0.6658 0.8160
No log 10.1395 436 0.7417 0.5513 0.7417 0.8612
No log 10.1860 438 0.7943 0.5412 0.7943 0.8912
No log 10.2326 440 0.8475 0.5389 0.8475 0.9206
No log 10.2791 442 0.8355 0.5365 0.8355 0.9141
No log 10.3256 444 0.7709 0.5712 0.7709 0.8780
No log 10.3721 446 0.7063 0.5244 0.7063 0.8404
No log 10.4186 448 0.6945 0.5690 0.6945 0.8333
No log 10.4651 450 0.7028 0.5376 0.7028 0.8383
No log 10.5116 452 0.6999 0.5732 0.6999 0.8366
No log 10.5581 454 0.7875 0.5934 0.7875 0.8874
No log 10.6047 456 0.8967 0.5389 0.8967 0.9470
No log 10.6512 458 0.9890 0.5448 0.9890 0.9945
No log 10.6977 460 1.0322 0.5809 1.0322 1.0160
No log 10.7442 462 1.0562 0.5809 1.0562 1.0277
No log 10.7907 464 0.9652 0.5584 0.9652 0.9824
No log 10.8372 466 0.8027 0.5151 0.8027 0.8959
No log 10.8837 468 0.7313 0.5635 0.7313 0.8552
No log 10.9302 470 0.7316 0.5706 0.7316 0.8554
No log 10.9767 472 0.7610 0.5614 0.7610 0.8723
No log 11.0233 474 0.7718 0.5614 0.7718 0.8785
No log 11.0698 476 0.7396 0.6214 0.7396 0.8600
No log 11.1163 478 0.6781 0.6005 0.6781 0.8234
No log 11.1628 480 0.6643 0.6371 0.6643 0.8150
No log 11.2093 482 0.6540 0.6041 0.6540 0.8087
No log 11.2558 484 0.6585 0.5898 0.6585 0.8115
No log 11.3023 486 0.7001 0.5674 0.7001 0.8367
No log 11.3488 488 0.7939 0.5696 0.7939 0.8910
No log 11.3953 490 0.8511 0.5591 0.8511 0.9226
No log 11.4419 492 0.9451 0.5262 0.9451 0.9721
No log 11.4884 494 0.8833 0.5384 0.8833 0.9398
No log 11.5349 496 0.7388 0.5650 0.7388 0.8595
No log 11.5814 498 0.6760 0.5381 0.6760 0.8222
0.3593 11.6279 500 0.6820 0.5404 0.6820 0.8259
0.3593 11.6744 502 0.7117 0.4142 0.7117 0.8436
0.3593 11.7209 504 0.7940 0.4858 0.7940 0.8911
0.3593 11.7674 506 0.9192 0.5552 0.9192 0.9587
0.3593 11.8140 508 1.0126 0.5425 1.0126 1.0063
0.3593 11.8605 510 0.9545 0.5262 0.9545 0.9770
0.3593 11.9070 512 0.8323 0.5638 0.8323 0.9123
0.3593 11.9535 514 0.7470 0.5220 0.7470 0.8643
0.3593 12.0 516 0.7382 0.5240 0.7382 0.8592
0.3593 12.0465 518 0.7863 0.5421 0.7863 0.8868
0.3593 12.0930 520 0.8475 0.5183 0.8475 0.9206
0.3593 12.1395 522 0.8754 0.5233 0.8754 0.9356
0.3593 12.1860 524 0.8660 0.5102 0.8660 0.9306
0.3593 12.2326 526 0.8411 0.4754 0.8411 0.9171
0.3593 12.2791 528 0.8333 0.4532 0.8333 0.9128

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k14_task2_organization

Finetuned
(4023)
this model