ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k14_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8187
  • Qwk: 0.6466
  • Mse: 0.8187
  • Rmse: 0.9048

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0312 2 6.8974 0.0242 6.8974 2.6263
No log 0.0625 4 4.2012 0.0976 4.2012 2.0497
No log 0.0938 6 2.9235 0.0633 2.9235 1.7098
No log 0.125 8 2.8412 0.0140 2.8412 1.6856
No log 0.1562 10 2.2184 0.0000 2.2184 1.4894
No log 0.1875 12 2.1129 0.0672 2.1129 1.4536
No log 0.2188 14 1.8287 0.1132 1.8287 1.3523
No log 0.25 16 1.8113 0.1132 1.8113 1.3458
No log 0.2812 18 2.0753 0.1626 2.0753 1.4406
No log 0.3125 20 1.7971 0.2478 1.7971 1.3406
No log 0.3438 22 1.6986 0.1509 1.6986 1.3033
No log 0.375 24 1.6116 0.0777 1.6116 1.2695
No log 0.4062 26 1.6370 0.0784 1.6370 1.2794
No log 0.4375 28 1.6724 0.1165 1.6724 1.2932
No log 0.4688 30 1.6654 0.1165 1.6654 1.2905
No log 0.5 32 1.5850 0.1165 1.5850 1.2590
No log 0.5312 34 1.5115 0.1165 1.5115 1.2294
No log 0.5625 36 1.4815 0.1165 1.4815 1.2172
No log 0.5938 38 1.4714 0.2243 1.4714 1.2130
No log 0.625 40 1.4695 0.2075 1.4695 1.2122
No log 0.6562 42 1.4875 0.1165 1.4875 1.2196
No log 0.6875 44 1.5555 0.2075 1.5555 1.2472
No log 0.7188 46 1.6029 0.1714 1.6029 1.2661
No log 0.75 48 1.5502 0.1887 1.5502 1.2451
No log 0.7812 50 1.3839 0.2407 1.3839 1.1764
No log 0.8125 52 1.2334 0.3894 1.2334 1.1106
No log 0.8438 54 1.2496 0.4483 1.2496 1.1179
No log 0.875 56 1.3653 0.4483 1.3653 1.1684
No log 0.9062 58 1.3225 0.3894 1.3225 1.1500
No log 0.9375 60 1.2341 0.3860 1.2341 1.1109
No log 0.9688 62 1.2422 0.3540 1.2422 1.1145
No log 1.0 64 1.2748 0.3273 1.2748 1.1291
No log 1.0312 66 1.2782 0.3273 1.2782 1.1306
No log 1.0625 68 1.2303 0.4746 1.2303 1.1092
No log 1.0938 70 1.1844 0.5714 1.1844 1.0883
No log 1.125 72 1.1483 0.5512 1.1483 1.0716
No log 1.1562 74 1.2889 0.4590 1.2889 1.1353
No log 1.1875 76 1.5207 0.3740 1.5207 1.2332
No log 1.2188 78 1.5733 0.3150 1.5733 1.2543
No log 1.25 80 1.5934 0.3088 1.5934 1.2623
No log 1.2812 82 1.4329 0.4296 1.4329 1.1970
No log 1.3125 84 1.4706 0.4412 1.4706 1.2127
No log 1.3438 86 1.5405 0.3768 1.5405 1.2412
No log 1.375 88 1.4200 0.4672 1.4200 1.1916
No log 1.4062 90 1.3071 0.4672 1.3071 1.1433
No log 1.4375 92 1.3580 0.4638 1.3580 1.1654
No log 1.4688 94 1.3153 0.4672 1.3153 1.1469
No log 1.5 96 1.1269 0.5571 1.1269 1.0616
No log 1.5312 98 1.2095 0.5522 1.2095 1.0998
No log 1.5625 100 1.3478 0.5180 1.3478 1.1610
No log 1.5938 102 1.4737 0.4397 1.4737 1.2139
No log 1.625 104 1.4120 0.5180 1.4120 1.1883
No log 1.6562 106 1.1506 0.5909 1.1506 1.0727
No log 1.6875 108 1.0495 0.5522 1.0495 1.0244
No log 1.7188 110 1.1154 0.5909 1.1154 1.0561
No log 1.75 112 1.2211 0.5909 1.2211 1.1050
No log 1.7812 114 1.1774 0.5238 1.1774 1.0851
No log 1.8125 116 1.0257 0.5669 1.0257 1.0128
No log 1.8438 118 0.9755 0.5954 0.9755 0.9877
No log 1.875 120 1.0774 0.6515 1.0774 1.0380
No log 1.9062 122 1.1063 0.6222 1.1063 1.0518
No log 1.9375 124 1.0628 0.6222 1.0628 1.0309
No log 1.9688 126 1.0009 0.6222 1.0009 1.0005
No log 2.0 128 0.9614 0.6377 0.9614 0.9805
No log 2.0312 130 1.0240 0.6277 1.0240 1.0119
No log 2.0625 132 1.0401 0.6107 1.0401 1.0198
No log 2.0938 134 1.0090 0.5625 1.0090 1.0045
No log 2.125 136 0.9468 0.6260 0.9468 0.9730
No log 2.1562 138 0.9574 0.5938 0.9574 0.9785
No log 2.1875 140 0.9266 0.5556 0.9266 0.9626
No log 2.2188 142 1.0669 0.5645 1.0669 1.0329
No log 2.25 144 1.3027 0.4068 1.3027 1.1414
No log 2.2812 146 1.1100 0.528 1.1100 1.0535
No log 2.3125 148 0.9312 0.5938 0.9312 0.9650
No log 2.3438 150 0.9719 0.6165 0.9719 0.9858
No log 2.375 152 1.0303 0.6222 1.0303 1.0150
No log 2.4062 154 1.0177 0.5970 1.0177 1.0088
No log 2.4375 156 1.0557 0.5865 1.0557 1.0275
No log 2.4688 158 1.0354 0.544 1.0354 1.0175
No log 2.5 160 0.9881 0.544 0.9881 0.9941
No log 2.5312 162 0.9717 0.5827 0.9717 0.9858
No log 2.5625 164 0.9907 0.5581 0.9907 0.9953
No log 2.5938 166 1.0914 0.5714 1.0914 1.0447
No log 2.625 168 1.0813 0.5397 1.0813 1.0399
No log 2.6562 170 1.0464 0.5116 1.0464 1.0229
No log 2.6875 172 1.0488 0.5821 1.0488 1.0241
No log 2.7188 174 1.0138 0.5781 1.0138 1.0069
No log 2.75 176 1.0844 0.5588 1.0844 1.0413
No log 2.7812 178 1.3038 0.5324 1.3038 1.1418
No log 2.8125 180 1.3241 0.5634 1.3241 1.1507
No log 2.8438 182 1.1764 0.4806 1.1764 1.0846
No log 2.875 184 1.2012 0.4844 1.2012 1.0960
No log 2.9062 186 1.2314 0.4580 1.2314 1.1097
No log 2.9375 188 1.3896 0.5 1.3896 1.1788
No log 2.9688 190 1.4257 0.4029 1.4257 1.1940
No log 3.0 192 1.2587 0.5231 1.2587 1.1219
No log 3.0312 194 1.1103 0.5238 1.1103 1.0537
No log 3.0625 196 1.1299 0.512 1.1299 1.0630
No log 3.0938 198 1.2190 0.5 1.2190 1.1041
No log 3.125 200 1.2464 0.5481 1.2464 1.1164
No log 3.1562 202 1.3351 0.4748 1.3351 1.1554
No log 3.1875 204 1.2422 0.5333 1.2422 1.1145
No log 3.2188 206 1.0729 0.6176 1.0729 1.0358
No log 3.25 208 1.0487 0.6074 1.0487 1.0241
No log 3.2812 210 1.1516 0.5970 1.1516 1.0731
No log 3.3125 212 1.3575 0.4930 1.3575 1.1651
No log 3.3438 214 1.3155 0.5109 1.3155 1.1469
No log 3.375 216 1.1232 0.6015 1.1232 1.0598
No log 3.4062 218 0.9749 0.6047 0.9749 0.9874
No log 3.4375 220 0.9422 0.5781 0.9422 0.9707
No log 3.4688 222 0.9408 0.6154 0.9408 0.9699
No log 3.5 224 1.0136 0.5846 1.0136 1.0068
No log 3.5312 226 1.2391 0.5926 1.2391 1.1131
No log 3.5625 228 1.4032 0.4143 1.4032 1.1846
No log 3.5938 230 1.3385 0.5147 1.3385 1.1569
No log 3.625 232 1.1460 0.5970 1.1460 1.0705
No log 3.6562 234 1.0078 0.6176 1.0078 1.0039
No log 3.6875 236 0.9888 0.6176 0.9888 0.9944
No log 3.7188 238 1.1148 0.6087 1.1148 1.0559
No log 3.75 240 1.3024 0.5652 1.3024 1.1412
No log 3.7812 242 1.4351 0.4476 1.4351 1.1980
No log 3.8125 244 1.4711 0.4196 1.4711 1.2129
No log 3.8438 246 1.3300 0.4526 1.3300 1.1533
No log 3.875 248 1.1781 0.5606 1.1781 1.0854
No log 3.9062 250 1.0333 0.6165 1.0333 1.0165
No log 3.9375 252 0.9777 0.6471 0.9777 0.9888
No log 3.9688 254 1.0510 0.6331 1.0510 1.0252
No log 4.0 256 1.1474 0.6331 1.1474 1.0712
No log 4.0312 258 1.1971 0.6099 1.1971 1.0941
No log 4.0625 260 1.1577 0.6099 1.1577 1.0760
No log 4.0938 262 1.1403 0.6131 1.1403 1.0678
No log 4.125 264 1.0847 0.6475 1.0847 1.0415
No log 4.1562 266 0.9654 0.7143 0.9654 0.9825
No log 4.1875 268 0.9030 0.6812 0.9030 0.9503
No log 4.2188 270 0.8570 0.6906 0.8570 0.9258
No log 4.25 272 0.8418 0.6950 0.8418 0.9175
No log 4.2812 274 0.8804 0.7234 0.8804 0.9383
No log 4.3125 276 1.0147 0.6197 1.0147 1.0073
No log 4.3438 278 1.1192 0.5789 1.1192 1.0579
No log 4.375 280 1.1127 0.6267 1.1127 1.0548
No log 4.4062 282 1.0595 0.6389 1.0595 1.0293
No log 4.4375 284 1.0177 0.6434 1.0177 1.0088
No log 4.4688 286 0.9833 0.6383 0.9833 0.9916
No log 4.5 288 1.0113 0.6383 1.0113 1.0056
No log 4.5312 290 1.1826 0.6316 1.1826 1.0875
No log 4.5625 292 1.2958 0.6 1.2958 1.1383
No log 4.5938 294 1.1959 0.6497 1.1959 1.0936
No log 4.625 296 1.0863 0.625 1.0863 1.0423
No log 4.6562 298 1.0592 0.6434 1.0592 1.0292
No log 4.6875 300 0.9965 0.6483 0.9965 0.9982
No log 4.7188 302 0.9795 0.6483 0.9795 0.9897
No log 4.75 304 0.8917 0.6713 0.8917 0.9443
No log 4.7812 306 0.8116 0.6897 0.8116 0.9009
No log 4.8125 308 0.8713 0.6525 0.8713 0.9335
No log 4.8438 310 1.0049 0.6301 1.0049 1.0024
No log 4.875 312 1.0923 0.5694 1.0923 1.0451
No log 4.9062 314 1.0022 0.6056 1.0022 1.0011
No log 4.9375 316 0.8826 0.6187 0.8826 0.9394
No log 4.9688 318 0.8453 0.6761 0.8453 0.9194
No log 5.0 320 0.9265 0.6154 0.9265 0.9625
No log 5.0312 322 1.1257 0.5563 1.1257 1.0610
No log 5.0625 324 1.1859 0.5882 1.1859 1.0890
No log 5.0938 326 1.0398 0.5793 1.0398 1.0197
No log 5.125 328 0.9038 0.6475 0.9038 0.9507
No log 5.1562 330 0.8643 0.6522 0.8643 0.9297
No log 5.1875 332 0.8345 0.6269 0.8345 0.9135
No log 5.2188 334 0.8362 0.6423 0.8362 0.9145
No log 5.25 336 1.0274 0.6 1.0274 1.0136
No log 5.2812 338 1.1337 0.6081 1.1337 1.0648
No log 5.3125 340 1.0053 0.6099 1.0053 1.0026
No log 5.3438 342 0.9042 0.6525 0.9042 0.9509
No log 5.375 344 0.9444 0.6525 0.9444 0.9718
No log 5.4062 346 0.9686 0.6301 0.9686 0.9842
No log 5.4375 348 1.0059 0.6752 1.0059 1.0029
No log 5.4688 350 0.9672 0.6452 0.9672 0.9835
No log 5.5 352 0.8986 0.6944 0.8986 0.9480
No log 5.5312 354 0.8289 0.6667 0.8289 0.9104
No log 5.5625 356 0.8178 0.6667 0.8178 0.9043
No log 5.5938 358 0.8626 0.6621 0.8626 0.9288
No log 5.625 360 0.8970 0.6143 0.8970 0.9471
No log 5.6562 362 0.9014 0.6395 0.9014 0.9494
No log 5.6875 364 0.8284 0.6187 0.8284 0.9102
No log 5.7188 366 0.7524 0.6716 0.7524 0.8674
No log 5.75 368 0.7216 0.7 0.7216 0.8494
No log 5.7812 370 0.7400 0.6812 0.7400 0.8602
No log 5.8125 372 0.7549 0.7172 0.7549 0.8689
No log 5.8438 374 0.8400 0.6486 0.8400 0.9165
No log 5.875 376 0.8111 0.6486 0.8111 0.9006
No log 5.9062 378 0.7311 0.7310 0.7311 0.8550
No log 5.9375 380 0.7746 0.6993 0.7746 0.8801
No log 5.9688 382 0.8109 0.6619 0.8109 0.9005
No log 6.0 384 0.8031 0.6757 0.8031 0.8961
No log 6.0312 386 0.7734 0.6939 0.7734 0.8795
No log 6.0625 388 0.7735 0.7075 0.7735 0.8795
No log 6.0938 390 0.7599 0.7092 0.7599 0.8717
No log 6.125 392 0.7548 0.7183 0.7548 0.8688
No log 6.1562 394 0.8042 0.7092 0.8042 0.8968
No log 6.1875 396 0.8644 0.6812 0.8644 0.9297
No log 6.2188 398 1.0085 0.6207 1.0085 1.0042
No log 6.25 400 1.0130 0.6131 1.0130 1.0065
No log 6.2812 402 0.9347 0.6370 0.9347 0.9668
No log 6.3125 404 0.9551 0.6176 0.9551 0.9773
No log 6.3438 406 0.9656 0.6383 0.9656 0.9827
No log 6.375 408 0.9819 0.6383 0.9819 0.9909
No log 6.4062 410 1.0372 0.6286 1.0372 1.0184
No log 6.4375 412 0.9150 0.6383 0.9150 0.9566
No log 6.4688 414 0.8529 0.6569 0.8529 0.9235
No log 6.5 416 0.8094 0.6519 0.8094 0.8996
No log 6.5312 418 0.8486 0.6316 0.8486 0.9212
No log 6.5625 420 0.9258 0.6277 0.9258 0.9622
No log 6.5938 422 1.0154 0.5882 1.0154 1.0077
No log 6.625 424 1.0819 0.5532 1.0819 1.0401
No log 6.6562 426 0.9841 0.6232 0.9841 0.9920
No log 6.6875 428 0.9345 0.6528 0.9345 0.9667
No log 6.7188 430 0.8645 0.6712 0.8645 0.9298
No log 6.75 432 0.9178 0.6667 0.9178 0.9580
No log 6.7812 434 0.9546 0.6528 0.9546 0.9770
No log 6.8125 436 1.0294 0.6275 1.0294 1.0146
No log 6.8438 438 1.0370 0.5816 1.0370 1.0184
No log 6.875 440 0.9262 0.6573 0.9262 0.9624
No log 6.9062 442 0.8585 0.6571 0.8585 0.9265
No log 6.9375 444 0.8063 0.6619 0.8063 0.8979
No log 6.9688 446 0.8145 0.6471 0.8145 0.9025
No log 7.0 448 0.8900 0.6471 0.8900 0.9434
No log 7.0312 450 0.9188 0.6277 0.9188 0.9586
No log 7.0625 452 0.8849 0.6277 0.8849 0.9407
No log 7.0938 454 0.8448 0.6277 0.8448 0.9191
No log 7.125 456 0.7493 0.6815 0.7493 0.8656
No log 7.1562 458 0.7293 0.6815 0.7293 0.8540
No log 7.1875 460 0.7229 0.6912 0.7229 0.8502
No log 7.2188 462 0.7367 0.6716 0.7367 0.8583
No log 7.25 464 0.8595 0.6176 0.8595 0.9271
No log 7.2812 466 0.8864 0.6286 0.8864 0.9415
No log 7.3125 468 0.7654 0.6667 0.7654 0.8749
No log 7.3438 470 0.6982 0.7059 0.6982 0.8356
No log 7.375 472 0.6986 0.6957 0.6986 0.8358
No log 7.4062 474 0.6953 0.7143 0.6953 0.8339
No log 7.4375 476 0.7917 0.7117 0.7917 0.8898
No log 7.4688 478 0.9967 0.6977 0.9967 0.9983
No log 7.5 480 1.0039 0.6932 1.0039 1.0019
No log 7.5312 482 0.9024 0.6977 0.9024 0.9499
No log 7.5625 484 0.8060 0.6623 0.8060 0.8978
No log 7.5938 486 0.7927 0.6429 0.7927 0.8903
No log 7.625 488 0.8484 0.6486 0.8484 0.9211
No log 7.6562 490 0.9504 0.64 0.9504 0.9749
No log 7.6875 492 1.0423 0.6309 1.0423 1.0209
No log 7.7188 494 0.9679 0.6014 0.9679 0.9838
No log 7.75 496 0.8531 0.6471 0.8531 0.9236
No log 7.7812 498 0.7851 0.6667 0.7851 0.8860
0.4393 7.8125 500 0.7817 0.6667 0.7817 0.8841
0.4393 7.8438 502 0.7158 0.7042 0.7158 0.8460
0.4393 7.875 504 0.6154 0.8 0.6154 0.7845
0.4393 7.9062 506 0.5821 0.8079 0.5821 0.7629
0.4393 7.9375 508 0.5587 0.8079 0.5587 0.7475
0.4393 7.9688 510 0.5540 0.8105 0.5540 0.7443
0.4393 8.0 512 0.6222 0.8323 0.6222 0.7888
0.4393 8.0312 514 0.7334 0.7237 0.7334 0.8564
0.4393 8.0625 516 0.8395 0.7067 0.8395 0.9163
0.4393 8.0938 518 0.8206 0.6901 0.8206 0.9059
0.4393 8.125 520 0.8104 0.7042 0.8104 0.9002
0.4393 8.1562 522 0.7842 0.7042 0.7842 0.8855
0.4393 8.1875 524 0.7933 0.7042 0.7933 0.8907
0.4393 8.2188 526 0.8186 0.7042 0.8186 0.9047
0.4393 8.25 528 0.8996 0.6897 0.8996 0.9485
0.4393 8.2812 530 0.9591 0.64 0.9591 0.9794
0.4393 8.3125 532 0.8901 0.6277 0.8901 0.9435
0.4393 8.3438 534 0.7775 0.6714 0.7775 0.8817
0.4393 8.375 536 0.7711 0.6667 0.7711 0.8781
0.4393 8.4062 538 0.8053 0.6522 0.8053 0.8974
0.4393 8.4375 540 0.9070 0.6187 0.9070 0.9523
0.4393 8.4688 542 0.8973 0.6222 0.8973 0.9473
0.4393 8.5 544 0.8187 0.6466 0.8187 0.9048

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k14_task1_organization

Finetuned
(4023)
this model