ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k10_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7626
  • Qwk: 0.4135
  • Mse: 0.7626
  • Rmse: 0.8733

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0588 2 4.0980 0.0086 4.0980 2.0243
No log 0.1176 4 2.2041 -0.0220 2.2041 1.4846
No log 0.1765 6 1.6845 -0.0195 1.6845 1.2979
No log 0.2353 8 1.3876 0.1337 1.3876 1.1780
No log 0.2941 10 1.3847 0.1428 1.3847 1.1767
No log 0.3529 12 1.5011 0.0741 1.5011 1.2252
No log 0.4118 14 1.7690 -0.0678 1.7690 1.3300
No log 0.4706 16 1.8361 -0.0046 1.8361 1.3550
No log 0.5294 18 1.6380 -0.0113 1.6380 1.2798
No log 0.5882 20 1.3948 0.0 1.3948 1.1810
No log 0.6471 22 1.3920 0.0 1.3920 1.1798
No log 0.7059 24 1.2735 -0.0411 1.2735 1.1285
No log 0.7647 26 1.2352 0.1525 1.2352 1.1114
No log 0.8235 28 1.2881 0.2317 1.2881 1.1349
No log 0.8824 30 1.3056 0.2101 1.3056 1.1426
No log 0.9412 32 1.2426 0.2217 1.2426 1.1147
No log 1.0 34 1.2307 0.1768 1.2307 1.1094
No log 1.0588 36 1.1977 0.1983 1.1977 1.0944
No log 1.1176 38 1.1391 0.1442 1.1391 1.0673
No log 1.1765 40 1.1221 0.2385 1.1221 1.0593
No log 1.2353 42 1.0780 0.2456 1.0780 1.0383
No log 1.2941 44 1.0482 0.2214 1.0482 1.0238
No log 1.3529 46 1.0285 0.2057 1.0285 1.0141
No log 1.4118 48 1.0444 0.2097 1.0444 1.0220
No log 1.4706 50 1.0197 0.2211 1.0197 1.0098
No log 1.5294 52 0.9713 0.2551 0.9713 0.9855
No log 1.5882 54 0.9582 0.2625 0.9582 0.9789
No log 1.6471 56 1.0076 0.2669 1.0076 1.0038
No log 1.7059 58 1.4160 0.2364 1.4160 1.1899
No log 1.7647 60 1.1757 0.25 1.1757 1.0843
No log 1.8235 62 0.8714 0.3970 0.8714 0.9335
No log 1.8824 64 0.8764 0.2865 0.8764 0.9362
No log 1.9412 66 1.0354 0.2236 1.0354 1.0176
No log 2.0 68 1.0828 0.2236 1.0828 1.0406
No log 2.0588 70 1.0195 0.2624 1.0195 1.0097
No log 2.1176 72 1.0374 0.2973 1.0374 1.0185
No log 2.1765 74 1.3074 0.2219 1.3074 1.1434
No log 2.2353 76 1.2196 0.3208 1.2196 1.1044
No log 2.2941 78 1.0120 0.3674 1.0120 1.0060
No log 2.3529 80 1.0395 0.2986 1.0395 1.0196
No log 2.4118 82 1.1143 0.1342 1.1143 1.0556
No log 2.4706 84 1.0927 0.1101 1.0927 1.0453
No log 2.5294 86 0.9975 0.1425 0.9975 0.9988
No log 2.5882 88 0.9501 0.3389 0.9501 0.9747
No log 2.6471 90 1.0505 0.2651 1.0505 1.0249
No log 2.7059 92 1.0688 0.2554 1.0688 1.0338
No log 2.7647 94 1.1255 0.2225 1.1255 1.0609
No log 2.8235 96 1.2064 0.2733 1.2064 1.0983
No log 2.8824 98 1.0767 0.3277 1.0767 1.0377
No log 2.9412 100 0.9329 0.3646 0.9329 0.9659
No log 3.0 102 0.8688 0.4398 0.8688 0.9321
No log 3.0588 104 0.9317 0.4401 0.9317 0.9653
No log 3.1176 106 0.9225 0.4219 0.9225 0.9604
No log 3.1765 108 0.8630 0.4830 0.8630 0.9290
No log 3.2353 110 0.8079 0.5102 0.8079 0.8988
No log 3.2941 112 0.8096 0.4565 0.8096 0.8998
No log 3.3529 114 0.8513 0.4843 0.8513 0.9226
No log 3.4118 116 0.9544 0.4982 0.9544 0.9769
No log 3.4706 118 0.9735 0.4784 0.9735 0.9866
No log 3.5294 120 0.8329 0.5025 0.8329 0.9126
No log 3.5882 122 0.8047 0.5025 0.8047 0.8970
No log 3.6471 124 0.9060 0.5567 0.9060 0.9518
No log 3.7059 126 0.8852 0.5294 0.8852 0.9408
No log 3.7647 128 0.7529 0.4869 0.7529 0.8677
No log 3.8235 130 0.7364 0.5523 0.7364 0.8582
No log 3.8824 132 0.7482 0.5184 0.7482 0.8650
No log 3.9412 134 0.7161 0.5439 0.7161 0.8462
No log 4.0 136 0.7548 0.5245 0.7548 0.8688
No log 4.0588 138 0.8096 0.5318 0.8096 0.8998
No log 4.1176 140 0.8001 0.5555 0.8001 0.8945
No log 4.1765 142 0.7855 0.5555 0.7855 0.8863
No log 4.2353 144 0.7300 0.5081 0.7300 0.8544
No log 4.2941 146 0.7307 0.5471 0.7307 0.8548
No log 4.3529 148 0.7599 0.5111 0.7599 0.8717
No log 4.4118 150 0.7826 0.4983 0.7826 0.8846
No log 4.4706 152 0.8549 0.4923 0.8549 0.9246
No log 4.5294 154 1.0837 0.4377 1.0837 1.0410
No log 4.5882 156 1.0763 0.4295 1.0763 1.0374
No log 4.6471 158 0.8319 0.5510 0.8319 0.9121
No log 4.7059 160 0.7640 0.5085 0.7640 0.8741
No log 4.7647 162 0.7740 0.5645 0.7740 0.8798
No log 4.8235 164 0.7874 0.5201 0.7874 0.8874
No log 4.8824 166 0.7722 0.4774 0.7722 0.8787
No log 4.9412 168 0.7444 0.5213 0.7444 0.8628
No log 5.0 170 0.7662 0.5279 0.7662 0.8753
No log 5.0588 172 0.7397 0.4728 0.7397 0.8600
No log 5.1176 174 0.7159 0.4643 0.7159 0.8461
No log 5.1765 176 0.7174 0.5288 0.7174 0.8470
No log 5.2353 178 0.7239 0.5416 0.7239 0.8508
No log 5.2941 180 0.8001 0.4924 0.8001 0.8945
No log 5.3529 182 0.7796 0.5175 0.7796 0.8829
No log 5.4118 184 0.7773 0.5625 0.7773 0.8817
No log 5.4706 186 0.8011 0.4494 0.8011 0.8951
No log 5.5294 188 0.7796 0.5633 0.7796 0.8830
No log 5.5882 190 0.6894 0.5120 0.6894 0.8303
No log 5.6471 192 0.7943 0.5532 0.7943 0.8913
No log 5.7059 194 1.0883 0.4186 1.0883 1.0432
No log 5.7647 196 1.2159 0.4382 1.2159 1.1027
No log 5.8235 198 1.0396 0.4177 1.0396 1.0196
No log 5.8824 200 0.8351 0.4902 0.8351 0.9138
No log 5.9412 202 0.7524 0.5502 0.7524 0.8674
No log 6.0 204 0.7785 0.5064 0.7785 0.8823
No log 6.0588 206 0.8675 0.5110 0.8675 0.9314
No log 6.1176 208 0.9026 0.4181 0.9026 0.9501
No log 6.1765 210 0.8447 0.4807 0.8447 0.9191
No log 6.2353 212 0.7596 0.5407 0.7596 0.8715
No log 6.2941 214 0.7329 0.5287 0.7329 0.8561
No log 6.3529 216 0.7475 0.5451 0.7475 0.8646
No log 6.4118 218 0.7703 0.5828 0.7703 0.8776
No log 6.4706 220 0.8172 0.4870 0.8172 0.9040
No log 6.5294 222 0.7836 0.5898 0.7836 0.8852
No log 6.5882 224 0.7717 0.6196 0.7717 0.8784
No log 6.6471 226 0.7670 0.4677 0.7670 0.8758
No log 6.7059 228 0.8250 0.5210 0.8250 0.9083
No log 6.7647 230 0.7778 0.5342 0.7778 0.8819
No log 6.8235 232 0.7853 0.5131 0.7853 0.8861
No log 6.8824 234 0.8878 0.5188 0.8878 0.9422
No log 6.9412 236 1.0217 0.4576 1.0217 1.0108
No log 7.0 238 0.9190 0.4882 0.9190 0.9586
No log 7.0588 240 0.7608 0.4989 0.7608 0.8722
No log 7.1176 242 0.7936 0.5548 0.7936 0.8908
No log 7.1765 244 0.8289 0.6135 0.8289 0.9104
No log 7.2353 246 0.7921 0.6377 0.7921 0.8900
No log 7.2941 248 0.8147 0.5212 0.8147 0.9026
No log 7.3529 250 0.8937 0.5167 0.8937 0.9454
No log 7.4118 252 0.9237 0.5290 0.9237 0.9611
No log 7.4706 254 1.0486 0.5297 1.0486 1.0240
No log 7.5294 256 1.0631 0.5296 1.0631 1.0310
No log 7.5882 258 1.0204 0.5128 1.0204 1.0101
No log 7.6471 260 0.9362 0.6072 0.9362 0.9676
No log 7.7059 262 0.8627 0.6063 0.8627 0.9288
No log 7.7647 264 0.8092 0.6032 0.8092 0.8995
No log 7.8235 266 0.7890 0.5329 0.7890 0.8882
No log 7.8824 268 0.7952 0.5766 0.7952 0.8917
No log 7.9412 270 0.8305 0.5137 0.8305 0.9113
No log 8.0 272 0.8663 0.5013 0.8663 0.9307
No log 8.0588 274 0.8381 0.4696 0.8381 0.9155
No log 8.1176 276 0.8088 0.4450 0.8088 0.8994
No log 8.1765 278 0.8344 0.3785 0.8344 0.9134
No log 8.2353 280 0.8857 0.4444 0.8857 0.9411
No log 8.2941 282 0.8849 0.4982 0.8849 0.9407
No log 8.3529 284 0.7503 0.5968 0.7503 0.8662
No log 8.4118 286 0.6686 0.6004 0.6686 0.8177
No log 8.4706 288 0.6630 0.6004 0.6630 0.8143
No log 8.5294 290 0.7092 0.5777 0.7092 0.8421
No log 8.5882 292 0.7144 0.5777 0.7144 0.8452
No log 8.6471 294 0.7407 0.5777 0.7407 0.8607
No log 8.7059 296 0.7192 0.5777 0.7192 0.8480
No log 8.7647 298 0.6656 0.5786 0.6656 0.8158
No log 8.8235 300 0.6605 0.6227 0.6605 0.8127
No log 8.8824 302 0.6659 0.6509 0.6659 0.8160
No log 8.9412 304 0.7146 0.5898 0.7146 0.8453
No log 9.0 306 0.9098 0.4965 0.9098 0.9538
No log 9.0588 308 0.9729 0.4669 0.9729 0.9864
No log 9.1176 310 0.8287 0.4810 0.8287 0.9103
No log 9.1765 312 0.6904 0.6167 0.6904 0.8309
No log 9.2353 314 0.7515 0.5337 0.7515 0.8669
No log 9.2941 316 0.7961 0.5348 0.7961 0.8923
No log 9.3529 318 0.7235 0.5219 0.7235 0.8506
No log 9.4118 320 0.6810 0.6206 0.6810 0.8252
No log 9.4706 322 0.7682 0.5244 0.7682 0.8764
No log 9.5294 324 1.0247 0.4669 1.0247 1.0123
No log 9.5882 326 1.0515 0.4856 1.0515 1.0254
No log 9.6471 328 0.8769 0.5389 0.8769 0.9364
No log 9.7059 330 0.7008 0.5728 0.7008 0.8371
No log 9.7647 332 0.6618 0.6003 0.6618 0.8135
No log 9.8235 334 0.6704 0.6003 0.6704 0.8188
No log 9.8824 336 0.7385 0.5867 0.7385 0.8594
No log 9.9412 338 0.7649 0.5331 0.7649 0.8746
No log 10.0 340 0.6942 0.5938 0.6942 0.8332
No log 10.0588 342 0.6507 0.6335 0.6507 0.8067
No log 10.1176 344 0.6549 0.6186 0.6549 0.8092
No log 10.1765 346 0.6741 0.4560 0.6741 0.8210
No log 10.2353 348 0.6935 0.5847 0.6935 0.8328
No log 10.2941 350 0.7378 0.5567 0.7378 0.8589
No log 10.3529 352 0.7298 0.5559 0.7298 0.8543
No log 10.4118 354 0.6976 0.5810 0.6976 0.8352
No log 10.4706 356 0.6874 0.6157 0.6874 0.8291
No log 10.5294 358 0.6669 0.6058 0.6669 0.8167
No log 10.5882 360 0.6584 0.6242 0.6584 0.8114
No log 10.6471 362 0.6354 0.6120 0.6354 0.7971
No log 10.7059 364 0.6284 0.6310 0.6284 0.7927
No log 10.7647 366 0.6405 0.6275 0.6405 0.8003
No log 10.8235 368 0.6377 0.6266 0.6377 0.7986
No log 10.8824 370 0.6118 0.6429 0.6118 0.7822
No log 10.9412 372 0.5808 0.6407 0.5808 0.7621
No log 11.0 374 0.5935 0.6234 0.5935 0.7704
No log 11.0588 376 0.5960 0.6516 0.5960 0.7720
No log 11.1176 378 0.5919 0.6280 0.5919 0.7694
No log 11.1765 380 0.6980 0.6132 0.6980 0.8355
No log 11.2353 382 0.8052 0.5190 0.8052 0.8973
No log 11.2941 384 0.7463 0.5706 0.7463 0.8639
No log 11.3529 386 0.6439 0.6138 0.6439 0.8024
No log 11.4118 388 0.6859 0.5951 0.6859 0.8282
No log 11.4706 390 0.7505 0.5258 0.7505 0.8663
No log 11.5294 392 0.7039 0.5025 0.7039 0.8390
No log 11.5882 394 0.6496 0.5409 0.6496 0.8060
No log 11.6471 396 0.7069 0.5651 0.7069 0.8407
No log 11.7059 398 0.8086 0.5332 0.8086 0.8992
No log 11.7647 400 0.8051 0.5447 0.8051 0.8973
No log 11.8235 402 0.7315 0.5195 0.7315 0.8553
No log 11.8824 404 0.6894 0.5107 0.6894 0.8303
No log 11.9412 406 0.6916 0.5107 0.6916 0.8316
No log 12.0 408 0.7160 0.5459 0.7160 0.8462
No log 12.0588 410 0.7260 0.5446 0.7260 0.8520
No log 12.1176 412 0.7119 0.6032 0.7119 0.8437
No log 12.1765 414 0.7060 0.5785 0.7060 0.8403
No log 12.2353 416 0.6937 0.5388 0.6937 0.8329
No log 12.2941 418 0.7119 0.5337 0.7119 0.8437
No log 12.3529 420 0.6968 0.5430 0.6968 0.8347
No log 12.4118 422 0.6631 0.5913 0.6631 0.8143
No log 12.4706 424 0.6956 0.5718 0.6956 0.8340
No log 12.5294 426 0.7086 0.5707 0.7086 0.8418
No log 12.5882 428 0.6801 0.5707 0.6801 0.8247
No log 12.6471 430 0.6509 0.5688 0.6509 0.8068
No log 12.7059 432 0.6428 0.5875 0.6428 0.8017
No log 12.7647 434 0.6504 0.5650 0.6504 0.8064
No log 12.8235 436 0.6170 0.6100 0.6170 0.7855
No log 12.8824 438 0.6048 0.6564 0.6048 0.7777
No log 12.9412 440 0.6012 0.7143 0.6012 0.7754
No log 13.0 442 0.6126 0.6602 0.6126 0.7827
No log 13.0588 444 0.6415 0.5933 0.6415 0.8009
No log 13.1176 446 0.6633 0.5256 0.6633 0.8144
No log 13.1765 448 0.6904 0.5242 0.6904 0.8309
No log 13.2353 450 0.6522 0.5622 0.6522 0.8076
No log 13.2941 452 0.6275 0.5964 0.6275 0.7922
No log 13.3529 454 0.6387 0.5303 0.6387 0.7992
No log 13.4118 456 0.6399 0.5741 0.6399 0.7999
No log 13.4706 458 0.6409 0.5763 0.6409 0.8006
No log 13.5294 460 0.6520 0.5763 0.6520 0.8074
No log 13.5882 462 0.6776 0.5188 0.6776 0.8232
No log 13.6471 464 0.7182 0.5509 0.7182 0.8474
No log 13.7059 466 0.7237 0.5170 0.7237 0.8507
No log 13.7647 468 0.7290 0.4935 0.7290 0.8538
No log 13.8235 470 0.7941 0.4929 0.7941 0.8911
No log 13.8824 472 0.8634 0.5208 0.8634 0.9292
No log 13.9412 474 0.9137 0.5405 0.9137 0.9559
No log 14.0 476 0.9177 0.4777 0.9177 0.9580
No log 14.0588 478 0.8358 0.5458 0.8358 0.9142
No log 14.1176 480 0.7520 0.5190 0.7520 0.8672
No log 14.1765 482 0.7329 0.5784 0.7329 0.8561
No log 14.2353 484 0.7475 0.6129 0.7475 0.8646
No log 14.2941 486 0.7528 0.6129 0.7528 0.8676
No log 14.3529 488 0.7703 0.5596 0.7703 0.8777
No log 14.4118 490 0.7890 0.5645 0.7890 0.8883
No log 14.4706 492 0.7646 0.5361 0.7646 0.8744
No log 14.5294 494 0.7606 0.5560 0.7606 0.8722
No log 14.5882 496 0.7603 0.5799 0.7603 0.8719
No log 14.6471 498 0.7762 0.5475 0.7762 0.8810
0.2873 14.7059 500 0.7057 0.5413 0.7057 0.8401
0.2873 14.7647 502 0.6776 0.5024 0.6776 0.8232
0.2873 14.8235 504 0.6967 0.5093 0.6967 0.8347
0.2873 14.8824 506 0.7055 0.5219 0.7055 0.8399
0.2873 14.9412 508 0.7024 0.5622 0.7024 0.8381
0.2873 15.0 510 0.7157 0.6176 0.7157 0.8460
0.2873 15.0588 512 0.7378 0.6510 0.7378 0.8590
0.2873 15.1176 514 0.7454 0.6051 0.7454 0.8634
0.2873 15.1765 516 0.7872 0.5816 0.7872 0.8872
0.2873 15.2353 518 0.7626 0.5816 0.7626 0.8733
0.2873 15.2941 520 0.7269 0.5773 0.7269 0.8526
0.2873 15.3529 522 0.7403 0.5766 0.7403 0.8604
0.2873 15.4118 524 0.7806 0.5439 0.7806 0.8835
0.2873 15.4706 526 0.7443 0.5153 0.7443 0.8628
0.2873 15.5294 528 0.6979 0.5262 0.6979 0.8354
0.2873 15.5882 530 0.7291 0.4386 0.7291 0.8539
0.2873 15.6471 532 0.7688 0.4767 0.7688 0.8768
0.2873 15.7059 534 0.7626 0.4135 0.7626 0.8733

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k10_task5_organization

Finetuned
(4019)
this model