ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k8_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8015
  • Qwk: 0.6515
  • Mse: 0.8015
  • Rmse: 0.8953

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0526 2 7.2210 0.0179 7.2210 2.6872
No log 0.1053 4 4.9123 0.0866 4.9122 2.2164
No log 0.1579 6 4.1318 -0.0597 4.1318 2.0327
No log 0.2105 8 4.7941 -0.0741 4.7941 2.1895
No log 0.2632 10 4.1707 -0.0796 4.1707 2.0422
No log 0.3158 12 2.7542 0.0526 2.7542 1.6596
No log 0.3684 14 2.1398 0.1618 2.1398 1.4628
No log 0.4211 16 2.2247 0.1014 2.2247 1.4916
No log 0.4737 18 2.2576 0.0735 2.2576 1.5025
No log 0.5263 20 2.0285 0.1000 2.0285 1.4243
No log 0.5789 22 1.7146 0.2243 1.7146 1.3094
No log 0.6316 24 1.5760 0.1905 1.5760 1.2554
No log 0.6842 26 1.6119 0.2222 1.6119 1.2696
No log 0.7368 28 1.7857 0.2881 1.7857 1.3363
No log 0.7895 30 1.7075 0.3419 1.7075 1.3067
No log 0.8421 32 1.5499 0.25 1.5499 1.2449
No log 0.8947 34 1.4502 0.2342 1.4502 1.2042
No log 0.9474 36 1.4030 0.2182 1.4030 1.1845
No log 1.0 38 1.4487 0.3304 1.4487 1.2036
No log 1.0526 40 1.6014 0.3390 1.6014 1.2655
No log 1.1053 42 1.8950 0.2836 1.8950 1.3766
No log 1.1579 44 2.1598 0.2302 2.1598 1.4696
No log 1.2105 46 2.1995 0.2128 2.1995 1.4831
No log 1.2632 48 1.9567 0.2353 1.9567 1.3988
No log 1.3158 50 1.4760 0.4496 1.4760 1.2149
No log 1.3684 52 1.3240 0.4516 1.3240 1.1506
No log 1.4211 54 1.3746 0.4252 1.3746 1.1724
No log 1.4737 56 1.7538 0.3582 1.7538 1.3243
No log 1.5263 58 2.0168 0.2740 2.0168 1.4201
No log 1.5789 60 1.9695 0.2897 1.9695 1.4034
No log 1.6316 62 1.8217 0.3000 1.8217 1.3497
No log 1.6842 64 1.5353 0.4545 1.5353 1.2391
No log 1.7368 66 1.3914 0.5075 1.3914 1.1796
No log 1.7895 68 1.5801 0.4118 1.5801 1.2570
No log 1.8421 70 1.8602 0.3787 1.8602 1.3639
No log 1.8947 72 2.0084 0.3687 2.0084 1.4172
No log 1.9474 74 1.9825 0.3468 1.9825 1.4080
No log 2.0 76 1.8568 0.4255 1.8568 1.3626
No log 2.0526 78 1.9780 0.2190 1.9780 1.4064
No log 2.1053 80 1.5901 0.4252 1.5901 1.2610
No log 2.1579 82 1.3697 0.4651 1.3697 1.1703
No log 2.2105 84 1.3231 0.4769 1.3231 1.1503
No log 2.2632 86 1.5770 0.3876 1.5770 1.2558
No log 2.3158 88 1.8945 0.2336 1.8945 1.3764
No log 2.3684 90 1.8998 0.25 1.8998 1.3783
No log 2.4211 92 1.4037 0.4733 1.4037 1.1848
No log 2.4737 94 0.9229 0.5954 0.9229 0.9607
No log 2.5263 96 0.8342 0.7059 0.8342 0.9134
No log 2.5789 98 0.8452 0.6812 0.8452 0.9194
No log 2.6316 100 0.9366 0.7034 0.9366 0.9678
No log 2.6842 102 1.1350 0.7020 1.1350 1.0654
No log 2.7368 104 1.1262 0.7020 1.1262 1.0612
No log 2.7895 106 0.9862 0.6892 0.9862 0.9931
No log 2.8421 108 0.8593 0.6667 0.8593 0.9270
No log 2.8947 110 0.9241 0.6377 0.9241 0.9613
No log 2.9474 112 0.8866 0.6423 0.8866 0.9416
No log 3.0 114 0.8665 0.6619 0.8665 0.9308
No log 3.0526 116 1.0394 0.6569 1.0394 1.0195
No log 3.1053 118 1.2926 0.4928 1.2926 1.1369
No log 3.1579 120 1.3159 0.4412 1.3159 1.1471
No log 3.2105 122 1.2690 0.4925 1.2690 1.1265
No log 3.2632 124 1.1716 0.5735 1.1716 1.0824
No log 3.3158 126 1.0724 0.6294 1.0724 1.0356
No log 3.3684 128 1.0512 0.6232 1.0512 1.0253
No log 3.4211 130 0.9926 0.5926 0.9926 0.9963
No log 3.4737 132 0.9370 0.6119 0.9370 0.9680
No log 3.5263 134 0.9012 0.6471 0.9012 0.9493
No log 3.5789 136 0.9628 0.6763 0.9628 0.9812
No log 3.6316 138 1.0033 0.6897 1.0033 1.0016
No log 3.6842 140 0.9803 0.6803 0.9803 0.9901
No log 3.7368 142 0.9940 0.6800 0.9940 0.9970
No log 3.7895 144 1.0122 0.6395 1.0122 1.0061
No log 3.8421 146 1.1402 0.5921 1.1402 1.0678
No log 3.8947 148 1.1472 0.5882 1.1472 1.0711
No log 3.9474 150 0.9689 0.6286 0.9689 0.9843
No log 4.0 152 0.8778 0.6763 0.8778 0.9369
No log 4.0526 154 0.9266 0.6370 0.9266 0.9626
No log 4.1053 156 0.8918 0.6715 0.8918 0.9443
No log 4.1579 158 0.8897 0.6370 0.8897 0.9432
No log 4.2105 160 1.0216 0.5634 1.0216 1.0108
No log 4.2632 162 1.1611 0.5946 1.1611 1.0775
No log 4.3158 164 1.0859 0.5793 1.0859 1.0420
No log 4.3684 166 0.9404 0.5581 0.9404 0.9697
No log 4.4211 168 0.9100 0.6466 0.9100 0.9539
No log 4.4737 170 0.9173 0.6212 0.9173 0.9577
No log 4.5263 172 0.9119 0.5802 0.9119 0.9549
No log 4.5789 174 0.9586 0.5970 0.9586 0.9791
No log 4.6316 176 0.9498 0.6286 0.9498 0.9746
No log 4.6842 178 0.8721 0.6377 0.8721 0.9338
No log 4.7368 180 0.8534 0.6667 0.8534 0.9238
No log 4.7895 182 0.9601 0.5736 0.9601 0.9798
No log 4.8421 184 1.1018 0.4839 1.1018 1.0497
No log 4.8947 186 1.2485 0.3802 1.2485 1.1174
No log 4.9474 188 1.0715 0.4098 1.0715 1.0351
No log 5.0 190 1.0145 0.5344 1.0145 1.0072
No log 5.0526 192 1.1505 0.5294 1.1505 1.0726
No log 5.1053 194 1.1523 0.5231 1.1523 1.0734
No log 5.1579 196 1.1283 0.5 1.1283 1.0622
No log 5.2105 198 1.0709 0.496 1.0709 1.0348
No log 5.2632 200 1.0177 0.5846 1.0177 1.0088
No log 5.3158 202 1.0210 0.5970 1.0210 1.0104
No log 5.3684 204 0.9519 0.6074 0.9519 0.9757
No log 5.4211 206 0.9432 0.6165 0.9432 0.9712
No log 5.4737 208 1.0536 0.6711 1.0536 1.0265
No log 5.5263 210 1.1748 0.6242 1.1748 1.0839
No log 5.5789 212 1.0523 0.6536 1.0523 1.0258
No log 5.6316 214 0.8616 0.7123 0.8616 0.9282
No log 5.6842 216 0.8244 0.6809 0.8244 0.9079
No log 5.7368 218 0.8561 0.7484 0.8561 0.9252
No log 5.7895 220 0.9692 0.6316 0.9692 0.9845
No log 5.8421 222 1.1304 0.6087 1.1304 1.0632
No log 5.8947 224 1.0497 0.6282 1.0497 1.0245
No log 5.9474 226 0.9370 0.6301 0.9370 0.9680
No log 6.0 228 0.8869 0.6475 0.8869 0.9417
No log 6.0526 230 0.9077 0.6324 0.9077 0.9527
No log 6.1053 232 0.8914 0.6475 0.8914 0.9441
No log 6.1579 234 0.9279 0.6301 0.9279 0.9633
No log 6.2105 236 1.1064 0.6115 1.1064 1.0518
No log 6.2632 238 1.0869 0.625 1.0869 1.0425
No log 6.3158 240 0.9186 0.6622 0.9186 0.9584
No log 6.3684 242 0.8687 0.6525 0.8687 0.9320
No log 6.4211 244 0.8577 0.6277 0.8577 0.9261
No log 6.4737 246 0.8769 0.6618 0.8769 0.9364
No log 6.5263 248 0.9165 0.6803 0.9165 0.9573
No log 6.5789 250 0.9221 0.6803 0.9221 0.9602
No log 6.6316 252 0.9270 0.6143 0.9270 0.9628
No log 6.6842 254 0.9864 0.5797 0.9864 0.9932
No log 6.7368 256 1.0942 0.5612 1.0942 1.0460
No log 6.7895 258 1.1866 0.5306 1.1866 1.0893
No log 6.8421 260 1.2524 0.5442 1.2524 1.1191
No log 6.8947 262 1.1639 0.5833 1.1639 1.0788
No log 6.9474 264 1.0523 0.5714 1.0523 1.0258
No log 7.0 266 0.9647 0.5827 0.9647 0.9822
No log 7.0526 268 0.9260 0.6364 0.9260 0.9623
No log 7.1053 270 0.9387 0.5954 0.9387 0.9689
No log 7.1579 272 0.8922 0.6269 0.8922 0.9445
No log 7.2105 274 0.8855 0.6857 0.8855 0.9410
No log 7.2632 276 0.9438 0.6486 0.9438 0.9715
No log 7.3158 278 1.1080 0.5867 1.1080 1.0526
No log 7.3684 280 1.0833 0.6053 1.0833 1.0408
No log 7.4211 282 0.9216 0.6447 0.9216 0.9600
No log 7.4737 284 0.8422 0.6809 0.8422 0.9177
No log 7.5263 286 0.8836 0.6471 0.8836 0.9400
No log 7.5789 288 1.0390 0.5714 1.0390 1.0193
No log 7.6316 290 1.1135 0.5 1.1135 1.0552
No log 7.6842 292 1.0800 0.4677 1.0800 1.0392
No log 7.7368 294 1.0248 0.5082 1.0248 1.0123
No log 7.7895 296 1.0273 0.4754 1.0273 1.0136
No log 7.8421 298 1.0229 0.4754 1.0229 1.0114
No log 7.8947 300 1.0172 0.5156 1.0172 1.0086
No log 7.9474 302 1.0258 0.5564 1.0258 1.0128
No log 8.0 304 0.9476 0.5821 0.9476 0.9734
No log 8.0526 306 0.8814 0.6519 0.8814 0.9389
No log 8.1053 308 0.8876 0.6620 0.8876 0.9421
No log 8.1579 310 1.0219 0.6093 1.0219 1.0109
No log 8.2105 312 1.2332 0.6049 1.2332 1.1105
No log 8.2632 314 1.4156 0.5952 1.4156 1.1898
No log 8.3158 316 1.3596 0.6038 1.3596 1.1660
No log 8.3684 318 1.1287 0.5612 1.1287 1.0624
No log 8.4211 320 1.0214 0.5481 1.0214 1.0107
No log 8.4737 322 0.9676 0.5630 0.9676 0.9837
No log 8.5263 324 1.0000 0.5735 1.0000 1.0000
No log 8.5789 326 1.0360 0.5857 1.0360 1.0179
No log 8.6316 328 1.0952 0.5931 1.0952 1.0465
No log 8.6842 330 1.1370 0.6369 1.1370 1.0663
No log 8.7368 332 1.0993 0.6369 1.0993 1.0485
No log 8.7895 334 1.1274 0.6093 1.1274 1.0618
No log 8.8421 336 1.1647 0.5517 1.1647 1.0792
No log 8.8947 338 1.0622 0.5630 1.0622 1.0306
No log 8.9474 340 1.0665 0.5821 1.0665 1.0327
No log 9.0 342 1.1668 0.5373 1.1668 1.0802
No log 9.0526 344 1.1780 0.5373 1.1780 1.0853
No log 9.1053 346 1.0198 0.5778 1.0198 1.0099
No log 9.1579 348 0.8675 0.6471 0.8675 0.9314
No log 9.2105 350 0.8350 0.6857 0.8350 0.9138
No log 9.2632 352 0.8521 0.6471 0.8521 0.9231
No log 9.3158 354 0.9664 0.6338 0.9664 0.9831
No log 9.3684 356 1.2302 0.5752 1.2302 1.1092
No log 9.4211 358 1.3033 0.5695 1.3033 1.1416
No log 9.4737 360 1.1797 0.5255 1.1797 1.0861
No log 9.5263 362 0.9707 0.5649 0.9707 0.9852
No log 9.5789 364 0.8812 0.625 0.8812 0.9387
No log 9.6316 366 0.8993 0.6299 0.8993 0.9483
No log 9.6842 368 0.9299 0.6142 0.9299 0.9643
No log 9.7368 370 0.9666 0.5354 0.9666 0.9831
No log 9.7895 372 1.0198 0.6027 1.0198 1.0099
No log 9.8421 374 1.0326 0.6 1.0326 1.0162
No log 9.8947 376 1.1784 0.6329 1.1784 1.0856
No log 9.9474 378 1.2066 0.6329 1.2066 1.0985
No log 10.0 380 1.0757 0.6194 1.0757 1.0372
No log 10.0526 382 0.9200 0.6447 0.9200 0.9592
No log 10.1053 384 0.8879 0.6490 0.8879 0.9423
No log 10.1579 386 0.9263 0.6536 0.9263 0.9624
No log 10.2105 388 0.9057 0.6536 0.9057 0.9517
No log 10.2632 390 0.8368 0.6892 0.8368 0.9148
No log 10.3158 392 0.7967 0.6716 0.7967 0.8926
No log 10.3684 394 0.8143 0.6617 0.8143 0.9024
No log 10.4211 396 0.8694 0.5909 0.8694 0.9324
No log 10.4737 398 0.8396 0.6212 0.8396 0.9163
No log 10.5263 400 0.8135 0.6519 0.8135 0.9020
No log 10.5789 402 0.8234 0.6176 0.8234 0.9074
No log 10.6316 404 0.8441 0.6483 0.8441 0.9187
No log 10.6842 406 0.8745 0.6755 0.8745 0.9352
No log 10.7368 408 0.8046 0.6939 0.8046 0.8970
No log 10.7895 410 0.7265 0.7133 0.7265 0.8524
No log 10.8421 412 0.7314 0.7153 0.7314 0.8552
No log 10.8947 414 0.7410 0.7206 0.7410 0.8608
No log 10.9474 416 0.7405 0.7050 0.7405 0.8605
No log 11.0 418 0.7943 0.6849 0.7943 0.8912
No log 11.0526 420 0.9473 0.6835 0.9473 0.9733
No log 11.1053 422 1.1210 0.6456 1.1210 1.0588
No log 11.1579 424 1.1774 0.6456 1.1774 1.0851
No log 11.2105 426 1.0312 0.6364 1.0312 1.0155
No log 11.2632 428 0.8691 0.6241 0.8691 0.9323
No log 11.3158 430 0.8232 0.6716 0.8232 0.9073
No log 11.3684 432 0.8255 0.6466 0.8255 0.9086
No log 11.4211 434 0.9159 0.5714 0.9159 0.9570
No log 11.4737 436 1.0707 0.5775 1.0707 1.0347
No log 11.5263 438 1.2484 0.5946 1.2484 1.1173
No log 11.5789 440 1.1323 0.6174 1.1323 1.0641
No log 11.6316 442 1.0662 0.6207 1.0662 1.0326
No log 11.6842 444 0.9167 0.6131 0.9167 0.9574
No log 11.7368 446 0.8429 0.5846 0.8429 0.9181
No log 11.7895 448 0.8640 0.6 0.8640 0.9295
No log 11.8421 450 0.9838 0.5522 0.9838 0.9919
No log 11.8947 452 1.2610 0.5789 1.2610 1.1230
No log 11.9474 454 1.4184 0.5897 1.4184 1.1909
No log 12.0 456 1.3222 0.6026 1.3222 1.1499
No log 12.0526 458 1.0098 0.6234 1.0098 1.0049
No log 12.1053 460 0.7950 0.6364 0.7950 0.8916
No log 12.1579 462 0.7720 0.7015 0.7720 0.8786
No log 12.2105 464 0.7794 0.6818 0.7794 0.8829
No log 12.2632 466 0.7556 0.7015 0.7556 0.8692
No log 12.3158 468 0.7615 0.7153 0.7615 0.8726
No log 12.3684 470 0.7854 0.6316 0.7854 0.8862
No log 12.4211 472 0.7451 0.7153 0.7451 0.8632
No log 12.4737 474 0.7352 0.7338 0.7352 0.8574
No log 12.5263 476 0.7629 0.7338 0.7629 0.8735
No log 12.5789 478 0.8001 0.7259 0.8001 0.8945
No log 12.6316 480 0.8188 0.6565 0.8188 0.9049
No log 12.6842 482 0.8382 0.6277 0.8382 0.9155
No log 12.7368 484 0.8490 0.6389 0.8490 0.9214
No log 12.7895 486 0.8069 0.6575 0.8069 0.8983
No log 12.8421 488 0.7584 0.6667 0.7584 0.8709
No log 12.8947 490 0.7126 0.7297 0.7126 0.8442
No log 12.9474 492 0.7190 0.6950 0.7190 0.8479
No log 13.0 494 0.7274 0.6763 0.7274 0.8529
No log 13.0526 496 0.7475 0.6761 0.7475 0.8646
No log 13.1053 498 0.6983 0.7310 0.6983 0.8357
0.4278 13.1579 500 0.6989 0.7383 0.6989 0.8360
0.4278 13.2105 502 0.7633 0.6533 0.7633 0.8737
0.4278 13.2632 504 0.7965 0.6710 0.7965 0.8925
0.4278 13.3158 506 0.7613 0.6533 0.7613 0.8725
0.4278 13.3684 508 0.7330 0.7211 0.7330 0.8562
0.4278 13.4211 510 0.7276 0.7534 0.7276 0.8530
0.4278 13.4737 512 0.7443 0.7465 0.7443 0.8628
0.4278 13.5263 514 0.7660 0.7338 0.7660 0.8752
0.4278 13.5789 516 0.7669 0.7338 0.7669 0.8758
0.4278 13.6316 518 0.7680 0.7338 0.7680 0.8764
0.4278 13.6842 520 0.7858 0.6912 0.7858 0.8864
0.4278 13.7368 522 0.8015 0.6515 0.8015 0.8953

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k8_task1_organization

Finetuned
(4019)
this model