ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k11_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5370
  • Qwk: 0.6118
  • Mse: 0.5370
  • Rmse: 0.7328

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0526 2 2.4188 -0.0262 2.4188 1.5553
No log 0.1053 4 1.2433 0.0736 1.2433 1.1150
No log 0.1579 6 0.9166 -0.0288 0.9166 0.9574
No log 0.2105 8 1.1279 -0.1914 1.1279 1.0620
No log 0.2632 10 1.4794 -0.3188 1.4794 1.2163
No log 0.3158 12 1.3061 -0.2562 1.3061 1.1429
No log 0.3684 14 1.0753 0.0952 1.0753 1.0370
No log 0.4211 16 0.8543 0.1648 0.8543 0.9243
No log 0.4737 18 0.7758 0.2751 0.7758 0.8808
No log 0.5263 20 0.7412 0.0428 0.7412 0.8609
No log 0.5789 22 0.7567 0.0 0.7567 0.8699
No log 0.6316 24 0.7719 0.0 0.7719 0.8786
No log 0.6842 26 0.7968 0.0 0.7968 0.8926
No log 0.7368 28 0.7350 0.0428 0.7350 0.8573
No log 0.7895 30 0.7190 0.1863 0.7190 0.8480
No log 0.8421 32 0.7293 0.1863 0.7293 0.8540
No log 0.8947 34 0.7361 0.0428 0.7361 0.8580
No log 0.9474 36 0.8620 0.0053 0.8620 0.9284
No log 1.0 38 0.9257 0.0509 0.9257 0.9621
No log 1.0526 40 1.0237 0.0203 1.0237 1.0118
No log 1.1053 42 0.9898 0.1739 0.9898 0.9949
No log 1.1579 44 0.8916 0.2109 0.8916 0.9443
No log 1.2105 46 0.7927 0.1365 0.7927 0.8903
No log 1.2632 48 0.7367 0.0327 0.7367 0.8583
No log 1.3158 50 0.7276 0.2537 0.7276 0.8530
No log 1.3684 52 0.7451 0.3243 0.7451 0.8632
No log 1.4211 54 0.7927 0.3092 0.7927 0.8903
No log 1.4737 56 0.7660 0.2537 0.7660 0.8752
No log 1.5263 58 0.7987 0.0679 0.7987 0.8937
No log 1.5789 60 0.8626 -0.0517 0.8626 0.9288
No log 1.6316 62 0.9719 0.0139 0.9719 0.9859
No log 1.6842 64 1.0397 -0.0173 1.0397 1.0196
No log 1.7368 66 0.9769 -0.0389 0.9769 0.9884
No log 1.7895 68 0.8425 0.0 0.8425 0.9179
No log 1.8421 70 0.7528 0.0 0.7528 0.8677
No log 1.8947 72 0.7182 0.0 0.7182 0.8475
No log 1.9474 74 0.7584 0.0 0.7584 0.8709
No log 2.0 76 0.9802 0.2183 0.9802 0.9900
No log 2.0526 78 1.0953 0.1535 1.0953 1.0466
No log 2.1053 80 1.1364 0.1755 1.1364 1.0660
No log 2.1579 82 1.0979 0.1819 1.0979 1.0478
No log 2.2105 84 0.9573 0.2875 0.9573 0.9784
No log 2.2632 86 0.8381 0.3492 0.8381 0.9155
No log 2.3158 88 0.8300 0.3090 0.8300 0.9111
No log 2.3684 90 0.8007 0.3090 0.8007 0.8948
No log 2.4211 92 0.8091 0.3398 0.8091 0.8995
No log 2.4737 94 0.7295 0.1358 0.7295 0.8541
No log 2.5263 96 0.7079 0.0444 0.7079 0.8414
No log 2.5789 98 0.7017 -0.0027 0.7017 0.8377
No log 2.6316 100 0.7301 0.0053 0.7301 0.8544
No log 2.6842 102 0.6968 0.0846 0.6968 0.8347
No log 2.7368 104 0.7447 0.1352 0.7447 0.8629
No log 2.7895 106 0.8718 0.2642 0.8718 0.9337
No log 2.8421 108 0.8623 0.2559 0.8623 0.9286
No log 2.8947 110 0.6937 0.2872 0.6937 0.8329
No log 2.9474 112 0.6675 0.3755 0.6675 0.8170
No log 3.0 114 0.8971 0.2308 0.8971 0.9471
No log 3.0526 116 0.8897 0.1504 0.8897 0.9433
No log 3.1053 118 0.8349 -0.0026 0.8349 0.9137
No log 3.1579 120 0.9784 0.1685 0.9784 0.9891
No log 3.2105 122 1.0145 0.1618 1.0145 1.0072
No log 3.2632 124 1.1045 0.0980 1.1045 1.0509
No log 3.3158 126 0.9946 0.1599 0.9946 0.9973
No log 3.3684 128 0.7739 0.1352 0.7739 0.8797
No log 3.4211 130 0.7101 0.2085 0.7101 0.8427
No log 3.4737 132 0.7805 0.3637 0.7805 0.8835
No log 3.5263 134 0.7434 0.3444 0.7434 0.8622
No log 3.5789 136 0.6953 0.2413 0.6953 0.8339
No log 3.6316 138 0.7040 0.3170 0.7040 0.8390
No log 3.6842 140 0.6684 0.3995 0.6684 0.8176
No log 3.7368 142 0.7049 0.3950 0.7049 0.8396
No log 3.7895 144 0.7407 0.3302 0.7407 0.8606
No log 3.8421 146 0.6939 0.3594 0.6939 0.8330
No log 3.8947 148 0.6145 0.4276 0.6145 0.7839
No log 3.9474 150 0.7237 0.4066 0.7237 0.8507
No log 4.0 152 0.8352 0.3551 0.8352 0.9139
No log 4.0526 154 0.8765 0.3761 0.8765 0.9362
No log 4.1053 156 0.7751 0.3872 0.7751 0.8804
No log 4.1579 158 0.6336 0.4635 0.6336 0.7960
No log 4.2105 160 0.5841 0.4052 0.5841 0.7643
No log 4.2632 162 0.6131 0.4306 0.6131 0.7830
No log 4.3158 164 0.6192 0.4035 0.6192 0.7869
No log 4.3684 166 0.6553 0.3789 0.6553 0.8095
No log 4.4211 168 0.6587 0.3950 0.6587 0.8116
No log 4.4737 170 0.6846 0.4430 0.6846 0.8274
No log 4.5263 172 0.6429 0.3622 0.6429 0.8018
No log 4.5789 174 0.6470 0.2336 0.6470 0.8044
No log 4.6316 176 0.6701 0.2405 0.6701 0.8186
No log 4.6842 178 0.6438 0.2405 0.6438 0.8024
No log 4.7368 180 0.6182 0.2336 0.6182 0.7863
No log 4.7895 182 0.6118 0.2641 0.6118 0.7822
No log 4.8421 184 0.6122 0.3659 0.6122 0.7824
No log 4.8947 186 0.6488 0.4298 0.6488 0.8055
No log 4.9474 188 0.7409 0.3707 0.7409 0.8607
No log 5.0 190 0.7190 0.3948 0.7190 0.8479
No log 5.0526 192 0.6369 0.4991 0.6369 0.7981
No log 5.1053 194 0.6694 0.5239 0.6694 0.8182
No log 5.1579 196 0.7010 0.5471 0.7010 0.8372
No log 5.2105 198 0.6277 0.5625 0.6277 0.7923
No log 5.2632 200 0.5910 0.5057 0.5910 0.7688
No log 5.3158 202 0.6067 0.5493 0.6067 0.7789
No log 5.3684 204 0.7122 0.3872 0.7122 0.8439
No log 5.4211 206 0.7615 0.3559 0.7615 0.8726
No log 5.4737 208 0.7182 0.4482 0.7182 0.8474
No log 5.5263 210 0.5950 0.4849 0.5950 0.7714
No log 5.5789 212 0.5588 0.4526 0.5588 0.7476
No log 5.6316 214 0.5636 0.4878 0.5636 0.7507
No log 5.6842 216 0.6178 0.4869 0.6178 0.7860
No log 5.7368 218 0.7503 0.3471 0.7503 0.8662
No log 5.7895 220 0.7560 0.3455 0.7560 0.8695
No log 5.8421 222 0.6523 0.4888 0.6523 0.8077
No log 5.8947 224 0.5655 0.4878 0.5655 0.7520
No log 5.9474 226 0.5965 0.5016 0.5965 0.7723
No log 6.0 228 0.5936 0.5081 0.5936 0.7704
No log 6.0526 230 0.5566 0.5404 0.5566 0.7461
No log 6.1053 232 0.5568 0.5143 0.5568 0.7462
No log 6.1579 234 0.6382 0.4914 0.6382 0.7989
No log 6.2105 236 0.6315 0.4290 0.6315 0.7947
No log 6.2632 238 0.5636 0.4789 0.5636 0.7507
No log 6.3158 240 0.6381 0.4606 0.6381 0.7988
No log 6.3684 242 0.8905 0.3608 0.8905 0.9437
No log 6.4211 244 0.9371 0.2872 0.9371 0.9681
No log 6.4737 246 0.7778 0.4102 0.7778 0.8819
No log 6.5263 248 0.6462 0.4035 0.6462 0.8038
No log 6.5789 250 0.6288 0.3717 0.6288 0.7930
No log 6.6316 252 0.6721 0.3953 0.6721 0.8198
No log 6.6842 254 0.7368 0.3675 0.7368 0.8584
No log 6.7368 256 0.7577 0.3593 0.7577 0.8705
No log 6.7895 258 0.7033 0.4064 0.7033 0.8386
No log 6.8421 260 0.6837 0.4205 0.6837 0.8269
No log 6.8947 262 0.6913 0.4425 0.6913 0.8314
No log 6.9474 264 0.6713 0.4391 0.6713 0.8193
No log 7.0 266 0.6009 0.3603 0.6009 0.7752
No log 7.0526 268 0.5780 0.4253 0.5780 0.7603
No log 7.1053 270 0.5800 0.3454 0.5800 0.7616
No log 7.1579 272 0.5887 0.3883 0.5887 0.7673
No log 7.2105 274 0.6276 0.4087 0.6276 0.7922
No log 7.2632 276 0.5891 0.4044 0.5891 0.7675
No log 7.3158 278 0.5597 0.4029 0.5597 0.7481
No log 7.3684 280 0.5597 0.4190 0.5597 0.7481
No log 7.4211 282 0.5627 0.6065 0.5627 0.7501
No log 7.4737 284 0.5947 0.3726 0.5947 0.7711
No log 7.5263 286 0.7172 0.4527 0.7172 0.8469
No log 7.5789 288 0.8820 0.4033 0.8820 0.9391
No log 7.6316 290 0.8188 0.3976 0.8188 0.9049
No log 7.6842 292 0.6332 0.4817 0.6332 0.7957
No log 7.7368 294 0.5270 0.6171 0.5270 0.7259
No log 7.7895 296 0.5300 0.6147 0.5300 0.7280
No log 7.8421 298 0.5179 0.6334 0.5179 0.7196
No log 7.8947 300 0.4903 0.6818 0.4903 0.7002
No log 7.9474 302 0.5064 0.6310 0.5064 0.7116
No log 8.0 304 0.5358 0.5674 0.5358 0.7320
No log 8.0526 306 0.5254 0.6100 0.5254 0.7249
No log 8.1053 308 0.5102 0.6158 0.5102 0.7143
No log 8.1579 310 0.4975 0.6479 0.4975 0.7053
No log 8.2105 312 0.5062 0.6426 0.5062 0.7115
No log 8.2632 314 0.5119 0.6517 0.5119 0.7155
No log 8.3158 316 0.5269 0.5009 0.5269 0.7259
No log 8.3684 318 0.5444 0.4635 0.5444 0.7379
No log 8.4211 320 0.5486 0.4813 0.5486 0.7407
No log 8.4737 322 0.5262 0.5939 0.5262 0.7254
No log 8.5263 324 0.5240 0.6405 0.5240 0.7239
No log 8.5789 326 0.5309 0.6336 0.5309 0.7286
No log 8.6316 328 0.5301 0.6923 0.5301 0.7281
No log 8.6842 330 0.5328 0.6765 0.5328 0.7299
No log 8.7368 332 0.5207 0.6835 0.5207 0.7216
No log 8.7895 334 0.5102 0.6541 0.5102 0.7143
No log 8.8421 336 0.5170 0.6541 0.5170 0.7191
No log 8.8947 338 0.5296 0.6001 0.5296 0.7277
No log 8.9474 340 0.5434 0.5867 0.5434 0.7371
No log 9.0 342 0.5337 0.6377 0.5337 0.7306
No log 9.0526 344 0.5510 0.6382 0.5510 0.7423
No log 9.1053 346 0.5588 0.6409 0.5588 0.7475
No log 9.1579 348 0.6242 0.4434 0.6242 0.7900
No log 9.2105 350 0.6885 0.4243 0.6885 0.8297
No log 9.2632 352 0.6701 0.4243 0.6701 0.8186
No log 9.3158 354 0.5596 0.5533 0.5596 0.7481
No log 9.3684 356 0.5318 0.6416 0.5318 0.7292
No log 9.4211 358 0.5289 0.6416 0.5289 0.7272
No log 9.4737 360 0.5537 0.4689 0.5537 0.7441
No log 9.5263 362 0.5118 0.6158 0.5118 0.7154
No log 9.5789 364 0.4995 0.6541 0.4995 0.7068
No log 9.6316 366 0.5041 0.6431 0.5041 0.7100
No log 9.6842 368 0.4997 0.6650 0.4997 0.7069
No log 9.7368 370 0.5046 0.6730 0.5046 0.7104
No log 9.7895 372 0.5210 0.6118 0.5210 0.7218
No log 9.8421 374 0.5197 0.6481 0.5197 0.7209
No log 9.8947 376 0.5303 0.6040 0.5303 0.7282
No log 9.9474 378 0.5309 0.6245 0.5309 0.7287
No log 10.0 380 0.5572 0.5344 0.5572 0.7464
No log 10.0526 382 0.5483 0.5589 0.5483 0.7405
No log 10.1053 384 0.5445 0.5589 0.5445 0.7379
No log 10.1579 386 0.5510 0.5589 0.5510 0.7423
No log 10.2105 388 0.5407 0.5610 0.5407 0.7353
No log 10.2632 390 0.5543 0.5434 0.5543 0.7445
No log 10.3158 392 0.5367 0.5741 0.5367 0.7326
No log 10.3684 394 0.5312 0.5363 0.5312 0.7289
No log 10.4211 396 0.5200 0.6645 0.5200 0.7211
No log 10.4737 398 0.5269 0.6561 0.5269 0.7259
No log 10.5263 400 0.5366 0.6918 0.5366 0.7325
No log 10.5789 402 0.5466 0.6416 0.5466 0.7393
No log 10.6316 404 0.5773 0.5185 0.5773 0.7598
No log 10.6842 406 0.5943 0.4835 0.5943 0.7709
No log 10.7368 408 0.5589 0.5884 0.5589 0.7476
No log 10.7895 410 0.5476 0.6593 0.5476 0.7400
No log 10.8421 412 0.5348 0.7004 0.5348 0.7313
No log 10.8947 414 0.5431 0.6593 0.5431 0.7370
No log 10.9474 416 0.5798 0.4942 0.5798 0.7615
No log 11.0 418 0.5799 0.5299 0.5799 0.7615
No log 11.0526 420 0.5356 0.6244 0.5356 0.7318
No log 11.1053 422 0.5317 0.6064 0.5317 0.7292
No log 11.1579 424 0.5317 0.6244 0.5317 0.7292
No log 11.2105 426 0.5302 0.5669 0.5302 0.7281
No log 11.2632 428 0.5346 0.5677 0.5346 0.7311
No log 11.3158 430 0.5705 0.4480 0.5705 0.7553
No log 11.3684 432 0.5710 0.4149 0.5710 0.7556
No log 11.4211 434 0.5577 0.4762 0.5577 0.7468
No log 11.4737 436 0.5453 0.4955 0.5453 0.7384
No log 11.5263 438 0.5462 0.5159 0.5462 0.7390
No log 11.5789 440 0.5716 0.5195 0.5716 0.7560
No log 11.6316 442 0.5580 0.5472 0.5580 0.7470
No log 11.6842 444 0.5511 0.6041 0.5511 0.7424
No log 11.7368 446 0.5577 0.6041 0.5577 0.7468
No log 11.7895 448 0.5992 0.5426 0.5992 0.7741
No log 11.8421 450 0.6497 0.4455 0.6497 0.8060
No log 11.8947 452 0.6230 0.3772 0.6230 0.7893
No log 11.9474 454 0.5796 0.5823 0.5796 0.7613
No log 12.0 456 0.5694 0.6083 0.5694 0.7546
No log 12.0526 458 0.5748 0.6083 0.5748 0.7582
No log 12.1053 460 0.5671 0.5840 0.5671 0.7531
No log 12.1579 462 0.6063 0.4693 0.6063 0.7786
No log 12.2105 464 0.6362 0.4290 0.6362 0.7976
No log 12.2632 466 0.5975 0.4692 0.5975 0.7730
No log 12.3158 468 0.5764 0.5891 0.5764 0.7592
No log 12.3684 470 0.5827 0.5961 0.5827 0.7634
No log 12.4211 472 0.5753 0.5891 0.5753 0.7585
No log 12.4737 474 0.5747 0.5891 0.5747 0.7581
No log 12.5263 476 0.5676 0.5701 0.5676 0.7534
No log 12.5789 478 0.5693 0.6307 0.5693 0.7545
No log 12.6316 480 0.5889 0.5867 0.5889 0.7674
No log 12.6842 482 0.6361 0.4811 0.6361 0.7976
No log 12.7368 484 0.6664 0.4470 0.6664 0.8163
No log 12.7895 486 0.7007 0.3810 0.7007 0.8371
No log 12.8421 488 0.6718 0.4195 0.6718 0.8196
No log 12.8947 490 0.6039 0.4817 0.6039 0.7771
No log 12.9474 492 0.5464 0.6388 0.5464 0.7392
No log 13.0 494 0.5599 0.6160 0.5599 0.7483
No log 13.0526 496 0.5986 0.5817 0.5986 0.7737
No log 13.1053 498 0.5675 0.6240 0.5675 0.7534
0.3976 13.1579 500 0.5337 0.6741 0.5337 0.7306
0.3976 13.2105 502 0.5617 0.5190 0.5617 0.7495
0.3976 13.2632 504 0.5656 0.5332 0.5656 0.7521
0.3976 13.3158 506 0.5302 0.6301 0.5302 0.7282
0.3976 13.3684 508 0.5254 0.6289 0.5254 0.7249
0.3976 13.4211 510 0.5370 0.6118 0.5370 0.7328

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k11_task7_organization

Finetuned
(4019)
this model