ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k12_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8597
  • Qwk: 0.3448
  • Mse: 0.8597
  • Rmse: 0.9272

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0526 2 4.7022 0.0010 4.7022 2.1685
No log 0.1053 4 3.1002 -0.0191 3.1002 1.7607
No log 0.1579 6 2.4807 -0.0614 2.4807 1.5750
No log 0.2105 8 1.8480 -0.0364 1.8480 1.3594
No log 0.2632 10 1.2560 0.1080 1.2560 1.1207
No log 0.3158 12 1.2535 0.0714 1.2535 1.1196
No log 0.3684 14 1.3432 0.0210 1.3432 1.1590
No log 0.4211 16 1.2302 0.1638 1.2302 1.1091
No log 0.4737 18 1.3626 0.0561 1.3626 1.1673
No log 0.5263 20 1.2569 0.0952 1.2569 1.1211
No log 0.5789 22 1.1787 0.2053 1.1787 1.0857
No log 0.6316 24 1.1781 0.0914 1.1781 1.0854
No log 0.6842 26 1.2206 0.0952 1.2206 1.1048
No log 0.7368 28 1.3662 0.0688 1.3662 1.1688
No log 0.7895 30 1.4241 0.0750 1.4241 1.1934
No log 0.8421 32 1.2377 0.1320 1.2377 1.1125
No log 0.8947 34 1.1744 0.1830 1.1744 1.0837
No log 0.9474 36 1.1351 0.1335 1.1351 1.0654
No log 1.0 38 1.1803 0.1417 1.1803 1.0864
No log 1.0526 40 1.2175 0.1952 1.2175 1.1034
No log 1.1053 42 1.0943 0.2440 1.0943 1.0461
No log 1.1579 44 1.1053 0.2454 1.1053 1.0514
No log 1.2105 46 1.1117 0.2555 1.1117 1.0543
No log 1.2632 48 1.0601 0.2893 1.0601 1.0296
No log 1.3158 50 1.0479 0.2300 1.0479 1.0237
No log 1.3684 52 1.0577 0.2199 1.0577 1.0284
No log 1.4211 54 1.0828 0.2401 1.0828 1.0406
No log 1.4737 56 1.1318 0.2016 1.1318 1.0638
No log 1.5263 58 1.1646 0.2358 1.1646 1.0792
No log 1.5789 60 1.1776 0.2199 1.1776 1.0852
No log 1.6316 62 1.2283 0.2592 1.2283 1.1083
No log 1.6842 64 1.2266 0.3493 1.2266 1.1075
No log 1.7368 66 1.2049 0.2764 1.2049 1.0977
No log 1.7895 68 1.1107 0.3379 1.1107 1.0539
No log 1.8421 70 1.2926 0.3200 1.2926 1.1369
No log 1.8947 72 1.5014 0.1300 1.5014 1.2253
No log 1.9474 74 1.3504 0.3041 1.3504 1.1621
No log 2.0 76 1.0446 0.3950 1.0446 1.0221
No log 2.0526 78 0.9352 0.4449 0.9352 0.9670
No log 2.1053 80 0.8954 0.4385 0.8954 0.9463
No log 2.1579 82 0.8800 0.4346 0.8800 0.9381
No log 2.2105 84 0.9006 0.4440 0.9006 0.9490
No log 2.2632 86 1.0656 0.5334 1.0656 1.0323
No log 2.3158 88 1.0020 0.4874 1.0020 1.0010
No log 2.3684 90 0.8813 0.5214 0.8813 0.9388
No log 2.4211 92 1.1200 0.4954 1.1200 1.0583
No log 2.4737 94 1.0323 0.4856 1.0323 1.0160
No log 2.5263 96 0.8677 0.5218 0.8677 0.9315
No log 2.5789 98 1.2274 0.4732 1.2274 1.1079
No log 2.6316 100 1.1978 0.4916 1.1978 1.0944
No log 2.6842 102 0.8885 0.5192 0.8885 0.9426
No log 2.7368 104 0.8017 0.5606 0.8017 0.8954
No log 2.7895 106 0.8004 0.5606 0.8004 0.8946
No log 2.8421 108 0.8253 0.5338 0.8253 0.9085
No log 2.8947 110 0.8704 0.5365 0.8704 0.9330
No log 2.9474 112 0.8306 0.5825 0.8306 0.9114
No log 3.0 114 0.9473 0.3847 0.9473 0.9733
No log 3.0526 116 1.1705 0.4524 1.1705 1.0819
No log 3.1053 118 1.2291 0.4404 1.2291 1.1086
No log 3.1579 120 1.0027 0.3738 1.0027 1.0014
No log 3.2105 122 0.9056 0.5673 0.9056 0.9516
No log 3.2632 124 0.9509 0.4236 0.9509 0.9752
No log 3.3158 126 0.9164 0.4241 0.9164 0.9573
No log 3.3684 128 0.9478 0.3966 0.9478 0.9736
No log 3.4211 130 0.9171 0.3861 0.9171 0.9576
No log 3.4737 132 0.8888 0.3896 0.8888 0.9428
No log 3.5263 134 0.9040 0.4002 0.9040 0.9508
No log 3.5789 136 0.9144 0.4465 0.9144 0.9563
No log 3.6316 138 0.9858 0.4412 0.9858 0.9929
No log 3.6842 140 1.3393 0.3481 1.3393 1.1573
No log 3.7368 142 1.5403 0.3239 1.5403 1.2411
No log 3.7895 144 1.2511 0.4186 1.2511 1.1185
No log 3.8421 146 0.9053 0.4846 0.9053 0.9515
No log 3.8947 148 0.9187 0.4042 0.9187 0.9585
No log 3.9474 150 0.9041 0.4175 0.9041 0.9508
No log 4.0 152 0.8842 0.4006 0.8842 0.9403
No log 4.0526 154 1.0290 0.3971 1.0290 1.0144
No log 4.1053 156 1.0742 0.3734 1.0742 1.0365
No log 4.1579 158 1.0335 0.4063 1.0335 1.0166
No log 4.2105 160 0.9044 0.4261 0.9044 0.9510
No log 4.2632 162 0.8267 0.4792 0.8267 0.9093
No log 4.3158 164 0.9927 0.4677 0.9927 0.9963
No log 4.3684 166 1.0646 0.4182 1.0646 1.0318
No log 4.4211 168 0.8752 0.5629 0.8752 0.9355
No log 4.4737 170 0.8894 0.5024 0.8894 0.9431
No log 4.5263 172 0.9471 0.5461 0.9471 0.9732
No log 4.5789 174 0.8625 0.5050 0.8625 0.9287
No log 4.6316 176 0.8360 0.5263 0.8360 0.9143
No log 4.6842 178 1.1067 0.4257 1.1067 1.0520
No log 4.7368 180 1.2777 0.3441 1.2777 1.1304
No log 4.7895 182 1.0841 0.4022 1.0841 1.0412
No log 4.8421 184 0.8597 0.4270 0.8597 0.9272
No log 4.8947 186 0.8727 0.4280 0.8727 0.9342
No log 4.9474 188 0.9548 0.4163 0.9548 0.9772
No log 5.0 190 0.9267 0.3401 0.9267 0.9627
No log 5.0526 192 0.8872 0.4016 0.8872 0.9419
No log 5.1053 194 0.9585 0.4700 0.9585 0.9790
No log 5.1579 196 0.9380 0.4271 0.9380 0.9685
No log 5.2105 198 0.9145 0.4460 0.9145 0.9563
No log 5.2632 200 0.8774 0.4079 0.8774 0.9367
No log 5.3158 202 0.9631 0.4165 0.9631 0.9814
No log 5.3684 204 0.9934 0.4681 0.9934 0.9967
No log 5.4211 206 0.8785 0.3966 0.8785 0.9373
No log 5.4737 208 0.8153 0.4324 0.8153 0.9030
No log 5.5263 210 0.8350 0.4466 0.8350 0.9138
No log 5.5789 212 0.8746 0.3020 0.8746 0.9352
No log 5.6316 214 0.9333 0.3976 0.9333 0.9661
No log 5.6842 216 1.0693 0.3249 1.0693 1.0341
No log 5.7368 218 1.0758 0.3857 1.0758 1.0372
No log 5.7895 220 0.9375 0.4166 0.9375 0.9682
No log 5.8421 222 0.8319 0.5580 0.8319 0.9121
No log 5.8947 224 0.8526 0.5263 0.8526 0.9234
No log 5.9474 226 0.8347 0.5821 0.8347 0.9136
No log 6.0 228 0.8614 0.4889 0.8614 0.9281
No log 6.0526 230 1.0671 0.4567 1.0671 1.0330
No log 6.1053 232 1.1038 0.4453 1.1038 1.0506
No log 6.1579 234 0.9843 0.3876 0.9843 0.9921
No log 6.2105 236 0.8567 0.3812 0.8567 0.9256
No log 6.2632 238 0.8280 0.4760 0.8280 0.9099
No log 6.3158 240 0.8283 0.5257 0.8283 0.9101
No log 6.3684 242 0.8014 0.4691 0.8014 0.8952
No log 6.4211 244 0.8073 0.4637 0.8073 0.8985
No log 6.4737 246 0.8321 0.4840 0.8321 0.9122
No log 6.5263 248 0.9488 0.4490 0.9488 0.9741
No log 6.5789 250 0.9459 0.4501 0.9459 0.9726
No log 6.6316 252 0.9033 0.3941 0.9033 0.9504
No log 6.6842 254 0.8446 0.4181 0.8446 0.9190
No log 6.7368 256 0.8553 0.3596 0.8553 0.9248
No log 6.7895 258 0.8543 0.3896 0.8543 0.9243
No log 6.8421 260 0.8694 0.3998 0.8694 0.9324
No log 6.8947 262 0.9763 0.3686 0.9763 0.9881
No log 6.9474 264 1.0661 0.4426 1.0661 1.0325
No log 7.0 266 1.0551 0.4339 1.0551 1.0272
No log 7.0526 268 0.9330 0.3854 0.9330 0.9659
No log 7.1053 270 0.9117 0.3834 0.9117 0.9548
No log 7.1579 272 0.9450 0.3936 0.9450 0.9721
No log 7.2105 274 1.0609 0.4383 1.0609 1.0300
No log 7.2632 276 1.1707 0.4234 1.1707 1.0820
No log 7.3158 278 1.1134 0.4041 1.1134 1.0552
No log 7.3684 280 1.0797 0.4041 1.0797 1.0391
No log 7.4211 282 0.9443 0.4089 0.9443 0.9717
No log 7.4737 284 0.8924 0.3993 0.8924 0.9447
No log 7.5263 286 0.8919 0.3897 0.8919 0.9444
No log 7.5789 288 0.9905 0.4318 0.9905 0.9952
No log 7.6316 290 1.0288 0.3824 1.0288 1.0143
No log 7.6842 292 1.0238 0.3958 1.0238 1.0118
No log 7.7368 294 1.0206 0.4224 1.0206 1.0102
No log 7.7895 296 0.9465 0.3744 0.9465 0.9729
No log 7.8421 298 0.8785 0.4025 0.8785 0.9373
No log 7.8947 300 0.9046 0.3989 0.9046 0.9511
No log 7.9474 302 0.8946 0.3762 0.8946 0.9459
No log 8.0 304 0.8258 0.4334 0.8258 0.9087
No log 8.0526 306 0.7759 0.5192 0.7759 0.8809
No log 8.1053 308 0.7676 0.5648 0.7676 0.8761
No log 8.1579 310 0.8018 0.5455 0.8018 0.8954
No log 8.2105 312 0.7796 0.5777 0.7796 0.8829
No log 8.2632 314 0.7614 0.5410 0.7614 0.8726
No log 8.3158 316 0.7705 0.5410 0.7705 0.8778
No log 8.3684 318 0.8069 0.4603 0.8069 0.8983
No log 8.4211 320 0.8325 0.4340 0.8325 0.9124
No log 8.4737 322 0.8059 0.3629 0.8059 0.8977
No log 8.5263 324 0.8063 0.4645 0.8063 0.8979
No log 8.5789 326 0.7939 0.4472 0.7939 0.8910
No log 8.6316 328 0.7980 0.4142 0.7980 0.8933
No log 8.6842 330 0.8760 0.5077 0.8760 0.9360
No log 8.7368 332 0.9374 0.4658 0.9374 0.9682
No log 8.7895 334 0.9081 0.4025 0.9081 0.9530
No log 8.8421 336 0.8720 0.3326 0.8720 0.9338
No log 8.8947 338 0.8604 0.3478 0.8604 0.9276
No log 8.9474 340 0.8548 0.3478 0.8548 0.9246
No log 9.0 342 0.8665 0.4260 0.8665 0.9308
No log 9.0526 344 0.8847 0.4260 0.8847 0.9406
No log 9.1053 346 0.8834 0.3299 0.8834 0.9399
No log 9.1579 348 0.8985 0.3734 0.8985 0.9479
No log 9.2105 350 0.9508 0.2917 0.9508 0.9751
No log 9.2632 352 0.9670 0.2791 0.9670 0.9834
No log 9.3158 354 0.9341 0.2917 0.9341 0.9665
No log 9.3684 356 0.8780 0.3326 0.8780 0.9370
No log 9.4211 358 0.8466 0.4180 0.8466 0.9201
No log 9.4737 360 0.8182 0.4180 0.8182 0.9046
No log 9.5263 362 0.8130 0.4428 0.8130 0.9016
No log 9.5789 364 0.8366 0.4242 0.8366 0.9146
No log 9.6316 366 0.9119 0.4031 0.9119 0.9549
No log 9.6842 368 0.9061 0.4164 0.9061 0.9519
No log 9.7368 370 0.8606 0.4102 0.8606 0.9277
No log 9.7895 372 0.8214 0.4620 0.8214 0.9063
No log 9.8421 374 0.8051 0.4620 0.8051 0.8973
No log 9.8947 376 0.7807 0.4995 0.7807 0.8836
No log 9.9474 378 0.7783 0.4620 0.7783 0.8822
No log 10.0 380 0.8402 0.4164 0.8402 0.9166
No log 10.0526 382 0.9671 0.4526 0.9671 0.9834
No log 10.1053 384 0.9964 0.4532 0.9964 0.9982
No log 10.1579 386 0.9692 0.4085 0.9692 0.9845
No log 10.2105 388 0.9238 0.3609 0.9238 0.9611
No log 10.2632 390 0.8810 0.3335 0.8810 0.9386
No log 10.3158 392 0.8212 0.3583 0.8212 0.9062
No log 10.3684 394 0.7703 0.4367 0.7703 0.8776
No log 10.4211 396 0.7517 0.4784 0.7517 0.8670
No log 10.4737 398 0.7452 0.5417 0.7452 0.8632
No log 10.5263 400 0.7131 0.5125 0.7131 0.8445
No log 10.5789 402 0.7133 0.5450 0.7133 0.8446
No log 10.6316 404 0.7372 0.5498 0.7372 0.8586
No log 10.6842 406 0.7317 0.4879 0.7317 0.8554
No log 10.7368 408 0.7471 0.4879 0.7471 0.8644
No log 10.7895 410 0.7350 0.4656 0.7350 0.8573
No log 10.8421 412 0.7475 0.4879 0.7475 0.8646
No log 10.8947 414 0.7797 0.4620 0.7797 0.8830
No log 10.9474 416 0.7871 0.4620 0.7871 0.8872
No log 11.0 418 0.7695 0.5094 0.7695 0.8772
No log 11.0526 420 0.7826 0.5029 0.7826 0.8847
No log 11.1053 422 0.8887 0.5002 0.8887 0.9427
No log 11.1579 424 0.9673 0.4590 0.9673 0.9835
No log 11.2105 426 0.9353 0.4596 0.9353 0.9671
No log 11.2632 428 0.8979 0.5135 0.8979 0.9476
No log 11.3158 430 0.8643 0.4087 0.8643 0.9297
No log 11.3684 432 0.8301 0.3891 0.8301 0.9111
No log 11.4211 434 0.8365 0.4028 0.8365 0.9146
No log 11.4737 436 0.8389 0.4028 0.8389 0.9159
No log 11.5263 438 0.8693 0.4473 0.8693 0.9324
No log 11.5789 440 0.9144 0.4649 0.9144 0.9562
No log 11.6316 442 0.9172 0.3667 0.9172 0.9577
No log 11.6842 444 0.8752 0.2865 0.8752 0.9355
No log 11.7368 446 0.8483 0.3095 0.8483 0.9210
No log 11.7895 448 0.8390 0.3510 0.8390 0.9160
No log 11.8421 450 0.8341 0.3095 0.8341 0.9133
No log 11.8947 452 0.8694 0.3700 0.8694 0.9324
No log 11.9474 454 0.9210 0.3270 0.9210 0.9597
No log 12.0 456 0.9327 0.3270 0.9327 0.9658
No log 12.0526 458 0.8963 0.3551 0.8963 0.9468
No log 12.1053 460 0.8278 0.3616 0.8278 0.9098
No log 12.1579 462 0.8131 0.3859 0.8131 0.9017
No log 12.2105 464 0.8218 0.3859 0.8218 0.9065
No log 12.2632 466 0.8384 0.3510 0.8384 0.9156
No log 12.3158 468 0.9346 0.2304 0.9346 0.9667
No log 12.3684 470 0.9864 0.2843 0.9864 0.9932
No log 12.4211 472 0.9584 0.2408 0.9584 0.9790
No log 12.4737 474 0.9114 0.2887 0.9114 0.9547
No log 12.5263 476 0.8388 0.3510 0.8388 0.9159
No log 12.5789 478 0.8178 0.3779 0.8178 0.9043
No log 12.6316 480 0.8525 0.4243 0.8525 0.9233
No log 12.6842 482 0.8852 0.3827 0.8852 0.9408
No log 12.7368 484 0.9252 0.3547 0.9252 0.9618
No log 12.7895 486 0.9377 0.3289 0.9377 0.9683
No log 12.8421 488 0.9130 0.2605 0.9130 0.9555
No log 12.8947 490 0.8774 0.3616 0.8774 0.9367
No log 12.9474 492 0.8737 0.3616 0.8737 0.9347
No log 13.0 494 0.8605 0.3478 0.8605 0.9276
No log 13.0526 496 0.8428 0.3957 0.8428 0.9180
No log 13.1053 498 0.8094 0.4488 0.8094 0.8997
0.3468 13.1579 500 0.7693 0.4488 0.7693 0.8771
0.3468 13.2105 502 0.7528 0.4737 0.7528 0.8676
0.3468 13.2632 504 0.7559 0.4993 0.7559 0.8694
0.3468 13.3158 506 0.7708 0.4321 0.7708 0.8780
0.3468 13.3684 508 0.7995 0.4526 0.7995 0.8941
0.3468 13.4211 510 0.8567 0.3931 0.8567 0.9256
0.3468 13.4737 512 0.9062 0.4552 0.9062 0.9519
0.3468 13.5263 514 0.8808 0.4040 0.8808 0.9385
0.3468 13.5789 516 0.8150 0.4334 0.8150 0.9028
0.3468 13.6316 518 0.7762 0.4413 0.7762 0.8810
0.3468 13.6842 520 0.7639 0.5027 0.7639 0.8740
0.3468 13.7368 522 0.7587 0.5186 0.7587 0.8710
0.3468 13.7895 524 0.7897 0.4662 0.7897 0.8887
0.3468 13.8421 526 0.9147 0.5029 0.9147 0.9564
0.3468 13.8947 528 1.0517 0.4839 1.0517 1.0255
0.3468 13.9474 530 1.0572 0.4850 1.0572 1.0282
0.3468 14.0 532 0.9970 0.4278 0.9970 0.9985
0.3468 14.0526 534 0.9656 0.4186 0.9656 0.9827
0.3468 14.1053 536 0.9243 0.3702 0.9243 0.9614
0.3468 14.1579 538 0.8706 0.3250 0.8706 0.9331
0.3468 14.2105 540 0.8597 0.3448 0.8597 0.9272

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k12_task2_organization

Finetuned
(4023)
this model