ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k3_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8997
  • Qwk: 0.3506
  • Mse: 0.8997
  • Rmse: 0.9485

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1667 2 3.9498 -0.0092 3.9498 1.9874
No log 0.3333 4 2.3184 0.0458 2.3184 1.5226
No log 0.5 6 1.6172 0.0203 1.6172 1.2717
No log 0.6667 8 1.1891 0.2409 1.1891 1.0905
No log 0.8333 10 1.1480 0.1447 1.1480 1.0714
No log 1.0 12 1.4367 -0.1723 1.4367 1.1986
No log 1.1667 14 1.4062 -0.1723 1.4062 1.1858
No log 1.3333 16 1.2486 0.0293 1.2486 1.1174
No log 1.5 18 1.2032 0.0510 1.2032 1.0969
No log 1.6667 20 1.1397 0.0950 1.1397 1.0676
No log 1.8333 22 1.1040 0.1284 1.1040 1.0507
No log 2.0 24 1.0984 0.2391 1.0984 1.0481
No log 2.1667 26 1.0753 0.1997 1.0753 1.0370
No log 2.3333 28 1.1075 0.1240 1.1075 1.0524
No log 2.5 30 1.1261 0.1416 1.1261 1.0612
No log 2.6667 32 1.0701 0.1603 1.0701 1.0345
No log 2.8333 34 1.1455 0.0672 1.1455 1.0703
No log 3.0 36 1.2468 0.1142 1.2468 1.1166
No log 3.1667 38 1.1044 0.1028 1.1044 1.0509
No log 3.3333 40 1.0260 0.2819 1.0260 1.0129
No log 3.5 42 1.2137 0.3571 1.2137 1.1017
No log 3.6667 44 1.0819 0.3222 1.0819 1.0401
No log 3.8333 46 0.9560 0.2910 0.9560 0.9777
No log 4.0 48 0.9549 0.2596 0.9549 0.9772
No log 4.1667 50 0.9896 0.2842 0.9896 0.9948
No log 4.3333 52 1.0492 0.3543 1.0492 1.0243
No log 4.5 54 1.0133 0.3293 1.0133 1.0066
No log 4.6667 56 0.9882 0.2933 0.9882 0.9941
No log 4.8333 58 1.0347 0.2834 1.0347 1.0172
No log 5.0 60 0.9738 0.2812 0.9738 0.9868
No log 5.1667 62 0.9905 0.3493 0.9905 0.9953
No log 5.3333 64 1.0621 0.3992 1.0621 1.0306
No log 5.5 66 1.0046 0.2974 1.0046 1.0023
No log 5.6667 68 1.0004 0.3211 1.0004 1.0002
No log 5.8333 70 1.1116 0.3158 1.1116 1.0543
No log 6.0 72 1.0993 0.3237 1.0993 1.0485
No log 6.1667 74 0.9498 0.3078 0.9498 0.9746
No log 6.3333 76 1.0064 0.3959 1.0064 1.0032
No log 6.5 78 1.0303 0.3543 1.0303 1.0150
No log 6.6667 80 0.9152 0.3112 0.9152 0.9567
No log 6.8333 82 0.9000 0.3467 0.9000 0.9487
No log 7.0 84 0.9004 0.3446 0.9004 0.9489
No log 7.1667 86 1.0037 0.3556 1.0037 1.0018
No log 7.3333 88 1.2115 0.1597 1.2115 1.1007
No log 7.5 90 1.1168 0.2171 1.1168 1.0568
No log 7.6667 92 1.0016 0.1573 1.0016 1.0008
No log 7.8333 94 1.0050 0.1893 1.0050 1.0025
No log 8.0 96 0.9709 0.3631 0.9709 0.9853
No log 8.1667 98 0.9284 0.2229 0.9284 0.9635
No log 8.3333 100 0.9176 0.3293 0.9176 0.9579
No log 8.5 102 0.9181 0.3209 0.9181 0.9582
No log 8.6667 104 0.9497 0.2254 0.9497 0.9745
No log 8.8333 106 0.9563 0.2910 0.9563 0.9779
No log 9.0 108 0.9980 0.3188 0.9980 0.9990
No log 9.1667 110 0.9879 0.3070 0.9879 0.9939
No log 9.3333 112 0.9751 0.1867 0.9751 0.9875
No log 9.5 114 1.0413 0.3066 1.0413 1.0204
No log 9.6667 116 0.9659 0.2923 0.9659 0.9828
No log 9.8333 118 0.9320 0.3474 0.9320 0.9654
No log 10.0 120 1.0570 0.4068 1.0570 1.0281
No log 10.1667 122 1.0312 0.3577 1.0312 1.0155
No log 10.3333 124 0.8946 0.3721 0.8946 0.9459
No log 10.5 126 0.9153 0.3700 0.9153 0.9567
No log 10.6667 128 0.9133 0.3056 0.9133 0.9557
No log 10.8333 130 0.9237 0.2865 0.9237 0.9611
No log 11.0 132 0.9361 0.3583 0.9361 0.9675
No log 11.1667 134 0.9333 0.2528 0.9333 0.9661
No log 11.3333 136 0.9188 0.2473 0.9188 0.9586
No log 11.5 138 0.9040 0.2911 0.9040 0.9508
No log 11.6667 140 0.9545 0.2593 0.9545 0.9770
No log 11.8333 142 0.9881 0.2616 0.9881 0.9940
No log 12.0 144 0.9298 0.2667 0.9298 0.9643
No log 12.1667 146 0.9416 0.4257 0.9416 0.9704
No log 12.3333 148 0.9460 0.4241 0.9460 0.9726
No log 12.5 150 0.9031 0.2932 0.9031 0.9503
No log 12.6667 152 0.9159 0.2879 0.9159 0.9570
No log 12.8333 154 0.9184 0.3196 0.9184 0.9583
No log 13.0 156 0.8766 0.3661 0.8766 0.9363
No log 13.1667 158 0.8709 0.3622 0.8709 0.9332
No log 13.3333 160 0.8895 0.3544 0.8895 0.9431
No log 13.5 162 0.8951 0.3257 0.8951 0.9461
No log 13.6667 164 0.8865 0.3215 0.8865 0.9415
No log 13.8333 166 0.8808 0.3229 0.8808 0.9385
No log 14.0 168 0.8727 0.3236 0.8727 0.9342
No log 14.1667 170 0.8932 0.3564 0.8932 0.9451
No log 14.3333 172 0.9011 0.3700 0.9011 0.9493
No log 14.5 174 0.8750 0.2788 0.8750 0.9354
No log 14.6667 176 0.8981 0.3941 0.8981 0.9477
No log 14.8333 178 0.9620 0.3264 0.9620 0.9808
No log 15.0 180 0.9409 0.3802 0.9409 0.9700
No log 15.1667 182 0.9159 0.4336 0.9159 0.9571
No log 15.3333 184 0.9123 0.4336 0.9123 0.9552
No log 15.5 186 0.8879 0.3799 0.8879 0.9423
No log 15.6667 188 0.8704 0.2812 0.8704 0.9330
No log 15.8333 190 0.8687 0.2974 0.8687 0.9320
No log 16.0 192 0.8868 0.4713 0.8868 0.9417
No log 16.1667 194 0.8984 0.4713 0.8984 0.9479
No log 16.3333 196 0.8429 0.3398 0.8429 0.9181
No log 16.5 198 0.8304 0.3685 0.8304 0.9112
No log 16.6667 200 0.8275 0.3476 0.8275 0.9097
No log 16.8333 202 0.8296 0.3476 0.8296 0.9108
No log 17.0 204 0.8351 0.3033 0.8351 0.9138
No log 17.1667 206 0.8504 0.3819 0.8504 0.9222
No log 17.3333 208 0.8248 0.3717 0.8248 0.9082
No log 17.5 210 0.8406 0.4277 0.8406 0.9168
No log 17.6667 212 0.8406 0.4866 0.8406 0.9169
No log 17.8333 214 0.8142 0.5010 0.8142 0.9023
No log 18.0 216 0.9927 0.4472 0.9927 0.9964
No log 18.1667 218 1.1513 0.4332 1.1513 1.0730
No log 18.3333 220 1.0477 0.4115 1.0477 1.0236
No log 18.5 222 0.8788 0.3958 0.8788 0.9374
No log 18.6667 224 0.8510 0.3476 0.8510 0.9225
No log 18.8333 226 0.8806 0.3222 0.8806 0.9384
No log 19.0 228 0.8736 0.3056 0.8736 0.9347
No log 19.1667 230 0.8650 0.3011 0.8650 0.9301
No log 19.3333 232 0.8838 0.3112 0.8838 0.9401
No log 19.5 234 0.9387 0.3207 0.9387 0.9689
No log 19.6667 236 0.9466 0.3523 0.9466 0.9730
No log 19.8333 238 0.9385 0.2647 0.9385 0.9688
No log 20.0 240 0.9094 0.2988 0.9094 0.9536
No log 20.1667 242 0.9068 0.2988 0.9068 0.9523
No log 20.3333 244 0.9255 0.3506 0.9255 0.9620
No log 20.5 246 0.9552 0.2771 0.9552 0.9773
No log 20.6667 248 0.9737 0.3063 0.9737 0.9868
No log 20.8333 250 0.9504 0.4075 0.9504 0.9749
No log 21.0 252 0.8852 0.3393 0.8852 0.9409
No log 21.1667 254 0.8481 0.4097 0.8481 0.9209
No log 21.3333 256 0.8392 0.3838 0.8392 0.9161
No log 21.5 258 0.8349 0.4241 0.8349 0.9137
No log 21.6667 260 0.8311 0.3548 0.8311 0.9116
No log 21.8333 262 0.8302 0.3942 0.8302 0.9112
No log 22.0 264 0.8472 0.4204 0.8472 0.9204
No log 22.1667 266 0.8945 0.4958 0.8945 0.9458
No log 22.3333 268 0.8830 0.4973 0.8830 0.9397
No log 22.5 270 0.8822 0.4730 0.8822 0.9392
No log 22.6667 272 0.8875 0.4730 0.8875 0.9421
No log 22.8333 274 0.8949 0.3631 0.8949 0.9460
No log 23.0 276 0.8780 0.3922 0.8780 0.9370
No log 23.1667 278 0.8621 0.3797 0.8621 0.9285
No log 23.3333 280 0.8539 0.3403 0.8539 0.9240
No log 23.5 282 0.8553 0.3569 0.8553 0.9248
No log 23.6667 284 0.8490 0.3610 0.8490 0.9214
No log 23.8333 286 0.8389 0.3301 0.8389 0.9159
No log 24.0 288 0.8332 0.3301 0.8332 0.9128
No log 24.1667 290 0.8272 0.3301 0.8272 0.9095
No log 24.3333 292 0.8297 0.3733 0.8297 0.9109
No log 24.5 294 0.8270 0.3693 0.8270 0.9094
No log 24.6667 296 0.8293 0.3693 0.8293 0.9106
No log 24.8333 298 0.8341 0.3403 0.8341 0.9133
No log 25.0 300 0.8464 0.3797 0.8464 0.9200
No log 25.1667 302 0.8360 0.3403 0.8360 0.9143
No log 25.3333 304 0.8401 0.3200 0.8401 0.9166
No log 25.5 306 0.8753 0.2366 0.8753 0.9356
No log 25.6667 308 0.8770 0.2366 0.8770 0.9365
No log 25.8333 310 0.8518 0.3200 0.8518 0.9229
No log 26.0 312 0.8480 0.3548 0.8480 0.9209
No log 26.1667 314 0.8510 0.3697 0.8510 0.9225
No log 26.3333 316 0.8443 0.3403 0.8443 0.9188
No log 26.5 318 0.8355 0.3569 0.8355 0.9141
No log 26.6667 320 0.8312 0.3583 0.8312 0.9117
No log 26.8333 322 0.8171 0.4248 0.8171 0.9039
No log 27.0 324 0.8187 0.3673 0.8187 0.9048
No log 27.1667 326 0.8196 0.3673 0.8196 0.9053
No log 27.3333 328 0.8133 0.3673 0.8133 0.9018
No log 27.5 330 0.8096 0.3817 0.8096 0.8998
No log 27.6667 332 0.8243 0.3528 0.8243 0.9079
No log 27.8333 334 0.8353 0.4204 0.8353 0.9139
No log 28.0 336 0.8506 0.3922 0.8506 0.9223
No log 28.1667 338 0.8407 0.3922 0.8407 0.9169
No log 28.3333 340 0.8223 0.3693 0.8223 0.9068
No log 28.5 342 0.8233 0.3837 0.8233 0.9074
No log 28.6667 344 0.8146 0.3837 0.8146 0.9025
No log 28.8333 346 0.8260 0.3979 0.8260 0.9088
No log 29.0 348 0.8281 0.3979 0.8281 0.9100
No log 29.1667 350 0.8279 0.2956 0.8279 0.9099
No log 29.3333 352 0.8555 0.3243 0.8555 0.9249
No log 29.5 354 0.8617 0.2834 0.8617 0.9283
No log 29.6667 356 0.8877 0.3403 0.8877 0.9422
No log 29.8333 358 0.9507 0.3663 0.9507 0.9751
No log 30.0 360 0.9879 0.3531 0.9879 0.9940
No log 30.1667 362 0.9674 0.3383 0.9674 0.9836
No log 30.3333 364 0.9059 0.3506 0.9059 0.9518
No log 30.5 366 0.8818 0.3403 0.8818 0.9390
No log 30.6667 368 0.8709 0.3403 0.8709 0.9332
No log 30.8333 370 0.8960 0.4729 0.8960 0.9466
No log 31.0 372 0.9517 0.4335 0.9517 0.9755
No log 31.1667 374 0.9975 0.4098 0.9975 0.9988
No log 31.3333 376 0.9969 0.4681 0.9969 0.9984
No log 31.5 378 0.9531 0.3782 0.9531 0.9763
No log 31.6667 380 0.8894 0.3922 0.8894 0.9431
No log 31.8333 382 0.8537 0.3266 0.8537 0.9240
No log 32.0 384 0.8646 0.2366 0.8646 0.9299
No log 32.1667 386 0.8640 0.2366 0.8640 0.9295
No log 32.3333 388 0.8491 0.2888 0.8491 0.9215
No log 32.5 390 0.8515 0.3403 0.8515 0.9227
No log 32.6667 392 0.8877 0.3941 0.8877 0.9422
No log 32.8333 394 0.9336 0.3921 0.9336 0.9662
No log 33.0 396 0.9479 0.2674 0.9479 0.9736
No log 33.1667 398 0.9374 0.2819 0.9374 0.9682
No log 33.3333 400 0.9189 0.3372 0.9189 0.9586
No log 33.5 402 0.8885 0.3631 0.8885 0.9426
No log 33.6667 404 0.8574 0.3797 0.8574 0.9260
No log 33.8333 406 0.8415 0.3961 0.8415 0.9173
No log 34.0 408 0.8424 0.2956 0.8424 0.9178
No log 34.1667 410 0.8427 0.2956 0.8427 0.9180
No log 34.3333 412 0.8318 0.3713 0.8318 0.9121
No log 34.5 414 0.8569 0.4490 0.8569 0.9257
No log 34.6667 416 0.9206 0.4595 0.9206 0.9595
No log 34.8333 418 0.9635 0.4828 0.9635 0.9816
No log 35.0 420 0.9528 0.4318 0.9528 0.9761
No log 35.1667 422 0.9251 0.3506 0.9251 0.9618
No log 35.3333 424 0.9049 0.3922 0.9049 0.9513
No log 35.5 426 0.8927 0.3797 0.8927 0.9448
No log 35.6667 428 0.8856 0.3797 0.8856 0.9411
No log 35.8333 430 0.8969 0.3922 0.8969 0.9470
No log 36.0 432 0.9392 0.3631 0.9392 0.9691
No log 36.1667 434 1.0144 0.2819 1.0144 1.0072
No log 36.3333 436 1.0267 0.3063 1.0267 1.0132
No log 36.5 438 0.9861 0.2771 0.9861 0.9930
No log 36.6667 440 0.9671 0.2771 0.9671 0.9834
No log 36.8333 442 0.9717 0.2771 0.9717 0.9857
No log 37.0 444 0.9838 0.2771 0.9838 0.9919
No log 37.1667 446 0.9744 0.2771 0.9744 0.9871
No log 37.3333 448 0.9426 0.3198 0.9426 0.9709
No log 37.5 450 0.9437 0.3198 0.9437 0.9715
No log 37.6667 452 0.9278 0.3198 0.9278 0.9632
No log 37.8333 454 0.9303 0.3506 0.9303 0.9645
No log 38.0 456 0.9197 0.3922 0.9197 0.9590
No log 38.1667 458 0.9212 0.3506 0.9212 0.9598
No log 38.3333 460 0.9395 0.3089 0.9395 0.9693
No log 38.5 462 0.9654 0.3214 0.9654 0.9826
No log 38.6667 464 0.9814 0.3089 0.9814 0.9907
No log 38.8333 466 0.9725 0.3089 0.9725 0.9861
No log 39.0 468 0.9478 0.3089 0.9478 0.9736
No log 39.1667 470 0.9064 0.3797 0.9064 0.9520
No log 39.3333 472 0.8911 0.3548 0.8911 0.9440
No log 39.5 474 0.8850 0.3548 0.8850 0.9407
No log 39.6667 476 0.8978 0.3548 0.8978 0.9475
No log 39.8333 478 0.9208 0.3797 0.9208 0.9596
No log 40.0 480 0.9554 0.2671 0.9554 0.9775
No log 40.1667 482 0.9818 0.2698 0.9818 0.9909
No log 40.3333 484 0.9855 0.2698 0.9855 0.9927
No log 40.5 486 0.9646 0.2671 0.9646 0.9822
No log 40.6667 488 0.9327 0.3797 0.9327 0.9657
No log 40.8333 490 0.9120 0.3243 0.9120 0.9550
No log 41.0 492 0.8990 0.3243 0.8990 0.9482
No log 41.1667 494 0.8883 0.3117 0.8883 0.9425
No log 41.3333 496 0.8839 0.3243 0.8839 0.9401
No log 41.5 498 0.8989 0.4067 0.8989 0.9481
0.2088 41.6667 500 0.9187 0.3506 0.9187 0.9585
0.2088 41.8333 502 0.9288 0.3506 0.9288 0.9638
0.2088 42.0 504 0.9304 0.3506 0.9304 0.9646
0.2088 42.1667 506 0.9410 0.3089 0.9410 0.9701
0.2088 42.3333 508 0.9253 0.3506 0.9253 0.9619
0.2088 42.5 510 0.8997 0.3506 0.8997 0.9485

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k3_task5_organization

Finetuned
(4019)
this model