ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k15_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0338
  • Qwk: 0.2263
  • Mse: 1.0338
  • Rmse: 1.0168

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0278 2 3.8015 -0.0387 3.8015 1.9497
No log 0.0556 4 2.0439 -0.0139 2.0439 1.4296
No log 0.0833 6 1.3448 0.0302 1.3448 1.1596
No log 0.1111 8 1.1231 0.1810 1.1231 1.0598
No log 0.1389 10 1.1408 0.1962 1.1408 1.0681
No log 0.1667 12 1.5141 -0.0197 1.5141 1.2305
No log 0.1944 14 1.5461 0.0455 1.5461 1.2434
No log 0.2222 16 1.2503 0.2247 1.2503 1.1182
No log 0.25 18 0.9803 0.2239 0.9803 0.9901
No log 0.2778 20 1.0360 0.1927 1.0360 1.0179
No log 0.3056 22 0.9874 0.2251 0.9874 0.9937
No log 0.3333 24 1.0643 0.2711 1.0643 1.0317
No log 0.3611 26 1.1605 0.1979 1.1605 1.0773
No log 0.3889 28 1.2748 0.2752 1.2748 1.1291
No log 0.4167 30 1.0840 0.2343 1.0840 1.0411
No log 0.4444 32 0.9507 0.3435 0.9507 0.9750
No log 0.4722 34 1.1055 0.2095 1.1055 1.0514
No log 0.5 36 1.0928 0.2611 1.0928 1.0454
No log 0.5278 38 0.9822 0.3243 0.9822 0.9911
No log 0.5556 40 1.2409 0.3108 1.2409 1.1140
No log 0.5833 42 1.4121 0.2283 1.4121 1.1883
No log 0.6111 44 1.2733 0.2376 1.2733 1.1284
No log 0.6389 46 1.0462 0.3578 1.0462 1.0228
No log 0.6667 48 0.9756 0.3447 0.9756 0.9877
No log 0.6944 50 0.9923 0.3256 0.9923 0.9961
No log 0.7222 52 1.0149 0.2972 1.0149 1.0074
No log 0.75 54 0.9950 0.2911 0.9950 0.9975
No log 0.7778 56 1.0027 0.3332 1.0027 1.0014
No log 0.8056 58 1.0155 0.4048 1.0155 1.0077
No log 0.8333 60 0.9733 0.3785 0.9733 0.9865
No log 0.8611 62 1.1057 0.2172 1.1057 1.0515
No log 0.8889 64 1.6857 0.0999 1.6857 1.2984
No log 0.9167 66 1.9563 -0.0509 1.9563 1.3987
No log 0.9444 68 1.5470 0.1321 1.5470 1.2438
No log 0.9722 70 1.0577 0.3217 1.0577 1.0284
No log 1.0 72 0.9233 0.3964 0.9233 0.9609
No log 1.0278 74 0.9194 0.4410 0.9194 0.9588
No log 1.0556 76 0.8983 0.3806 0.8983 0.9478
No log 1.0833 78 0.9662 0.4094 0.9662 0.9830
No log 1.1111 80 1.0201 0.3400 1.0201 1.0100
No log 1.1389 82 1.0060 0.2944 1.0060 1.0030
No log 1.1667 84 1.0013 0.3129 1.0013 1.0007
No log 1.1944 86 1.0986 0.3316 1.0986 1.0481
No log 1.2222 88 1.0418 0.2747 1.0418 1.0207
No log 1.25 90 0.9471 0.3958 0.9471 0.9732
No log 1.2778 92 0.8995 0.3959 0.8995 0.9484
No log 1.3056 94 0.8822 0.4647 0.8822 0.9392
No log 1.3333 96 0.8803 0.4720 0.8803 0.9383
No log 1.3611 98 0.8578 0.4553 0.8578 0.9262
No log 1.3889 100 0.8677 0.4244 0.8677 0.9315
No log 1.4167 102 0.8913 0.4244 0.8913 0.9441
No log 1.4444 104 0.9121 0.3744 0.9121 0.9550
No log 1.4722 106 0.9494 0.3968 0.9494 0.9744
No log 1.5 108 0.9366 0.4493 0.9366 0.9678
No log 1.5278 110 0.8620 0.4691 0.8620 0.9284
No log 1.5556 112 0.8280 0.4514 0.8280 0.9099
No log 1.5833 114 0.8415 0.4155 0.8415 0.9173
No log 1.6111 116 0.9017 0.4218 0.9017 0.9496
No log 1.6389 118 0.9059 0.4433 0.9059 0.9518
No log 1.6667 120 0.9022 0.4428 0.9022 0.9498
No log 1.6944 122 0.9150 0.4428 0.9150 0.9565
No log 1.7222 124 0.8834 0.4889 0.8834 0.9399
No log 1.75 126 0.9270 0.3945 0.9270 0.9628
No log 1.7778 128 1.0333 0.3881 1.0333 1.0165
No log 1.8056 130 1.0347 0.3897 1.0347 1.0172
No log 1.8333 132 1.0385 0.3778 1.0385 1.0191
No log 1.8611 134 0.9146 0.3615 0.9146 0.9563
No log 1.8889 136 0.8835 0.3367 0.8835 0.9400
No log 1.9167 138 0.8656 0.3465 0.8656 0.9304
No log 1.9444 140 0.8810 0.3687 0.8810 0.9386
No log 1.9722 142 0.9071 0.3668 0.9071 0.9524
No log 2.0 144 0.8804 0.3668 0.8804 0.9383
No log 2.0278 146 0.8460 0.3272 0.8460 0.9198
No log 2.0556 148 0.8546 0.3552 0.8546 0.9244
No log 2.0833 150 0.8573 0.3820 0.8573 0.9259
No log 2.1111 152 0.8632 0.4264 0.8632 0.9291
No log 2.1389 154 0.9885 0.4579 0.9885 0.9942
No log 2.1667 156 1.0313 0.4356 1.0313 1.0155
No log 2.1944 158 0.9654 0.4586 0.9654 0.9825
No log 2.2222 160 0.9034 0.3465 0.9034 0.9505
No log 2.25 162 0.9010 0.5312 0.9010 0.9492
No log 2.2778 164 1.1209 0.3961 1.1209 1.0587
No log 2.3056 166 1.0496 0.4510 1.0496 1.0245
No log 2.3333 168 0.8677 0.5021 0.8677 0.9315
No log 2.3611 170 0.9002 0.4395 0.9002 0.9488
No log 2.3889 172 1.1106 0.3847 1.1106 1.0539
No log 2.4167 174 1.1831 0.3021 1.1831 1.0877
No log 2.4444 176 1.1091 0.3314 1.1091 1.0531
No log 2.4722 178 1.0940 0.3578 1.0940 1.0460
No log 2.5 180 1.1281 0.3298 1.1281 1.0621
No log 2.5278 182 1.1745 0.3624 1.1745 1.0837
No log 2.5556 184 1.1962 0.3624 1.1962 1.0937
No log 2.5833 186 1.1050 0.3641 1.1050 1.0512
No log 2.6111 188 1.0856 0.3596 1.0856 1.0419
No log 2.6389 190 1.1782 0.3578 1.1782 1.0855
No log 2.6667 192 1.3274 0.3383 1.3274 1.1521
No log 2.6944 194 1.3447 0.3158 1.3447 1.1596
No log 2.7222 196 1.2545 0.3344 1.2545 1.1200
No log 2.75 198 1.1711 0.3741 1.1711 1.0822
No log 2.7778 200 1.1245 0.3721 1.1245 1.0604
No log 2.8056 202 1.1375 0.3699 1.1375 1.0665
No log 2.8333 204 1.1968 0.3553 1.1968 1.0940
No log 2.8611 206 1.2046 0.3553 1.2046 1.0975
No log 2.8889 208 1.1245 0.3790 1.1245 1.0604
No log 2.9167 210 0.9902 0.3584 0.9902 0.9951
No log 2.9444 212 0.9525 0.3551 0.9525 0.9760
No log 2.9722 214 0.9699 0.2723 0.9699 0.9848
No log 3.0 216 1.0296 0.2886 1.0296 1.0147
No log 3.0278 218 1.1649 0.3391 1.1649 1.0793
No log 3.0556 220 1.2851 0.3703 1.2851 1.1336
No log 3.0833 222 1.4030 0.3722 1.4030 1.1845
No log 3.1111 224 1.2955 0.3755 1.2955 1.1382
No log 3.1389 226 1.0725 0.4156 1.0725 1.0356
No log 3.1667 228 0.9940 0.4299 0.9940 0.9970
No log 3.1944 230 1.0304 0.4470 1.0304 1.0151
No log 3.2222 232 1.0292 0.4111 1.0292 1.0145
No log 3.25 234 0.9904 0.3298 0.9904 0.9952
No log 3.2778 236 0.9849 0.3687 0.9849 0.9924
No log 3.3056 238 1.0359 0.3863 1.0359 1.0178
No log 3.3333 240 1.0654 0.3913 1.0654 1.0322
No log 3.3611 242 1.1143 0.3928 1.1143 1.0556
No log 3.3889 244 1.0624 0.3928 1.0624 1.0308
No log 3.4167 246 1.0131 0.3826 1.0131 1.0065
No log 3.4444 248 0.9719 0.3124 0.9719 0.9858
No log 3.4722 250 0.9948 0.3124 0.9948 0.9974
No log 3.5 252 1.0560 0.3826 1.0560 1.0276
No log 3.5278 254 1.0755 0.3880 1.0755 1.0370
No log 3.5556 256 1.0567 0.3880 1.0567 1.0279
No log 3.5833 258 1.0034 0.4218 1.0034 1.0017
No log 3.6111 260 0.9569 0.4444 0.9569 0.9782
No log 3.6389 262 0.9080 0.3902 0.9080 0.9529
No log 3.6667 264 0.8888 0.4169 0.8888 0.9428
No log 3.6944 266 0.8957 0.3902 0.8957 0.9464
No log 3.7222 268 0.8881 0.3902 0.8881 0.9424
No log 3.75 270 0.9036 0.3902 0.9036 0.9506
No log 3.7778 272 0.9149 0.4115 0.9149 0.9565
No log 3.8056 274 0.9242 0.4166 0.9242 0.9614
No log 3.8333 276 0.8942 0.4175 0.8942 0.9456
No log 3.8611 278 0.9046 0.3914 0.9046 0.9511
No log 3.8889 280 0.8962 0.3914 0.8962 0.9467
No log 3.9167 282 0.8550 0.3781 0.8550 0.9247
No log 3.9444 284 0.8940 0.4984 0.8940 0.9455
No log 3.9722 286 0.9259 0.4444 0.9259 0.9622
No log 4.0 288 0.8617 0.4208 0.8617 0.9283
No log 4.0278 290 0.8981 0.3914 0.8981 0.9477
No log 4.0556 292 0.8771 0.3897 0.8771 0.9365
No log 4.0833 294 0.8557 0.4025 0.8557 0.9250
No log 4.1111 296 0.8476 0.3896 0.8476 0.9206
No log 4.1389 298 0.8567 0.3879 0.8567 0.9256
No log 4.1667 300 0.8764 0.3897 0.8764 0.9362
No log 4.1944 302 0.8395 0.3897 0.8395 0.9163
No log 4.2222 304 0.8397 0.3358 0.8397 0.9164
No log 4.25 306 0.8780 0.3223 0.8780 0.9370
No log 4.2778 308 0.9463 0.3861 0.9463 0.9728
No log 4.3056 310 0.8964 0.3202 0.8964 0.9468
No log 4.3333 312 0.8709 0.3795 0.8709 0.9332
No log 4.3611 314 0.9716 0.4396 0.9716 0.9857
No log 4.3889 316 0.9860 0.4140 0.9860 0.9930
No log 4.4167 318 0.9226 0.3236 0.9226 0.9605
No log 4.4444 320 0.9897 0.3124 0.9897 0.9948
No log 4.4722 322 1.2114 0.3553 1.2114 1.1007
No log 4.5 324 1.2566 0.3601 1.2566 1.1210
No log 4.5278 326 1.1946 0.3829 1.1946 1.0930
No log 4.5556 328 1.0338 0.3558 1.0338 1.0167
No log 4.5833 330 0.9715 0.3842 0.9715 0.9856
No log 4.6111 332 0.9427 0.3338 0.9427 0.9709
No log 4.6389 334 0.9341 0.3822 0.9341 0.9665
No log 4.6667 336 0.9409 0.3957 0.9409 0.9700
No log 4.6944 338 0.9241 0.3957 0.9241 0.9613
No log 4.7222 340 0.8712 0.4335 0.8712 0.9334
No log 4.75 342 0.8479 0.4261 0.8479 0.9208
No log 4.7778 344 0.8514 0.3095 0.8514 0.9227
No log 4.8056 346 0.8442 0.3095 0.8442 0.9188
No log 4.8333 348 0.8408 0.3977 0.8408 0.9170
No log 4.8611 350 0.9083 0.3957 0.9083 0.9530
No log 4.8889 352 1.0533 0.3808 1.0533 1.0263
No log 4.9167 354 1.1153 0.3985 1.1153 1.0561
No log 4.9444 356 1.0563 0.3972 1.0563 1.0278
No log 4.9722 358 1.0135 0.3956 1.0135 1.0067
No log 5.0 360 1.0528 0.3956 1.0528 1.0261
No log 5.0278 362 1.0954 0.3921 1.0954 1.0466
No log 5.0556 364 1.0699 0.3938 1.0699 1.0344
No log 5.0833 366 1.0841 0.3921 1.0841 1.0412
No log 5.1111 368 1.0546 0.3921 1.0546 1.0269
No log 5.1389 370 0.9925 0.3691 0.9925 0.9962
No log 5.1667 372 0.9601 0.3005 0.9601 0.9798
No log 5.1944 374 0.9667 0.3005 0.9667 0.9832
No log 5.2222 376 1.0174 0.3363 1.0174 1.0087
No log 5.25 378 1.0755 0.3511 1.0755 1.0371
No log 5.2778 380 1.1215 0.3808 1.1215 1.0590
No log 5.3056 382 1.1069 0.3787 1.1069 1.0521
No log 5.3333 384 1.0623 0.3787 1.0623 1.0307
No log 5.3611 386 1.0543 0.3424 1.0543 1.0268
No log 5.3889 388 1.0025 0.3361 1.0025 1.0013
No log 5.4167 390 0.9928 0.2655 0.9928 0.9964
No log 5.4444 392 1.0042 0.1927 1.0042 1.0021
No log 5.4722 394 1.0074 0.1927 1.0074 1.0037
No log 5.5 396 0.9658 0.1927 0.9658 0.9828
No log 5.5278 398 0.9081 0.2499 0.9081 0.9529
No log 5.5556 400 0.8621 0.3285 0.8621 0.9285
No log 5.5833 402 0.8447 0.3974 0.8447 0.9191
No log 5.6111 404 0.7834 0.4388 0.7834 0.8851
No log 5.6389 406 0.8045 0.4462 0.8045 0.8969
No log 5.6667 408 0.8008 0.4180 0.8008 0.8949
No log 5.6944 410 0.7744 0.4119 0.7744 0.8800
No log 5.7222 412 0.8778 0.3861 0.8778 0.9369
No log 5.75 414 1.0305 0.4458 1.0305 1.0152
No log 5.7778 416 1.0575 0.4458 1.0575 1.0284
No log 5.8056 418 0.9561 0.3913 0.9561 0.9778
No log 5.8333 420 0.9395 0.3897 0.9395 0.9693
No log 5.8611 422 0.9616 0.3654 0.9616 0.9806
No log 5.8889 424 0.9968 0.3921 0.9968 0.9984
No log 5.9167 426 0.9959 0.3648 0.9959 0.9980
No log 5.9444 428 0.9815 0.3278 0.9815 0.9907
No log 5.9722 430 0.9375 0.3577 0.9375 0.9683
No log 6.0 432 0.9339 0.3577 0.9339 0.9664
No log 6.0278 434 0.9521 0.3558 0.9521 0.9758
No log 6.0556 436 1.0083 0.3511 1.0083 1.0042
No log 6.0833 438 1.0670 0.3222 1.0670 1.0330
No log 6.1111 440 1.1570 0.3654 1.1570 1.0757
No log 6.1389 442 1.1796 0.3546 1.1796 1.0861
No log 6.1667 444 1.1154 0.3787 1.1154 1.0561
No log 6.1944 446 1.0117 0.3472 1.0117 1.0058
No log 6.2222 448 0.9543 0.2228 0.9543 0.9769
No log 6.25 450 0.9475 0.2651 0.9475 0.9734
No log 6.2778 452 0.9798 0.2346 0.9798 0.9898
No log 6.3056 454 1.0599 0.2325 1.0599 1.0295
No log 6.3333 456 1.1369 0.3374 1.1369 1.0663
No log 6.3611 458 1.1778 0.3135 1.1778 1.0852
No log 6.3889 460 1.1621 0.3761 1.1621 1.0780
No log 6.4167 462 1.0528 0.3677 1.0528 1.0261
No log 6.4444 464 0.9640 0.3824 0.9640 0.9819
No log 6.4722 466 0.9313 0.2951 0.9313 0.9650
No log 6.5 468 0.9523 0.3771 0.9523 0.9759
No log 6.5278 470 0.9585 0.3790 0.9585 0.9790
No log 6.5556 472 0.9810 0.4253 0.9810 0.9904
No log 6.5833 474 0.9529 0.4234 0.9529 0.9762
No log 6.6111 476 0.9137 0.4471 0.9137 0.9559
No log 6.6389 478 0.8776 0.3285 0.8776 0.9368
No log 6.6667 480 0.8399 0.3859 0.8399 0.9165
No log 6.6944 482 0.8264 0.3878 0.8264 0.9091
No log 6.7222 484 0.8143 0.4012 0.8143 0.9024
No log 6.75 486 0.8081 0.4128 0.8081 0.8990
No log 6.7778 488 0.8017 0.4014 0.8017 0.8954
No log 6.8056 490 0.8120 0.3740 0.8120 0.9011
No log 6.8333 492 0.8345 0.4361 0.8345 0.9135
No log 6.8611 494 0.8847 0.4102 0.8847 0.9406
No log 6.8889 496 0.8957 0.3957 0.8957 0.9464
No log 6.9167 498 0.8589 0.3697 0.8589 0.9268
0.266 6.9444 500 0.8700 0.3740 0.8700 0.9328
0.266 6.9722 502 0.8957 0.3740 0.8957 0.9464
0.266 7.0 504 0.9315 0.3762 0.9315 0.9651
0.266 7.0278 506 0.9757 0.3662 0.9757 0.9878
0.266 7.0556 508 1.0009 0.3409 1.0009 1.0005
0.266 7.0833 510 1.0352 0.2770 1.0352 1.0174
0.266 7.1111 512 1.0415 0.2322 1.0415 1.0205
0.266 7.1389 514 1.0338 0.2263 1.0338 1.0168

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k15_task5_organization

Finetuned
(4019)
this model