ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k10_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7252
  • Qwk: 0.2464
  • Mse: 0.7252
  • Rmse: 0.8516

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0444 2 3.2300 -0.0149 3.2300 1.7972
No log 0.0889 4 1.6866 -0.0070 1.6866 1.2987
No log 0.1333 6 1.4718 0.0255 1.4718 1.2132
No log 0.1778 8 0.8949 0.1673 0.8949 0.9460
No log 0.2222 10 0.5655 0.0222 0.5655 0.7520
No log 0.2667 12 0.5613 0.0569 0.5613 0.7492
No log 0.3111 14 0.5551 0.0303 0.5551 0.7451
No log 0.3556 16 0.5454 0.0569 0.5454 0.7385
No log 0.4 18 0.5520 -0.0159 0.5520 0.7430
No log 0.4444 20 0.8311 0.2000 0.8311 0.9116
No log 0.4889 22 0.7803 0.2300 0.7803 0.8834
No log 0.5333 24 0.6153 0.0720 0.6153 0.7844
No log 0.5778 26 0.7098 0.2000 0.7098 0.8425
No log 0.6222 28 0.6655 0.1888 0.6655 0.8158
No log 0.6667 30 0.6458 -0.1007 0.6458 0.8036
No log 0.7111 32 0.7409 0.0345 0.7409 0.8607
No log 0.7556 34 0.6909 -0.0390 0.6909 0.8312
No log 0.8 36 0.6797 -0.0072 0.6797 0.8244
No log 0.8444 38 0.6871 0.0199 0.6871 0.8289
No log 0.8889 40 0.7641 0.0311 0.7641 0.8741
No log 0.9333 42 0.9004 -0.0275 0.9004 0.9489
No log 0.9778 44 0.7186 0.1628 0.7186 0.8477
No log 1.0222 46 0.7551 0.0737 0.7551 0.8690
No log 1.0667 48 0.6982 0.1373 0.6982 0.8356
No log 1.1111 50 0.7325 0.0728 0.7325 0.8559
No log 1.1556 52 1.0763 -0.0154 1.0763 1.0374
No log 1.2 54 0.7183 0.0769 0.7183 0.8476
No log 1.2444 56 0.6764 0.1345 0.6764 0.8224
No log 1.2889 58 0.7061 0.0617 0.7061 0.8403
No log 1.3333 60 1.0978 -0.0277 1.0978 1.0478
No log 1.3778 62 1.0735 0.0252 1.0735 1.0361
No log 1.4222 64 0.8700 0.0891 0.8700 0.9328
No log 1.4667 66 0.8345 0.0891 0.8345 0.9135
No log 1.5111 68 0.7870 0.0802 0.7870 0.8871
No log 1.5556 70 1.0944 0.0 1.0944 1.0461
No log 1.6 72 1.0873 0.0038 1.0873 1.0427
No log 1.6444 74 1.0292 0.0169 1.0292 1.0145
No log 1.6889 76 1.2861 0.0270 1.2861 1.1341
No log 1.7333 78 0.9785 0.0717 0.9785 0.9892
No log 1.7778 80 0.7560 0.2577 0.7560 0.8695
No log 1.8222 82 0.8102 0.2153 0.8102 0.9001
No log 1.8667 84 1.4464 0.0464 1.4464 1.2027
No log 1.9111 86 1.4896 0.0502 1.4896 1.2205
No log 1.9556 88 1.2051 0.0262 1.2051 1.0978
No log 2.0 90 0.9645 0.0308 0.9645 0.9821
No log 2.0444 92 0.7751 0.2727 0.7751 0.8804
No log 2.0889 94 0.8084 0.2444 0.8084 0.8991
No log 2.1333 96 0.9620 0.0539 0.9620 0.9808
No log 2.1778 98 1.6203 0.1111 1.6203 1.2729
No log 2.2222 100 1.3509 0.1169 1.3509 1.1623
No log 2.2667 102 0.8183 0.1538 0.8183 0.9046
No log 2.3111 104 0.8755 0.2146 0.8755 0.9357
No log 2.3556 106 0.9242 0.2489 0.9242 0.9614
No log 2.4 108 0.9547 0.2208 0.9547 0.9771
No log 2.4444 110 1.3326 0.0789 1.3326 1.1544
No log 2.4889 112 0.8853 0.1269 0.8853 0.9409
No log 2.5333 114 0.7701 0.1919 0.7701 0.8776
No log 2.5778 116 0.8002 0.2315 0.8002 0.8945
No log 2.6222 118 0.7511 0.2350 0.7511 0.8666
No log 2.6667 120 0.8243 0.1600 0.8243 0.9079
No log 2.7111 122 1.1654 0.0312 1.1654 1.0796
No log 2.7556 124 1.1737 0.0111 1.1737 1.0834
No log 2.8 126 0.7277 0.2513 0.7277 0.8531
No log 2.8444 128 0.7175 0.2626 0.7175 0.8470
No log 2.8889 130 0.7618 0.2000 0.7618 0.8728
No log 2.9333 132 0.8250 0.1269 0.8250 0.9083
No log 2.9778 134 1.0376 -0.0164 1.0376 1.0186
No log 3.0222 136 1.0807 -0.0164 1.0807 1.0396
No log 3.0667 138 0.7983 0.1086 0.7983 0.8935
No log 3.1111 140 0.8270 0.0512 0.8270 0.9094
No log 3.1556 142 0.7617 0.2889 0.7617 0.8727
No log 3.2 144 1.1630 0.0840 1.1630 1.0784
No log 3.2444 146 1.8142 0.0787 1.8142 1.3469
No log 3.2889 148 1.3955 0.1325 1.3955 1.1813
No log 3.3333 150 0.6669 0.1910 0.6669 0.8167
No log 3.3778 152 0.6600 0.2393 0.6600 0.8124
No log 3.4222 154 0.6507 0.2749 0.6507 0.8067
No log 3.4667 156 0.6683 0.2258 0.6683 0.8175
No log 3.5111 158 0.7364 0.1568 0.7364 0.8581
No log 3.5556 160 0.7322 0.2653 0.7322 0.8557
No log 3.6 162 0.7172 0.2990 0.7172 0.8469
No log 3.6444 164 0.7420 0.1628 0.7420 0.8614
No log 3.6889 166 0.7474 0.2965 0.7474 0.8645
No log 3.7333 168 0.9132 0.1781 0.9132 0.9556
No log 3.7778 170 0.9228 0.1786 0.9228 0.9606
No log 3.8222 172 0.7122 0.2245 0.7122 0.8439
No log 3.8667 174 0.8060 0.2676 0.8060 0.8978
No log 3.9111 176 0.8894 0.1336 0.8894 0.9431
No log 3.9556 178 0.6875 0.2787 0.6875 0.8292
No log 4.0 180 0.8985 0.1453 0.8985 0.9479
No log 4.0444 182 1.2645 0.1409 1.2645 1.1245
No log 4.0889 184 1.0187 0.1008 1.0187 1.0093
No log 4.1333 186 0.6656 0.2298 0.6656 0.8159
No log 4.1778 188 0.7599 0.1357 0.7599 0.8717
No log 4.2222 190 0.7582 0.1841 0.7582 0.8707
No log 4.2667 192 0.6713 0.3136 0.6713 0.8193
No log 4.3111 194 0.8292 0.0714 0.8292 0.9106
No log 4.3556 196 1.0383 0.0949 1.0383 1.0190
No log 4.4 198 0.8862 0.1392 0.8862 0.9414
No log 4.4444 200 0.8202 0.1273 0.8202 0.9057
No log 4.4889 202 0.7186 0.2653 0.7186 0.8477
No log 4.5333 204 0.7369 0.2323 0.7369 0.8584
No log 4.5778 206 0.8505 0.1660 0.8505 0.9222
No log 4.6222 208 0.8253 0.1628 0.8253 0.9084
No log 4.6667 210 0.7015 0.2340 0.7015 0.8376
No log 4.7111 212 0.6842 0.3061 0.6842 0.8272
No log 4.7556 214 0.6146 0.3520 0.6146 0.7839
No log 4.8 216 0.6974 0.2165 0.6974 0.8351
No log 4.8444 218 0.6578 0.2487 0.6578 0.8111
No log 4.8889 220 0.6037 0.2865 0.6037 0.7770
No log 4.9333 222 0.5459 0.3797 0.5459 0.7388
No log 4.9778 224 0.5539 0.3913 0.5539 0.7443
No log 5.0222 226 0.5709 0.4105 0.5709 0.7556
No log 5.0667 228 0.6205 0.3641 0.6205 0.7877
No log 5.1111 230 0.6389 0.3641 0.6389 0.7993
No log 5.1556 232 0.6886 0.2963 0.6886 0.8298
No log 5.2 234 0.9103 0.2000 0.9103 0.9541
No log 5.2444 236 1.0206 0.1506 1.0206 1.0102
No log 5.2889 238 0.8372 0.2348 0.8372 0.9150
No log 5.3333 240 0.6978 0.3548 0.6978 0.8354
No log 5.3778 242 0.7519 0.2287 0.7519 0.8671
No log 5.4222 244 0.6975 0.3469 0.6975 0.8351
No log 5.4667 246 0.7209 0.1919 0.7209 0.8491
No log 5.5111 248 0.6954 0.2239 0.6954 0.8339
No log 5.5556 250 0.6836 0.3927 0.6836 0.8268
No log 5.6 252 0.8046 0.2072 0.8046 0.8970
No log 5.6444 254 0.8427 0.2281 0.8427 0.9180
No log 5.6889 256 0.7290 0.2963 0.7290 0.8538
No log 5.7333 258 0.8200 0.2000 0.8200 0.9055
No log 5.7778 260 0.9714 0.1169 0.9714 0.9856
No log 5.8222 262 0.8523 0.1705 0.8523 0.9232
No log 5.8667 264 0.7340 0.2233 0.7340 0.8567
No log 5.9111 266 0.6937 0.2842 0.6937 0.8329
No log 5.9556 268 0.6790 0.2917 0.6790 0.8240
No log 6.0 270 0.6959 0.1917 0.6959 0.8342
No log 6.0444 272 0.7125 0.1753 0.7125 0.8441
No log 6.0889 274 0.6606 0.2088 0.6606 0.8128
No log 6.1333 276 0.6208 0.2174 0.6208 0.7879
No log 6.1778 278 0.6444 0.2340 0.6444 0.8027
No log 6.2222 280 0.7016 0.1667 0.7016 0.8376
No log 6.2667 282 0.6538 0.2340 0.6538 0.8086
No log 6.3111 284 0.6297 0.3407 0.6297 0.7936
No log 6.3556 286 0.6545 0.2941 0.6545 0.8090
No log 6.4 288 0.7237 0.2157 0.7237 0.8507
No log 6.4444 290 0.7713 0.1781 0.7713 0.8782
No log 6.4889 292 0.8236 0.1549 0.8236 0.9075
No log 6.5333 294 0.7875 0.2432 0.7875 0.8874
No log 6.5778 296 0.7506 0.2475 0.7506 0.8664
No log 6.6222 298 0.7670 0.1481 0.7670 0.8758
No log 6.6667 300 0.8354 0.1238 0.8354 0.9140
No log 6.7111 302 0.7641 0.1238 0.7641 0.8741
No log 6.7556 304 0.6813 0.2258 0.6813 0.8254
No log 6.8 306 0.6658 0.3191 0.6658 0.8159
No log 6.8444 308 0.6707 0.2746 0.6707 0.8190
No log 6.8889 310 0.6911 0.2917 0.6911 0.8313
No log 6.9333 312 0.7084 0.2917 0.7084 0.8416
No log 6.9778 314 0.7124 0.2536 0.7124 0.8441
No log 7.0222 316 0.7266 0.2536 0.7266 0.8524
No log 7.0667 318 0.7239 0.2917 0.7239 0.8509
No log 7.1111 320 0.7234 0.2917 0.7234 0.8505
No log 7.1556 322 0.7055 0.2917 0.7055 0.8400
No log 7.2 324 0.6826 0.2464 0.6826 0.8262
No log 7.2444 326 0.6700 0.3299 0.6700 0.8185
No log 7.2889 328 0.6589 0.3043 0.6589 0.8117
No log 7.3333 330 0.6596 0.2265 0.6596 0.8121
No log 7.3778 332 0.7024 0.1556 0.7024 0.8381
No log 7.4222 334 0.7381 0.2000 0.7381 0.8591
No log 7.4667 336 0.7035 0.1913 0.7035 0.8388
No log 7.5111 338 0.6726 0.2727 0.6726 0.8201
No log 7.5556 340 0.7029 0.2549 0.7029 0.8384
No log 7.6 342 0.7122 0.2549 0.7122 0.8439
No log 7.6444 344 0.6876 0.3077 0.6876 0.8292
No log 7.6889 346 0.6972 0.28 0.6972 0.8350
No log 7.7333 348 0.7747 0.2300 0.7747 0.8802
No log 7.7778 350 0.7985 0.1927 0.7985 0.8936
No log 7.8222 352 0.7558 0.2593 0.7558 0.8694
No log 7.8667 354 0.7167 0.2563 0.7167 0.8466
No log 7.9111 356 0.7115 0.2563 0.7115 0.8435
No log 7.9556 358 0.7395 0.2986 0.7395 0.8599
No log 8.0 360 0.7676 0.1928 0.7676 0.8761
No log 8.0444 362 0.7557 0.2692 0.7557 0.8693
No log 8.0889 364 0.7382 0.2692 0.7382 0.8592
No log 8.1333 366 0.6938 0.2653 0.6938 0.8329
No log 8.1778 368 0.6745 0.2549 0.6745 0.8213
No log 8.2222 370 0.6961 0.3171 0.6961 0.8343
No log 8.2667 372 0.7020 0.2762 0.7020 0.8378
No log 8.3111 374 0.7081 0.2390 0.7081 0.8415
No log 8.3556 376 0.7511 0.2621 0.7511 0.8667
No log 8.4 378 0.7744 0.2212 0.7744 0.8800
No log 8.4444 380 0.7501 0.2621 0.7501 0.8661
No log 8.4889 382 0.7372 0.2637 0.7372 0.8586
No log 8.5333 384 0.7133 0.2464 0.7133 0.8446
No log 8.5778 386 0.7017 0.24 0.7017 0.8377
No log 8.6222 388 0.6969 0.2549 0.6969 0.8348
No log 8.6667 390 0.6986 0.2487 0.6986 0.8358
No log 8.7111 392 0.7210 0.2637 0.7210 0.8491
No log 8.7556 394 0.7451 0.2637 0.7451 0.8632
No log 8.8 396 0.7346 0.2637 0.7346 0.8571
No log 8.8444 398 0.7283 0.2323 0.7283 0.8534
No log 8.8889 400 0.7056 0.2653 0.7056 0.8400
No log 8.9333 402 0.6945 0.2475 0.6945 0.8334
No log 8.9778 404 0.6943 0.2475 0.6943 0.8333
No log 9.0222 406 0.6909 0.2821 0.6909 0.8312
No log 9.0667 408 0.6962 0.2727 0.6962 0.8344
No log 9.1111 410 0.7025 0.28 0.7025 0.8381
No log 9.1556 412 0.7148 0.2475 0.7148 0.8454
No log 9.2 414 0.7231 0.2453 0.7231 0.8503
No log 9.2444 416 0.7358 0.2453 0.7358 0.8578
No log 9.2889 418 0.7495 0.2593 0.7495 0.8657
No log 9.3333 420 0.7678 0.2593 0.7678 0.8762
No log 9.3778 422 0.7627 0.2593 0.7627 0.8734
No log 9.4222 424 0.7540 0.2593 0.7540 0.8683
No log 9.4667 426 0.7421 0.2637 0.7421 0.8615
No log 9.5111 428 0.7385 0.2637 0.7385 0.8593
No log 9.5556 430 0.7276 0.2453 0.7276 0.8530
No log 9.6 432 0.7207 0.2475 0.7207 0.8490
No log 9.6444 434 0.7210 0.2475 0.7210 0.8491
No log 9.6889 436 0.7215 0.2475 0.7215 0.8494
No log 9.7333 438 0.7216 0.2475 0.7216 0.8495
No log 9.7778 440 0.7206 0.2475 0.7206 0.8489
No log 9.8222 442 0.7215 0.2475 0.7215 0.8494
No log 9.8667 444 0.7227 0.2475 0.7227 0.8501
No log 9.9111 446 0.7240 0.2464 0.7240 0.8509
No log 9.9556 448 0.7250 0.2464 0.7250 0.8514
No log 10.0 450 0.7252 0.2464 0.7252 0.8516

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k10_task3_organization

Finetuned
(4023)
this model