ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k12_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8567
  • Qwk: 0.4956
  • Mse: 0.8567
  • Rmse: 0.9256

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0323 2 4.1923 -0.0344 4.1923 2.0475
No log 0.0645 4 2.4963 0.0358 2.4963 1.5800
No log 0.0968 6 1.1958 0.0864 1.1958 1.0935
No log 0.1290 8 1.0574 -0.0055 1.0574 1.0283
No log 0.1613 10 1.0113 -0.0791 1.0113 1.0056
No log 0.1935 12 0.8226 0.0502 0.8226 0.9070
No log 0.2258 14 0.8246 0.0831 0.8246 0.9081
No log 0.2581 16 0.9048 -0.1354 0.9048 0.9512
No log 0.2903 18 0.7697 0.1732 0.7697 0.8773
No log 0.3226 20 0.7270 0.1404 0.7270 0.8526
No log 0.3548 22 0.7658 0.0772 0.7658 0.8751
No log 0.3871 24 0.7754 0.1063 0.7754 0.8806
No log 0.4194 26 0.7240 0.1225 0.7240 0.8509
No log 0.4516 28 0.7029 0.1794 0.7029 0.8384
No log 0.4839 30 0.7249 0.1511 0.7249 0.8514
No log 0.5161 32 0.8473 0.1739 0.8473 0.9205
No log 0.5484 34 0.8413 0.1604 0.8413 0.9172
No log 0.5806 36 0.7495 0.1477 0.7495 0.8657
No log 0.6129 38 0.6934 0.3144 0.6934 0.8327
No log 0.6452 40 0.7369 0.3078 0.7369 0.8584
No log 0.6774 42 0.8758 0.3076 0.8758 0.9359
No log 0.7097 44 0.8694 0.3444 0.8694 0.9324
No log 0.7419 46 0.7460 0.3015 0.7460 0.8637
No log 0.7742 48 0.8201 0.3452 0.8201 0.9056
No log 0.8065 50 0.7437 0.3841 0.7437 0.8624
No log 0.8387 52 0.7993 0.3661 0.7993 0.8940
No log 0.8710 54 0.8922 0.3579 0.8922 0.9446
No log 0.9032 56 0.9822 0.2762 0.9822 0.9911
No log 0.9355 58 1.1207 0.2497 1.1207 1.0586
No log 0.9677 60 0.9058 0.2817 0.9058 0.9517
No log 1.0 62 0.9086 0.2632 0.9086 0.9532
No log 1.0323 64 1.2749 0.2434 1.2749 1.1291
No log 1.0645 66 1.4545 0.1654 1.4545 1.2060
No log 1.0968 68 1.1676 0.2841 1.1676 1.0805
No log 1.1290 70 0.7917 0.3655 0.7917 0.8898
No log 1.1613 72 0.7207 0.4300 0.7207 0.8489
No log 1.1935 74 0.7565 0.3952 0.7565 0.8698
No log 1.2258 76 0.9247 0.3115 0.9247 0.9616
No log 1.2581 78 0.8662 0.3235 0.8662 0.9307
No log 1.2903 80 0.8227 0.4122 0.8227 0.9070
No log 1.3226 82 0.9862 0.3751 0.9862 0.9931
No log 1.3548 84 1.3858 0.3781 1.3858 1.1772
No log 1.3871 86 1.8830 0.2658 1.8830 1.3722
No log 1.4194 88 1.6535 0.3253 1.6535 1.2859
No log 1.4516 90 1.2444 0.4088 1.2444 1.1155
No log 1.4839 92 1.1324 0.4089 1.1324 1.0642
No log 1.5161 94 1.1238 0.3861 1.1238 1.0601
No log 1.5484 96 1.0806 0.3976 1.0806 1.0395
No log 1.5806 98 1.0329 0.3900 1.0329 1.0163
No log 1.6129 100 1.2878 0.3835 1.2878 1.1348
No log 1.6452 102 0.9314 0.3830 0.9314 0.9651
No log 1.6774 104 0.7459 0.5117 0.7459 0.8636
No log 1.7097 106 0.7680 0.4720 0.7680 0.8764
No log 1.7419 108 0.8143 0.5023 0.8143 0.9024
No log 1.7742 110 1.0197 0.3664 1.0197 1.0098
No log 1.8065 112 1.6490 0.3525 1.6490 1.2841
No log 1.8387 114 1.5727 0.3466 1.5727 1.2541
No log 1.8710 116 1.3267 0.3426 1.3267 1.1518
No log 1.9032 118 1.2035 0.3584 1.2035 1.0971
No log 1.9355 120 1.2866 0.3544 1.2866 1.1343
No log 1.9677 122 1.5671 0.3586 1.5671 1.2518
No log 2.0 124 1.8005 0.2733 1.8005 1.3418
No log 2.0323 126 1.4195 0.3855 1.4195 1.1914
No log 2.0645 128 1.0335 0.3646 1.0335 1.0166
No log 2.0968 130 0.9462 0.4579 0.9462 0.9727
No log 2.1290 132 0.9507 0.4191 0.9507 0.9751
No log 2.1613 134 1.2730 0.3829 1.2730 1.1283
No log 2.1935 136 1.4554 0.3847 1.4554 1.2064
No log 2.2258 138 1.3239 0.3929 1.3239 1.1506
No log 2.2581 140 1.1957 0.3663 1.1957 1.0935
No log 2.2903 142 1.0769 0.3790 1.0769 1.0377
No log 2.3226 144 1.1943 0.3643 1.1943 1.0929
No log 2.3548 146 1.5198 0.3606 1.5198 1.2328
No log 2.3871 148 1.5899 0.3601 1.5899 1.2609
No log 2.4194 150 1.4131 0.3653 1.4131 1.1887
No log 2.4516 152 1.2314 0.3589 1.2314 1.1097
No log 2.4839 154 1.1568 0.3669 1.1568 1.0756
No log 2.5161 156 1.0299 0.3798 1.0299 1.0149
No log 2.5484 158 1.0340 0.3849 1.0340 1.0169
No log 2.5806 160 0.9988 0.4040 0.9988 0.9994
No log 2.6129 162 0.8326 0.4228 0.8326 0.9125
No log 2.6452 164 0.7624 0.4151 0.7624 0.8731
No log 2.6774 166 0.7759 0.3768 0.7759 0.8809
No log 2.7097 168 0.7968 0.4161 0.7968 0.8926
No log 2.7419 170 0.8338 0.4311 0.8338 0.9131
No log 2.7742 172 0.9266 0.4155 0.9266 0.9626
No log 2.8065 174 1.0121 0.3965 1.0121 1.0060
No log 2.8387 176 0.8940 0.4308 0.8940 0.9455
No log 2.8710 178 0.8481 0.4590 0.8481 0.9209
No log 2.9032 180 0.8008 0.5070 0.8008 0.8949
No log 2.9355 182 0.8303 0.4640 0.8303 0.9112
No log 2.9677 184 0.9083 0.4014 0.9083 0.9531
No log 3.0 186 1.1222 0.4187 1.1222 1.0594
No log 3.0323 188 1.1936 0.4221 1.1936 1.0925
No log 3.0645 190 1.0153 0.3798 1.0153 1.0076
No log 3.0968 192 0.8335 0.4670 0.8335 0.9130
No log 3.1290 194 0.8458 0.4806 0.8458 0.9197
No log 3.1613 196 0.8821 0.4773 0.8821 0.9392
No log 3.1935 198 1.0093 0.4021 1.0093 1.0047
No log 3.2258 200 1.1502 0.4083 1.1502 1.0725
No log 3.2581 202 1.0661 0.3947 1.0661 1.0325
No log 3.2903 204 0.9320 0.4148 0.9320 0.9654
No log 3.3226 206 0.8470 0.4321 0.8470 0.9203
No log 3.3548 208 0.8148 0.4457 0.8148 0.9027
No log 3.3871 210 0.6975 0.5077 0.6975 0.8352
No log 3.4194 212 0.7491 0.4688 0.7491 0.8655
No log 3.4516 214 0.7352 0.5060 0.7352 0.8575
No log 3.4839 216 0.7087 0.5127 0.7087 0.8418
No log 3.5161 218 0.8947 0.4872 0.8947 0.9459
No log 3.5484 220 1.0028 0.4561 1.0028 1.0014
No log 3.5806 222 0.9169 0.4830 0.9169 0.9575
No log 3.6129 224 0.8874 0.5036 0.8874 0.9420
No log 3.6452 226 0.9728 0.4700 0.9728 0.9863
No log 3.6774 228 1.2958 0.3977 1.2958 1.1383
No log 3.7097 230 1.7146 0.3302 1.7146 1.3094
No log 3.7419 232 1.6631 0.3256 1.6631 1.2896
No log 3.7742 234 1.3003 0.4124 1.3003 1.1403
No log 3.8065 236 0.9369 0.4758 0.9369 0.9680
No log 3.8387 238 0.7978 0.4744 0.7978 0.8932
No log 3.8710 240 0.7486 0.4791 0.7486 0.8652
No log 3.9032 242 0.8136 0.5017 0.8136 0.9020
No log 3.9355 244 0.8577 0.4738 0.8577 0.9261
No log 3.9677 246 0.9739 0.4666 0.9739 0.9868
No log 4.0 248 0.9478 0.4703 0.9478 0.9736
No log 4.0323 250 0.9407 0.4827 0.9407 0.9699
No log 4.0645 252 0.8555 0.4511 0.8555 0.9249
No log 4.0968 254 0.8809 0.4571 0.8809 0.9386
No log 4.1290 256 0.8625 0.4603 0.8625 0.9287
No log 4.1613 258 1.0147 0.4575 1.0147 1.0073
No log 4.1935 260 1.0431 0.4617 1.0431 1.0213
No log 4.2258 262 0.9715 0.4453 0.9715 0.9856
No log 4.2581 264 0.8652 0.4885 0.8652 0.9301
No log 4.2903 266 0.7624 0.5023 0.7624 0.8732
No log 4.3226 268 0.7389 0.4722 0.7389 0.8596
No log 4.3548 270 0.7540 0.4776 0.7540 0.8683
No log 4.3871 272 0.8437 0.5253 0.8437 0.9185
No log 4.4194 274 1.1489 0.4398 1.1489 1.0719
No log 4.4516 276 1.3852 0.3857 1.3852 1.1770
No log 4.4839 278 1.2549 0.4050 1.2549 1.1202
No log 4.5161 280 0.9695 0.4544 0.9695 0.9846
No log 4.5484 282 0.8432 0.4958 0.8432 0.9183
No log 4.5806 284 0.8186 0.4941 0.8186 0.9047
No log 4.6129 286 0.8201 0.5112 0.8201 0.9056
No log 4.6452 288 0.8445 0.5036 0.8445 0.9190
No log 4.6774 290 0.9258 0.4200 0.9258 0.9622
No log 4.7097 292 1.0225 0.4054 1.0225 1.0112
No log 4.7419 294 1.0015 0.4054 1.0015 1.0007
No log 4.7742 296 0.8921 0.4250 0.8921 0.9445
No log 4.8065 298 0.8272 0.4986 0.8272 0.9095
No log 4.8387 300 0.7946 0.5094 0.7946 0.8914
No log 4.8710 302 0.7964 0.5042 0.7964 0.8924
No log 4.9032 304 0.8581 0.4544 0.8581 0.9264
No log 4.9355 306 0.8894 0.4653 0.8894 0.9431
No log 4.9677 308 0.8250 0.5147 0.8250 0.9083
No log 5.0 310 0.7523 0.5123 0.7523 0.8674
No log 5.0323 312 0.7320 0.5520 0.7320 0.8555
No log 5.0645 314 0.7306 0.5312 0.7306 0.8548
No log 5.0968 316 0.7737 0.5225 0.7737 0.8796
No log 5.1290 318 0.9141 0.4202 0.9141 0.9561
No log 5.1613 320 0.9835 0.4184 0.9835 0.9917
No log 5.1935 322 1.0071 0.4342 1.0071 1.0036
No log 5.2258 324 0.9501 0.4295 0.9501 0.9747
No log 5.2581 326 0.8006 0.5292 0.8006 0.8947
No log 5.2903 328 0.7145 0.5358 0.7145 0.8453
No log 5.3226 330 0.7207 0.5519 0.7207 0.8490
No log 5.3548 332 0.8040 0.5530 0.8040 0.8967
No log 5.3871 334 0.9898 0.4430 0.9898 0.9949
No log 5.4194 336 1.1880 0.4195 1.1880 1.0900
No log 5.4516 338 1.1946 0.4224 1.1946 1.0930
No log 5.4839 340 1.0595 0.4458 1.0595 1.0293
No log 5.5161 342 0.9976 0.4589 0.9976 0.9988
No log 5.5484 344 0.9890 0.4467 0.9890 0.9945
No log 5.5806 346 0.9219 0.4405 0.9219 0.9601
No log 5.6129 348 0.9159 0.4563 0.9159 0.9570
No log 5.6452 350 0.9513 0.4516 0.9513 0.9753
No log 5.6774 352 0.9567 0.4389 0.9567 0.9781
No log 5.7097 354 0.8482 0.4720 0.8482 0.9210
No log 5.7419 356 0.7637 0.5054 0.7637 0.8739
No log 5.7742 358 0.7488 0.5288 0.7488 0.8653
No log 5.8065 360 0.7929 0.5091 0.7929 0.8905
No log 5.8387 362 0.9142 0.4457 0.9142 0.9561
No log 5.8710 364 1.0309 0.4555 1.0309 1.0153
No log 5.9032 366 1.0418 0.4555 1.0418 1.0207
No log 5.9355 368 0.9470 0.4360 0.9470 0.9731
No log 5.9677 370 0.8817 0.5024 0.8817 0.9390
No log 6.0 372 0.8437 0.4813 0.8437 0.9185
No log 6.0323 374 0.8468 0.4781 0.8468 0.9202
No log 6.0645 376 0.8376 0.4781 0.8376 0.9152
No log 6.0968 378 0.8454 0.4941 0.8454 0.9194
No log 6.1290 380 0.8151 0.5084 0.8151 0.9028
No log 6.1613 382 0.7688 0.5014 0.7688 0.8768
No log 6.1935 384 0.7394 0.4982 0.7394 0.8599
No log 6.2258 386 0.7167 0.5 0.7167 0.8466
No log 6.2581 388 0.7021 0.5 0.7021 0.8379
No log 6.2903 390 0.7247 0.5043 0.7247 0.8513
No log 6.3226 392 0.7662 0.5124 0.7662 0.8753
No log 6.3548 394 0.7975 0.5054 0.7975 0.8930
No log 6.3871 396 0.8514 0.4911 0.8514 0.9227
No log 6.4194 398 0.9669 0.4705 0.9669 0.9833
No log 6.4516 400 0.9802 0.4716 0.9802 0.9900
No log 6.4839 402 0.8887 0.4754 0.8887 0.9427
No log 6.5161 404 0.8083 0.5120 0.8083 0.8990
No log 6.5484 406 0.7767 0.4973 0.7767 0.8813
No log 6.5806 408 0.7540 0.5278 0.7540 0.8684
No log 6.6129 410 0.7515 0.4907 0.7515 0.8669
No log 6.6452 412 0.7985 0.5290 0.7985 0.8936
No log 6.6774 414 0.8353 0.4966 0.8353 0.9139
No log 6.7097 416 0.8265 0.4901 0.8265 0.9091
No log 6.7419 418 0.8466 0.4824 0.8466 0.9201
No log 6.7742 420 0.8328 0.4823 0.8328 0.9126
No log 6.8065 422 0.7811 0.5450 0.7811 0.8838
No log 6.8387 424 0.7428 0.5590 0.7428 0.8619
No log 6.8710 426 0.7683 0.5633 0.7683 0.8766
No log 6.9032 428 0.8645 0.4708 0.8645 0.9298
No log 6.9355 430 0.9746 0.4398 0.9746 0.9872
No log 6.9677 432 0.9873 0.4384 0.9873 0.9937
No log 7.0 434 0.9030 0.4653 0.9030 0.9502
No log 7.0323 436 0.8254 0.5558 0.8254 0.9085
No log 7.0645 438 0.8184 0.5407 0.8184 0.9047
No log 7.0968 440 0.8512 0.5104 0.8512 0.9226
No log 7.1290 442 0.9455 0.4711 0.9455 0.9723
No log 7.1613 444 1.0954 0.4153 1.0954 1.0466
No log 7.1935 446 1.2953 0.4053 1.2953 1.1381
No log 7.2258 448 1.3534 0.4058 1.3534 1.1634
No log 7.2581 450 1.2851 0.4089 1.2851 1.1336
No log 7.2903 452 1.1359 0.4209 1.1359 1.0658
No log 7.3226 454 1.0052 0.4468 1.0052 1.0026
No log 7.3548 456 0.9088 0.4739 0.9088 0.9533
No log 7.3871 458 0.8951 0.4836 0.8951 0.9461
No log 7.4194 460 0.9376 0.4356 0.9376 0.9683
No log 7.4516 462 1.0233 0.4321 1.0233 1.0116
No log 7.4839 464 1.0834 0.4120 1.0834 1.0409
No log 7.5161 466 1.0887 0.4120 1.0887 1.0434
No log 7.5484 468 1.0399 0.4161 1.0399 1.0198
No log 7.5806 470 0.9595 0.4218 0.9595 0.9795
No log 7.6129 472 0.8821 0.4823 0.8821 0.9392
No log 7.6452 474 0.8249 0.5461 0.8249 0.9082
No log 7.6774 476 0.8148 0.5146 0.8148 0.9027
No log 7.7097 478 0.8343 0.5146 0.8343 0.9134
No log 7.7419 480 0.8912 0.4912 0.8912 0.9440
No log 7.7742 482 0.9791 0.4590 0.9791 0.9895
No log 7.8065 484 1.0369 0.4678 1.0369 1.0183
No log 7.8387 486 1.0767 0.4459 1.0767 1.0377
No log 7.8710 488 1.0344 0.4678 1.0344 1.0170
No log 7.9032 490 0.9460 0.4652 0.9460 0.9726
No log 7.9355 492 0.8366 0.5206 0.8366 0.9147
No log 7.9677 494 0.7699 0.5140 0.7699 0.8774
No log 8.0 496 0.7443 0.4853 0.7443 0.8627
No log 8.0323 498 0.7507 0.5014 0.7507 0.8665
0.4636 8.0645 500 0.7755 0.5431 0.7755 0.8806
0.4636 8.0968 502 0.8285 0.5360 0.8285 0.9102
0.4636 8.1290 504 0.9202 0.4554 0.9202 0.9593
0.4636 8.1613 506 0.9759 0.4309 0.9759 0.9879
0.4636 8.1935 508 1.0227 0.4203 1.0227 1.0113
0.4636 8.2258 510 1.0091 0.4203 1.0091 1.0046
0.4636 8.2581 512 0.9474 0.4567 0.9474 0.9734
0.4636 8.2903 514 0.8672 0.4948 0.8672 0.9312
0.4636 8.3226 516 0.8326 0.5710 0.8326 0.9125
0.4636 8.3548 518 0.8007 0.5643 0.8007 0.8948
0.4636 8.3871 520 0.7789 0.5490 0.7789 0.8826
0.4636 8.4194 522 0.7652 0.5377 0.7652 0.8747
0.4636 8.4516 524 0.7661 0.5377 0.7661 0.8753
0.4636 8.4839 526 0.7836 0.5545 0.7836 0.8852
0.4636 8.5161 528 0.8138 0.5464 0.8138 0.9021
0.4636 8.5484 530 0.8360 0.5426 0.8360 0.9143
0.4636 8.5806 532 0.8453 0.5172 0.8453 0.9194
0.4636 8.6129 534 0.8457 0.5414 0.8457 0.9196
0.4636 8.6452 536 0.8616 0.4929 0.8616 0.9282
0.4636 8.6774 538 0.8640 0.4916 0.8640 0.9295
0.4636 8.7097 540 0.8601 0.4922 0.8601 0.9274
0.4636 8.7419 542 0.8749 0.4775 0.8749 0.9354
0.4636 8.7742 544 0.8925 0.4766 0.8925 0.9447
0.4636 8.8065 546 0.8935 0.4766 0.8935 0.9453
0.4636 8.8387 548 0.8866 0.4883 0.8866 0.9416
0.4636 8.8710 550 0.8717 0.4897 0.8717 0.9336
0.4636 8.9032 552 0.8511 0.5414 0.8511 0.9225
0.4636 8.9355 554 0.8409 0.5398 0.8409 0.9170
0.4636 8.9677 556 0.8514 0.5154 0.8514 0.9227
0.4636 9.0 558 0.8788 0.4897 0.8788 0.9374
0.4636 9.0323 560 0.9147 0.4756 0.9147 0.9564
0.4636 9.0645 562 0.9353 0.4534 0.9353 0.9671
0.4636 9.0968 564 0.9357 0.4534 0.9357 0.9673
0.4636 9.1290 566 0.9413 0.4534 0.9413 0.9702
0.4636 9.1613 568 0.9450 0.4534 0.9450 0.9721
0.4636 9.1935 570 0.9349 0.4534 0.9349 0.9669
0.4636 9.2258 572 0.9142 0.4641 0.9142 0.9562
0.4636 9.2581 574 0.8914 0.4766 0.8914 0.9441
0.4636 9.2903 576 0.8840 0.4883 0.8840 0.9402
0.4636 9.3226 578 0.8776 0.4883 0.8776 0.9368
0.4636 9.3548 580 0.8635 0.5115 0.8635 0.9292
0.4636 9.3871 582 0.8551 0.5303 0.8551 0.9247
0.4636 9.4194 584 0.8382 0.5269 0.8382 0.9155
0.4636 9.4516 586 0.8230 0.5476 0.8230 0.9072
0.4636 9.4839 588 0.8148 0.5627 0.8148 0.9027
0.4636 9.5161 590 0.8129 0.5627 0.8129 0.9016
0.4636 9.5484 592 0.8140 0.5544 0.8140 0.9022
0.4636 9.5806 594 0.8188 0.5656 0.8188 0.9049
0.4636 9.6129 596 0.8304 0.5420 0.8304 0.9112
0.4636 9.6452 598 0.8442 0.5269 0.8442 0.9188
0.4636 9.6774 600 0.8520 0.5313 0.8520 0.9230
0.4636 9.7097 602 0.8546 0.5076 0.8546 0.9244
0.4636 9.7419 604 0.8613 0.5068 0.8613 0.9280
0.4636 9.7742 606 0.8656 0.4998 0.8656 0.9304
0.4636 9.8065 608 0.8645 0.4998 0.8645 0.9298
0.4636 9.8387 610 0.8640 0.4998 0.8640 0.9295
0.4636 9.8710 612 0.8627 0.4949 0.8627 0.9288
0.4636 9.9032 614 0.8602 0.4949 0.8602 0.9275
0.4636 9.9355 616 0.8585 0.4956 0.8585 0.9266
0.4636 9.9677 618 0.8572 0.4956 0.8572 0.9258
0.4636 10.0 620 0.8567 0.4956 0.8567 0.9256

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
3
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k12_task2_organization

Finetuned
(4023)
this model