ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k13_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8341
  • Qwk: 0.3590
  • Mse: 0.8341
  • Rmse: 0.9133

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0476 2 4.1274 0.0130 4.1274 2.0316
No log 0.0952 4 2.5328 0.0126 2.5328 1.5915
No log 0.1429 6 2.2970 0.0458 2.2970 1.5156
No log 0.1905 8 1.8055 0.1179 1.8055 1.3437
No log 0.2381 10 1.1105 0.2781 1.1105 1.0538
No log 0.2857 12 1.0330 0.3896 1.0330 1.0163
No log 0.3333 14 1.2105 0.0760 1.2105 1.1002
No log 0.3810 16 1.4923 0.0 1.4923 1.2216
No log 0.4286 18 1.5578 0.0 1.5578 1.2481
No log 0.4762 20 1.4850 0.0 1.4850 1.2186
No log 0.5238 22 1.3394 0.0 1.3394 1.1573
No log 0.5714 24 1.2428 0.0 1.2428 1.1148
No log 0.6190 26 1.1314 0.1471 1.1314 1.0637
No log 0.6667 28 1.1012 0.2711 1.1012 1.0494
No log 0.7143 30 1.1347 0.2119 1.1347 1.0652
No log 0.7619 32 1.2425 0.0611 1.2425 1.1147
No log 0.8095 34 1.1715 0.0883 1.1715 1.0824
No log 0.8571 36 1.1761 0.0 1.1761 1.0845
No log 0.9048 38 1.2350 0.0 1.2350 1.1113
No log 0.9524 40 1.1355 0.1114 1.1355 1.0656
No log 1.0 42 1.0671 0.2734 1.0671 1.0330
No log 1.0476 44 1.0352 0.3332 1.0352 1.0174
No log 1.0952 46 1.1802 0.3307 1.1802 1.0864
No log 1.1429 48 1.6303 0.0997 1.6303 1.2768
No log 1.1905 50 1.4463 0.1956 1.4463 1.2026
No log 1.2381 52 1.0354 0.2639 1.0354 1.0176
No log 1.2857 54 0.9785 0.2764 0.9785 0.9892
No log 1.3333 56 0.9658 0.2615 0.9658 0.9827
No log 1.3810 58 0.9853 0.2956 0.9853 0.9926
No log 1.4286 60 1.0190 0.3143 1.0190 1.0095
No log 1.4762 62 1.0529 0.2196 1.0529 1.0261
No log 1.5238 64 1.1359 0.1086 1.1359 1.0658
No log 1.5714 66 1.2213 0.0731 1.2213 1.1051
No log 1.6190 68 1.1537 0.0849 1.1537 1.0741
No log 1.6667 70 1.1574 0.0849 1.1574 1.0758
No log 1.7143 72 1.0914 0.1471 1.0914 1.0447
No log 1.7619 74 1.0104 0.2343 1.0104 1.0052
No log 1.8095 76 0.8979 0.3457 0.8979 0.9476
No log 1.8571 78 0.8485 0.4002 0.8485 0.9211
No log 1.9048 80 0.8465 0.3288 0.8465 0.9201
No log 1.9524 82 0.8602 0.2643 0.8602 0.9275
No log 2.0 84 0.8834 0.2770 0.8834 0.9399
No log 2.0476 86 0.8722 0.2770 0.8722 0.9339
No log 2.0952 88 0.8856 0.2770 0.8856 0.9411
No log 2.1429 90 0.8449 0.2944 0.8449 0.9192
No log 2.1905 92 0.8357 0.2788 0.8357 0.9142
No log 2.2381 94 0.8640 0.3445 0.8640 0.9295
No log 2.2857 96 0.8835 0.3398 0.8835 0.9399
No log 2.3333 98 0.9982 0.3625 0.9982 0.9991
No log 2.3810 100 1.0828 0.3574 1.0828 1.0406
No log 2.4286 102 0.9594 0.3988 0.9594 0.9795
No log 2.4762 104 0.9214 0.3820 0.9214 0.9599
No log 2.5238 106 0.8717 0.4254 0.8717 0.9337
No log 2.5714 108 0.9336 0.4470 0.9336 0.9662
No log 2.6190 110 1.1721 0.3493 1.1721 1.0827
No log 2.6667 112 1.2195 0.2881 1.2195 1.1043
No log 2.7143 114 1.0523 0.2791 1.0523 1.0258
No log 2.7619 116 0.8778 0.4828 0.8778 0.9369
No log 2.8095 118 0.8010 0.4903 0.8010 0.8950
No log 2.8571 120 0.8239 0.4613 0.8239 0.9077
No log 2.9048 122 0.8505 0.4712 0.8505 0.9222
No log 2.9524 124 0.8165 0.4599 0.8165 0.9036
No log 3.0 126 0.8336 0.4712 0.8336 0.9130
No log 3.0476 128 0.8458 0.4824 0.8458 0.9197
No log 3.0952 130 0.9637 0.4334 0.9637 0.9817
No log 3.1429 132 0.9637 0.4078 0.9637 0.9817
No log 3.1905 134 0.8754 0.4568 0.8754 0.9357
No log 3.2381 136 0.7896 0.4097 0.7896 0.8886
No log 3.2857 138 0.7922 0.4277 0.7922 0.8900
No log 3.3333 140 0.7939 0.3817 0.7939 0.8910
No log 3.3810 142 0.7949 0.4353 0.7949 0.8916
No log 3.4286 144 0.8554 0.4946 0.8554 0.9249
No log 3.4762 146 0.9155 0.5385 0.9155 0.9568
No log 3.5238 148 0.8351 0.5266 0.8351 0.9139
No log 3.5714 150 0.8160 0.5383 0.8160 0.9033
No log 3.6190 152 0.8352 0.5383 0.8352 0.9139
No log 3.6667 154 0.8356 0.5173 0.8356 0.9141
No log 3.7143 156 0.8425 0.5208 0.8425 0.9179
No log 3.7619 158 0.8522 0.3747 0.8522 0.9232
No log 3.8095 160 0.8541 0.3256 0.8541 0.9242
No log 3.8571 162 0.8661 0.4966 0.8661 0.9307
No log 3.9048 164 0.9456 0.4417 0.9456 0.9724
No log 3.9524 166 0.9279 0.3957 0.9279 0.9633
No log 4.0 168 0.9541 0.3024 0.9541 0.9768
No log 4.0476 170 0.9291 0.4234 0.9291 0.9639
No log 4.0952 172 0.8858 0.5128 0.8858 0.9412
No log 4.1429 174 0.8580 0.5255 0.8580 0.9263
No log 4.1905 176 0.8790 0.5302 0.8790 0.9375
No log 4.2381 178 0.9195 0.4570 0.9195 0.9589
No log 4.2857 180 0.9392 0.4427 0.9392 0.9691
No log 4.3333 182 0.9178 0.3908 0.9178 0.9580
No log 4.3810 184 0.9342 0.4031 0.9342 0.9665
No log 4.4286 186 0.8873 0.3733 0.8873 0.9420
No log 4.4762 188 0.8652 0.3733 0.8652 0.9302
No log 4.5238 190 0.8462 0.4598 0.8462 0.9199
No log 4.5714 192 0.8212 0.4824 0.8212 0.9062
No log 4.6190 194 0.8022 0.4555 0.8022 0.8957
No log 4.6667 196 0.8018 0.4676 0.8018 0.8955
No log 4.7143 198 0.8031 0.5116 0.8031 0.8961
No log 4.7619 200 0.8469 0.5076 0.8469 0.9203
No log 4.8095 202 0.8404 0.4982 0.8404 0.9168
No log 4.8571 204 0.8581 0.4725 0.8581 0.9263
No log 4.9048 206 0.8237 0.5116 0.8237 0.9076
No log 4.9524 208 0.8660 0.4439 0.8660 0.9306
No log 5.0 210 0.9112 0.4444 0.9112 0.9545
No log 5.0476 212 0.9125 0.4681 0.9125 0.9552
No log 5.0952 214 0.9758 0.3814 0.9758 0.9878
No log 5.1429 216 0.9718 0.3696 0.9718 0.9858
No log 5.1905 218 0.9050 0.4459 0.9050 0.9513
No log 5.2381 220 0.8668 0.4752 0.8668 0.9310
No log 5.2857 222 0.8554 0.4537 0.8554 0.9249
No log 5.3333 224 0.9045 0.4085 0.9045 0.9511
No log 5.3810 226 0.9724 0.4002 0.9724 0.9861
No log 5.4286 228 0.9461 0.4186 0.9461 0.9727
No log 5.4762 230 0.9472 0.4553 0.9472 0.9733
No log 5.5238 232 0.9708 0.4313 0.9708 0.9853
No log 5.5714 234 1.0847 0.3675 1.0847 1.0415
No log 5.6190 236 1.1206 0.2807 1.1206 1.0586
No log 5.6667 238 1.0129 0.3602 1.0129 1.0064
No log 5.7143 240 0.9604 0.4439 0.9604 0.9800
No log 5.7619 242 0.9606 0.4304 0.9606 0.9801
No log 5.8095 244 0.9528 0.4030 0.9528 0.9761
No log 5.8571 246 0.9389 0.3474 0.9389 0.9690
No log 5.9048 248 0.9250 0.3445 0.9250 0.9618
No log 5.9524 250 0.9052 0.3590 0.9052 0.9514
No log 6.0 252 0.9359 0.3695 0.9359 0.9674
No log 6.0476 254 0.8797 0.3879 0.8797 0.9379
No log 6.0952 256 0.8572 0.4329 0.8572 0.9258
No log 6.1429 258 0.8584 0.4065 0.8584 0.9265
No log 6.1905 260 0.8793 0.3765 0.8793 0.9377
No log 6.2381 262 0.8826 0.3596 0.8826 0.9394
No log 6.2857 264 0.8959 0.3762 0.8959 0.9465
No log 6.3333 266 0.9086 0.3802 0.9086 0.9532
No log 6.3810 268 0.9266 0.4086 0.9266 0.9626
No log 6.4286 270 0.9351 0.4976 0.9351 0.9670
No log 6.4762 272 0.9195 0.4976 0.9195 0.9589
No log 6.5238 274 0.9098 0.4752 0.9098 0.9538
No log 6.5714 276 0.8729 0.4086 0.8729 0.9343
No log 6.6190 278 0.8748 0.4861 0.8748 0.9353
No log 6.6667 280 0.9460 0.5128 0.9460 0.9726
No log 6.7143 282 1.1153 0.3516 1.1153 1.0561
No log 6.7619 284 1.0788 0.4100 1.0788 1.0387
No log 6.8095 286 0.8964 0.4220 0.8964 0.9468
No log 6.8571 288 0.8596 0.4269 0.8596 0.9271
No log 6.9048 290 0.8560 0.5062 0.8560 0.9252
No log 6.9524 292 0.9966 0.4361 0.9966 0.9983
No log 7.0 294 0.9577 0.4898 0.9577 0.9786
No log 7.0476 296 0.9003 0.5137 0.9003 0.9488
No log 7.0952 298 0.8789 0.5254 0.8789 0.9375
No log 7.1429 300 0.8382 0.4940 0.8382 0.9155
No log 7.1905 302 0.8421 0.4940 0.8421 0.9177
No log 7.2381 304 0.7845 0.4757 0.7845 0.8857
No log 7.2857 306 0.7830 0.4297 0.7830 0.8849
No log 7.3333 308 0.8523 0.5035 0.8523 0.9232
No log 7.3810 310 0.8791 0.5134 0.8791 0.9376
No log 7.4286 312 0.8210 0.5173 0.8210 0.9061
No log 7.4762 314 0.8199 0.5482 0.8199 0.9055
No log 7.5238 316 0.9092 0.48 0.9092 0.9535
No log 7.5714 318 1.0393 0.4786 1.0393 1.0195
No log 7.6190 320 0.9844 0.4497 0.9844 0.9922
No log 7.6667 322 0.9342 0.4707 0.9342 0.9665
No log 7.7143 324 0.8424 0.5279 0.8424 0.9178
No log 7.7619 326 0.8503 0.5067 0.8503 0.9221
No log 7.8095 328 0.8934 0.4575 0.8934 0.9452
No log 7.8571 330 0.8379 0.4966 0.8379 0.9153
No log 7.9048 332 0.8190 0.4368 0.8190 0.9050
No log 7.9524 334 0.8241 0.4960 0.8241 0.9078
No log 8.0 336 0.8375 0.5173 0.8375 0.9151
No log 8.0476 338 0.9074 0.4695 0.9074 0.9526
No log 8.0952 340 0.9257 0.4577 0.9257 0.9621
No log 8.1429 342 0.9120 0.3710 0.9120 0.9550
No log 8.1905 344 0.9389 0.2651 0.9389 0.9690
No log 8.2381 346 0.9503 0.2651 0.9503 0.9748
No log 8.2857 348 0.8926 0.2873 0.8926 0.9448
No log 8.3333 350 0.8589 0.4138 0.8589 0.9268
No log 8.3810 352 0.8485 0.4065 0.8485 0.9211
No log 8.4286 354 0.8427 0.4423 0.8427 0.9180
No log 8.4762 356 0.9176 0.4588 0.9176 0.9579
No log 8.5238 358 0.9572 0.4815 0.9572 0.9783
No log 8.5714 360 0.9676 0.4695 0.9676 0.9837
No log 8.6190 362 0.9577 0.4573 0.9577 0.9786
No log 8.6667 364 0.9494 0.4326 0.9494 0.9744
No log 8.7143 366 0.9419 0.4326 0.9419 0.9705
No log 8.7619 368 0.9431 0.4326 0.9431 0.9711
No log 8.8095 370 0.8859 0.5089 0.8859 0.9412
No log 8.8571 372 0.9342 0.4711 0.9342 0.9666
No log 8.9048 374 1.0265 0.4464 1.0265 1.0131
No log 8.9524 376 1.0069 0.4807 1.0069 1.0035
No log 9.0 378 0.9064 0.4711 0.9064 0.9521
No log 9.0476 380 0.8619 0.4515 0.8619 0.9284
No log 9.0952 382 0.8703 0.4254 0.8703 0.9329
No log 9.1429 384 0.9180 0.4343 0.9180 0.9581
No log 9.1905 386 0.8812 0.4352 0.8812 0.9387
No log 9.2381 388 0.8510 0.4515 0.8510 0.9225
No log 9.2857 390 0.8546 0.4507 0.8546 0.9244
No log 9.3333 392 0.8869 0.4829 0.8869 0.9418
No log 9.3810 394 0.9424 0.5255 0.9424 0.9708
No log 9.4286 396 1.0838 0.3802 1.0838 1.0410
No log 9.4762 398 1.0323 0.4278 1.0323 1.0160
No log 9.5238 400 0.8824 0.4257 0.8824 0.9393
No log 9.5714 402 0.8518 0.4145 0.8518 0.9229
No log 9.6190 404 0.8475 0.4145 0.8475 0.9206
No log 9.6667 406 0.8821 0.4089 0.8821 0.9392
No log 9.7143 408 0.9330 0.4916 0.9330 0.9659
No log 9.7619 410 1.0115 0.4681 1.0115 1.0058
No log 9.8095 412 0.9500 0.4807 0.9500 0.9747
No log 9.8571 414 0.8622 0.4853 0.8622 0.9286
No log 9.9048 416 0.8432 0.4336 0.8432 0.9182
No log 9.9524 418 0.8625 0.5059 0.8625 0.9287
No log 10.0 420 0.9765 0.4907 0.9765 0.9882
No log 10.0476 422 1.0012 0.4581 1.0012 1.0006
No log 10.0952 424 0.9030 0.4588 0.9030 0.9502
No log 10.1429 426 0.8287 0.3994 0.8287 0.9103
No log 10.1905 428 0.8444 0.3097 0.8444 0.9189
No log 10.2381 430 0.8421 0.2788 0.8421 0.9177
No log 10.2857 432 0.8393 0.3838 0.8393 0.9161
No log 10.3333 434 0.8649 0.3682 0.8649 0.9300
No log 10.3810 436 0.8959 0.4455 0.8959 0.9465
No log 10.4286 438 0.8810 0.4119 0.8810 0.9386
No log 10.4762 440 0.8657 0.4499 0.8657 0.9305
No log 10.5238 442 0.9176 0.5039 0.9176 0.9579
No log 10.5714 444 0.9390 0.5039 0.9390 0.9690
No log 10.6190 446 0.9063 0.4829 0.9063 0.9520
No log 10.6667 448 0.9063 0.4603 0.9063 0.9520
No log 10.7143 450 0.9312 0.4603 0.9312 0.9650
No log 10.7619 452 0.9106 0.4603 0.9106 0.9543
No log 10.8095 454 0.8950 0.4257 0.8950 0.9460
No log 10.8571 456 0.8995 0.4257 0.8995 0.9484
No log 10.9048 458 0.9349 0.4826 0.9349 0.9669
No log 10.9524 460 0.9290 0.4709 0.9290 0.9639
No log 11.0 462 0.9190 0.4604 0.9190 0.9586
No log 11.0476 464 0.8675 0.4280 0.8675 0.9314
No log 11.0952 466 0.8648 0.4280 0.8648 0.9299
No log 11.1429 468 0.8976 0.4119 0.8976 0.9474
No log 11.1905 470 0.9058 0.4131 0.9058 0.9517
No log 11.2381 472 0.8617 0.4503 0.8617 0.9283
No log 11.2857 474 0.8639 0.4507 0.8639 0.9295
No log 11.3333 476 0.9284 0.5039 0.9284 0.9635
No log 11.3810 478 0.9612 0.4920 0.9612 0.9804
No log 11.4286 480 0.8940 0.5365 0.8940 0.9455
No log 11.4762 482 0.8208 0.4768 0.8208 0.9060
No log 11.5238 484 0.8086 0.4534 0.8086 0.8992
No log 11.5714 486 0.8000 0.4531 0.8000 0.8944
No log 11.6190 488 0.8163 0.4407 0.8163 0.9035
No log 11.6667 490 0.8847 0.3824 0.8847 0.9406
No log 11.7143 492 0.9233 0.4555 0.9233 0.9609
No log 11.7619 494 0.8744 0.4570 0.8744 0.9351
No log 11.8095 496 0.8847 0.5013 0.8847 0.9406
No log 11.8571 498 0.9265 0.4898 0.9265 0.9626
0.3038 11.9048 500 0.9193 0.5013 0.9193 0.9588
0.3038 11.9524 502 0.8905 0.5013 0.8905 0.9437
0.3038 12.0 504 0.9127 0.5013 0.9127 0.9553
0.3038 12.0476 506 0.8915 0.5013 0.8915 0.9442
0.3038 12.0952 508 0.8456 0.5013 0.8456 0.9195
0.3038 12.1429 510 0.8038 0.5041 0.8038 0.8966
0.3038 12.1905 512 0.8431 0.5013 0.8431 0.9182
0.3038 12.2381 514 0.8390 0.4695 0.8390 0.9160
0.3038 12.2857 516 0.8123 0.5041 0.8123 0.9013
0.3038 12.3333 518 0.8072 0.5041 0.8072 0.8985
0.3038 12.3810 520 0.7938 0.4712 0.7938 0.8910
0.3038 12.4286 522 0.8004 0.3860 0.8004 0.8946
0.3038 12.4762 524 0.8081 0.3149 0.8081 0.8989
0.3038 12.5238 526 0.8279 0.3311 0.8279 0.9099
0.3038 12.5714 528 0.8341 0.3590 0.8341 0.9133

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k13_task5_organization

Finetuned
(4019)
this model