ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k14_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5750
  • Qwk: 0.7815
  • Mse: 0.5750
  • Rmse: 0.7583

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0303 2 6.8048 0.0239 6.8048 2.6086
No log 0.0606 4 4.3326 0.0779 4.3326 2.0815
No log 0.0909 6 3.2583 -0.0235 3.2583 1.8051
No log 0.1212 8 2.3963 0.0588 2.3963 1.5480
No log 0.1515 10 2.3642 0.0323 2.3642 1.5376
No log 0.1818 12 2.2371 0.0168 2.2371 1.4957
No log 0.2121 14 1.6736 0.2037 1.6736 1.2937
No log 0.2424 16 2.5288 0.0544 2.5288 1.5902
No log 0.2727 18 3.3273 0.0588 3.3273 1.8241
No log 0.3030 20 2.8383 0.0497 2.8383 1.6847
No log 0.3333 22 1.9842 0.0945 1.9842 1.4086
No log 0.3636 24 1.4727 0.2385 1.4727 1.2135
No log 0.3939 26 1.4542 0.3091 1.4542 1.2059
No log 0.4242 28 1.5773 0.1869 1.5773 1.2559
No log 0.4545 30 1.9017 0.2435 1.9017 1.3790
No log 0.4848 32 2.3291 -0.0141 2.3291 1.5261
No log 0.5152 34 2.7412 0.0 2.7412 1.6556
No log 0.5455 36 3.0232 0.0727 3.0232 1.7387
No log 0.5758 38 2.9944 0.0727 2.9944 1.7304
No log 0.6061 40 2.7726 0.0629 2.7726 1.6651
No log 0.6364 42 2.3466 0.1370 2.3466 1.5319
No log 0.6667 44 2.0539 0.2519 2.0539 1.4331
No log 0.6970 46 2.0418 0.2353 2.0418 1.4289
No log 0.7273 48 2.1151 0.2676 2.1151 1.4544
No log 0.7576 50 1.8982 0.3143 1.8982 1.3778
No log 0.7879 52 1.9557 0.3099 1.9557 1.3985
No log 0.8182 54 2.3719 0.2323 2.3719 1.5401
No log 0.8485 56 2.8996 0.2105 2.8996 1.7028
No log 0.8788 58 3.2513 0.1485 3.2513 1.8031
No log 0.9091 60 2.9194 0.2174 2.9194 1.7086
No log 0.9394 62 2.7612 0.2703 2.7612 1.6617
No log 0.9697 64 2.4625 0.2793 2.4625 1.5692
No log 1.0 66 2.1567 0.3699 2.1567 1.4686
No log 1.0303 68 1.5443 0.4615 1.5443 1.2427
No log 1.0606 70 1.3352 0.4122 1.3352 1.1555
No log 1.0909 72 1.2966 0.4889 1.2966 1.1387
No log 1.1212 74 1.3357 0.4928 1.3357 1.1557
No log 1.1515 76 1.2812 0.5362 1.2812 1.1319
No log 1.1818 78 1.0504 0.5414 1.0504 1.0249
No log 1.2121 80 0.9545 0.6232 0.9545 0.9770
No log 1.2424 82 1.0937 0.6065 1.0937 1.0458
No log 1.2727 84 1.6117 0.5217 1.6117 1.2695
No log 1.3030 86 1.4665 0.5652 1.4665 1.2110
No log 1.3333 88 0.9558 0.6541 0.9558 0.9777
No log 1.3636 90 0.8601 0.7355 0.8601 0.9274
No log 1.3939 92 0.8936 0.7342 0.8936 0.9453
No log 1.4242 94 0.9324 0.6710 0.9324 0.9656
No log 1.4545 96 0.9349 0.6710 0.9349 0.9669
No log 1.4848 98 0.9365 0.6143 0.9365 0.9678
No log 1.5152 100 0.9384 0.6143 0.9384 0.9687
No log 1.5455 102 0.9491 0.6225 0.9491 0.9742
No log 1.5758 104 1.1166 0.6456 1.1166 1.0567
No log 1.6061 106 1.6842 0.5528 1.6842 1.2978
No log 1.6364 108 1.6815 0.5528 1.6815 1.2967
No log 1.6667 110 1.3229 0.5889 1.3229 1.1502
No log 1.6970 112 1.1122 0.6289 1.1122 1.0546
No log 1.7273 114 0.8960 0.6405 0.8960 0.9466
No log 1.7576 116 0.7982 0.6897 0.7982 0.8934
No log 1.7879 118 0.8376 0.6483 0.8376 0.9152
No log 1.8182 120 1.1409 0.6104 1.1409 1.0681
No log 1.8485 122 1.2750 0.6199 1.2750 1.1292
No log 1.8788 124 1.0995 0.6456 1.0995 1.0486
No log 1.9091 126 1.0231 0.6914 1.0231 1.0115
No log 1.9394 128 0.8192 0.7284 0.8192 0.9051
No log 1.9697 130 0.8177 0.7485 0.8177 0.9043
No log 2.0 132 0.9070 0.7283 0.9070 0.9524
No log 2.0303 134 0.8725 0.7386 0.8725 0.9341
No log 2.0606 136 0.8981 0.7126 0.8981 0.9477
No log 2.0909 138 0.7954 0.7407 0.7954 0.8919
No log 2.1212 140 0.8536 0.7089 0.8536 0.9239
No log 2.1515 142 0.9976 0.6790 0.9976 0.9988
No log 2.1818 144 1.2551 0.6591 1.2551 1.1203
No log 2.2121 146 1.2064 0.6778 1.2064 1.0984
No log 2.2424 148 0.7530 0.7237 0.7530 0.8678
No log 2.2727 150 0.6756 0.7368 0.6756 0.8220
No log 2.3030 152 0.7321 0.7308 0.7321 0.8556
No log 2.3333 154 0.8242 0.7394 0.8242 0.9079
No log 2.3636 156 0.6986 0.7607 0.6986 0.8358
No log 2.3939 158 0.6359 0.7733 0.6359 0.7974
No log 2.4242 160 0.7112 0.7237 0.7112 0.8433
No log 2.4545 162 0.6614 0.7898 0.6614 0.8133
No log 2.4848 164 0.8504 0.7293 0.8504 0.9222
No log 2.5152 166 1.0075 0.6995 1.0075 1.0038
No log 2.5455 168 0.8820 0.6826 0.8820 0.9391
No log 2.5758 170 0.8461 0.7190 0.8461 0.9198
No log 2.6061 172 0.8629 0.7020 0.8629 0.9289
No log 2.6364 174 1.0784 0.6707 1.0784 1.0385
No log 2.6667 176 1.2121 0.6740 1.2121 1.1010
No log 2.6970 178 1.0610 0.6821 1.0610 1.0300
No log 2.7273 180 0.9010 0.7219 0.9010 0.9492
No log 2.7576 182 1.0027 0.6821 1.0027 1.0013
No log 2.7879 184 1.1098 0.6705 1.1098 1.0535
No log 2.8182 186 1.0315 0.6591 1.0315 1.0156
No log 2.8485 188 0.8681 0.6962 0.8681 0.9317
No log 2.8788 190 0.8561 0.7248 0.8561 0.9253
No log 2.9091 192 0.9197 0.6883 0.9197 0.9590
No log 2.9394 194 1.0324 0.6190 1.0324 1.0161
No log 2.9697 196 1.0266 0.6437 1.0266 1.0132
No log 3.0 198 0.8521 0.7125 0.8521 0.9231
No log 3.0303 200 0.7187 0.7361 0.7187 0.8478
No log 3.0606 202 0.7887 0.7183 0.7887 0.8881
No log 3.0909 204 0.7213 0.7092 0.7213 0.8493
No log 3.1212 206 0.6804 0.7517 0.6804 0.8249
No log 3.1515 208 0.9165 0.6893 0.9165 0.9573
No log 3.1818 210 1.1211 0.6358 1.1211 1.0588
No log 3.2121 212 1.0346 0.6629 1.0346 1.0172
No log 3.2424 214 0.8061 0.7531 0.8061 0.8978
No log 3.2727 216 0.7957 0.7413 0.7957 0.8920
No log 3.3030 218 0.9818 0.7111 0.9818 0.9908
No log 3.3333 220 0.9185 0.7111 0.9185 0.9584
No log 3.3636 222 0.7742 0.7273 0.7742 0.8799
No log 3.3939 224 0.8575 0.7097 0.8575 0.9260
No log 3.4242 226 1.1411 0.6061 1.1411 1.0682
No log 3.4545 228 1.2181 0.6061 1.2181 1.1037
No log 3.4848 230 1.0425 0.6145 1.0425 1.0210
No log 3.5152 232 0.8155 0.6800 0.8155 0.9030
No log 3.5455 234 0.6700 0.7133 0.6700 0.8185
No log 3.5758 236 0.6797 0.7552 0.6797 0.8244
No log 3.6061 238 0.6645 0.7448 0.6645 0.8152
No log 3.6364 240 0.6303 0.7310 0.6303 0.7939
No log 3.6667 242 0.6286 0.76 0.6286 0.7928
No log 3.6970 244 0.6348 0.7792 0.6348 0.7967
No log 3.7273 246 0.6446 0.7712 0.6446 0.8028
No log 3.7576 248 0.6991 0.8050 0.6991 0.8361
No log 3.7879 250 0.7934 0.7262 0.7934 0.8907
No log 3.8182 252 0.7748 0.7515 0.7748 0.8802
No log 3.8485 254 0.7417 0.7712 0.7417 0.8612
No log 3.8788 256 0.7236 0.7712 0.7236 0.8507
No log 3.9091 258 0.6980 0.7975 0.6980 0.8355
No log 3.9394 260 0.6697 0.7975 0.6697 0.8184
No log 3.9697 262 0.6497 0.7949 0.6497 0.8060
No log 4.0 264 0.6449 0.7871 0.6449 0.8030
No log 4.0303 266 0.6572 0.8284 0.6572 0.8107
No log 4.0606 268 0.8805 0.7143 0.8805 0.9384
No log 4.0909 270 1.1404 0.6947 1.1404 1.0679
No log 4.1212 272 1.0995 0.6774 1.0995 1.0486
No log 4.1515 274 0.8756 0.7079 0.8756 0.9357
No log 4.1818 276 0.7334 0.8121 0.7334 0.8564
No log 4.2121 278 0.7583 0.7564 0.7583 0.8708
No log 4.2424 280 0.8075 0.7190 0.8075 0.8986
No log 4.2727 282 0.9641 0.6623 0.9641 0.9819
No log 4.3030 284 1.1215 0.5479 1.1215 1.0590
No log 4.3333 286 1.0705 0.6267 1.0705 1.0346
No log 4.3636 288 0.9364 0.6883 0.9364 0.9677
No log 4.3939 290 0.7510 0.7595 0.7510 0.8666
No log 4.4242 292 0.6556 0.775 0.6556 0.8097
No log 4.4545 294 0.6419 0.7821 0.6419 0.8012
No log 4.4848 296 0.6494 0.775 0.6494 0.8058
No log 4.5152 298 0.7548 0.7665 0.7548 0.8688
No log 4.5455 300 1.1197 0.6474 1.1197 1.0581
No log 4.5758 302 1.3629 0.5889 1.3629 1.1674
No log 4.6061 304 1.3104 0.6180 1.3104 1.1447
No log 4.6364 306 1.0263 0.6506 1.0263 1.0130
No log 4.6667 308 0.7658 0.7665 0.7658 0.8751
No log 4.6970 310 0.6923 0.7771 0.6923 0.8320
No log 4.7273 312 0.6879 0.7742 0.6879 0.8294
No log 4.7576 314 0.7061 0.7662 0.7061 0.8403
No log 4.7879 316 0.7088 0.7662 0.7088 0.8419
No log 4.8182 318 0.6860 0.7702 0.6860 0.8282
No log 4.8485 320 0.7049 0.7625 0.7049 0.8396
No log 4.8788 322 0.6542 0.7673 0.6542 0.8088
No log 4.9091 324 0.6280 0.7662 0.6280 0.7925
No log 4.9394 326 0.6326 0.7632 0.6326 0.7954
No log 4.9697 328 0.6191 0.7682 0.6191 0.7868
No log 5.0 330 0.6230 0.7950 0.6230 0.7893
No log 5.0303 332 0.7428 0.7614 0.7428 0.8619
No log 5.0606 334 0.8553 0.7444 0.8553 0.9248
No log 5.0909 336 0.7473 0.7486 0.7473 0.8645
No log 5.1212 338 0.6517 0.7791 0.6517 0.8073
No log 5.1515 340 0.5965 0.7879 0.5965 0.7723
No log 5.1818 342 0.6064 0.7901 0.6064 0.7787
No log 5.2121 344 0.6370 0.7607 0.6370 0.7981
No log 5.2424 346 0.7142 0.7738 0.7142 0.8451
No log 5.2727 348 0.6849 0.7636 0.6849 0.8276
No log 5.3030 350 0.6101 0.7879 0.6101 0.7811
No log 5.3333 352 0.6090 0.7879 0.6090 0.7804
No log 5.3636 354 0.6593 0.7816 0.6593 0.8120
No log 5.3939 356 0.6485 0.7907 0.6485 0.8053
No log 5.4242 358 0.6079 0.8047 0.6079 0.7797
No log 5.4545 360 0.5715 0.7925 0.5715 0.7560
No log 5.4848 362 0.5649 0.7901 0.5649 0.7516
No log 5.5152 364 0.5953 0.7976 0.5953 0.7716
No log 5.5455 366 0.6248 0.7931 0.6248 0.7904
No log 5.5758 368 0.6083 0.7882 0.6083 0.7800
No log 5.6061 370 0.5691 0.7843 0.5691 0.7544
No log 5.6364 372 0.5633 0.8052 0.5633 0.7505
No log 5.6667 374 0.5628 0.8 0.5628 0.7502
No log 5.6970 376 0.5895 0.7953 0.5895 0.7678
No log 5.7273 378 0.6438 0.7861 0.6438 0.8024
No log 5.7576 380 0.7377 0.7556 0.7377 0.8589
No log 5.7879 382 0.7724 0.7717 0.7724 0.8788
No log 5.8182 384 0.8010 0.7650 0.8010 0.8950
No log 5.8485 386 0.8682 0.7403 0.8682 0.9318
No log 5.8788 388 0.8724 0.7403 0.8724 0.9340
No log 5.9091 390 0.7580 0.7368 0.7580 0.8706
No log 5.9394 392 0.6957 0.7746 0.6957 0.8341
No log 5.9697 394 0.6585 0.8047 0.6585 0.8115
No log 6.0 396 0.6907 0.8047 0.6907 0.8311
No log 6.0303 398 0.7823 0.7294 0.7823 0.8845
No log 6.0606 400 0.9415 0.7151 0.9415 0.9703
No log 6.0909 402 1.0059 0.7302 1.0059 1.0029
No log 6.1212 404 0.8967 0.7322 0.8967 0.9469
No log 6.1515 406 0.7387 0.7746 0.7387 0.8595
No log 6.1818 408 0.6740 0.7861 0.6740 0.8209
No log 6.2121 410 0.6745 0.8075 0.6745 0.8213
No log 6.2424 412 0.6784 0.7791 0.6784 0.8236
No log 6.2727 414 0.6671 0.7746 0.6671 0.8168
No log 6.3030 416 0.6028 0.8095 0.6028 0.7764
No log 6.3333 418 0.5820 0.8193 0.5820 0.7629
No log 6.3636 420 0.5990 0.8095 0.5990 0.7740
No log 6.3939 422 0.7051 0.7865 0.7051 0.8397
No log 6.4242 424 0.7788 0.7778 0.7788 0.8825
No log 6.4545 426 0.7237 0.7865 0.7237 0.8507
No log 6.4848 428 0.6743 0.8095 0.6743 0.8212
No log 6.5152 430 0.6465 0.7826 0.6465 0.8041
No log 6.5455 432 0.6485 0.7771 0.6485 0.8053
No log 6.5758 434 0.6513 0.7925 0.6513 0.8070
No log 6.6061 436 0.7047 0.8024 0.7047 0.8394
No log 6.6364 438 0.7489 0.7614 0.7489 0.8654
No log 6.6667 440 0.8218 0.7650 0.8218 0.9065
No log 6.6970 442 0.8192 0.7513 0.8192 0.9051
No log 6.7273 444 0.9179 0.7263 0.9179 0.9580
No log 6.7576 446 0.8854 0.7322 0.8854 0.9409
No log 6.7879 448 0.8396 0.7471 0.8396 0.9163
No log 6.8182 450 0.6788 0.7882 0.6788 0.8239
No log 6.8485 452 0.6003 0.7821 0.6003 0.7748
No log 6.8788 454 0.5884 0.7821 0.5884 0.7670
No log 6.9091 456 0.5964 0.8193 0.5964 0.7723
No log 6.9394 458 0.6789 0.7701 0.6789 0.8240
No log 6.9697 460 0.8128 0.7528 0.8128 0.9016
No log 7.0 462 0.9492 0.6893 0.9492 0.9743
No log 7.0303 464 0.9221 0.7209 0.9221 0.9603
No log 7.0606 466 0.7858 0.7381 0.7858 0.8864
No log 7.0909 468 0.6554 0.8049 0.6554 0.8096
No log 7.1212 470 0.6245 0.7483 0.6245 0.7902
No log 7.1515 472 0.6277 0.76 0.6277 0.7923
No log 7.1818 474 0.6244 0.7712 0.6244 0.7902
No log 7.2121 476 0.6503 0.7730 0.6503 0.8064
No log 7.2424 478 0.7298 0.7647 0.7298 0.8543
No log 7.2727 480 0.7633 0.7574 0.7633 0.8736
No log 7.3030 482 0.7093 0.7647 0.7093 0.8422
No log 7.3333 484 0.6392 0.7613 0.6392 0.7995
No log 7.3636 486 0.6268 0.7517 0.6268 0.7917
No log 7.3939 488 0.6351 0.7172 0.6351 0.7969
No log 7.4242 490 0.6004 0.76 0.6004 0.7748
No log 7.4545 492 0.6290 0.8046 0.6290 0.7931
No log 7.4848 494 0.7432 0.7416 0.7432 0.8621
No log 7.5152 496 0.7882 0.7403 0.7882 0.8878
No log 7.5455 498 0.7699 0.7345 0.7699 0.8774
0.4203 7.5758 500 0.7020 0.7907 0.7020 0.8378
0.4203 7.6061 502 0.6615 0.7826 0.6615 0.8133
0.4203 7.6364 504 0.6598 0.7805 0.6598 0.8123
0.4203 7.6667 506 0.6857 0.7811 0.6857 0.8281
0.4203 7.6970 508 0.6784 0.8 0.6784 0.8236
0.4203 7.7273 510 0.6321 0.7882 0.6321 0.7950
0.4203 7.7576 512 0.5820 0.7879 0.5820 0.7629
0.4203 7.7879 514 0.5746 0.7815 0.5746 0.7580
0.4203 7.8182 516 0.5887 0.7619 0.5887 0.7673
0.4203 7.8485 518 0.5750 0.7815 0.5750 0.7583

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k14_task1_organization

Finetuned
(4019)
this model