ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k3_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8761
  • Qwk: 0.4444
  • Mse: 0.8761
  • Rmse: 0.9360

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.2 2 4.6658 0.0010 4.6658 2.1600
No log 0.4 4 2.6463 0.0248 2.6463 1.6267
No log 0.6 6 2.0322 -0.0233 2.0322 1.4256
No log 0.8 8 2.0414 -0.0233 2.0414 1.4288
No log 1.0 10 1.5623 0.0082 1.5623 1.2499
No log 1.2 12 1.4239 0.0784 1.4239 1.1933
No log 1.4 14 1.3272 0.0208 1.3272 1.1520
No log 1.6 16 1.5016 0.1080 1.5016 1.2254
No log 1.8 18 1.5581 0.0248 1.5581 1.2482
No log 2.0 20 1.8508 0.0363 1.8508 1.3604
No log 2.2 22 1.5109 0.0082 1.5109 1.2292
No log 2.4 24 1.2663 0.1530 1.2663 1.1253
No log 2.6 26 1.3070 0.1679 1.3070 1.1433
No log 2.8 28 1.3786 0.1838 1.3786 1.1741
No log 3.0 30 1.4237 0.0477 1.4237 1.1932
No log 3.2 32 1.2937 0.2019 1.2937 1.1374
No log 3.4 34 1.1688 0.1542 1.1688 1.0811
No log 3.6 36 1.3481 0.2151 1.3481 1.1611
No log 3.8 38 2.0196 0.0473 2.0196 1.4211
No log 4.0 40 1.9269 0.0622 1.9269 1.3881
No log 4.2 42 1.4303 0.2096 1.4303 1.1960
No log 4.4 44 1.2082 0.2366 1.2082 1.0992
No log 4.6 46 1.1412 0.1481 1.1412 1.0683
No log 4.8 48 1.1422 0.1982 1.1422 1.0687
No log 5.0 50 1.3716 0.2028 1.3716 1.1712
No log 5.2 52 1.4224 0.2028 1.4224 1.1926
No log 5.4 54 1.2953 0.2431 1.2953 1.1381
No log 5.6 56 1.3258 0.2736 1.3258 1.1514
No log 5.8 58 1.3350 0.2825 1.3350 1.1554
No log 6.0 60 1.5080 0.2411 1.5080 1.2280
No log 6.2 62 1.5423 0.2000 1.5423 1.2419
No log 6.4 64 1.6616 0.2159 1.6616 1.2890
No log 6.6 66 1.2361 0.3380 1.2361 1.1118
No log 6.8 68 1.0510 0.3435 1.0510 1.0252
No log 7.0 70 1.0308 0.3204 1.0308 1.0153
No log 7.2 72 0.9747 0.4176 0.9747 0.9873
No log 7.4 74 1.3584 0.2941 1.3584 1.1655
No log 7.6 76 1.2811 0.2823 1.2811 1.1319
No log 7.8 78 0.9781 0.3782 0.9781 0.9890
No log 8.0 80 0.9396 0.3699 0.9396 0.9693
No log 8.2 82 1.0741 0.2937 1.0741 1.0364
No log 8.4 84 1.3477 0.2593 1.3477 1.1609
No log 8.6 86 1.2754 0.2333 1.2754 1.1293
No log 8.8 88 1.0031 0.3036 1.0031 1.0015
No log 9.0 90 1.0114 0.3571 1.0114 1.0057
No log 9.2 92 1.1195 0.4171 1.1195 1.0580
No log 9.4 94 1.1632 0.3697 1.1632 1.0785
No log 9.6 96 1.3954 0.2800 1.3954 1.1813
No log 9.8 98 1.5294 0.2634 1.5294 1.2367
No log 10.0 100 1.4504 0.2512 1.4504 1.2043
No log 10.2 102 1.0912 0.4171 1.0912 1.0446
No log 10.4 104 0.9819 0.3418 0.9819 0.9909
No log 10.6 106 0.9881 0.3110 0.9881 0.9940
No log 10.8 108 1.1117 0.3640 1.1117 1.0544
No log 11.0 110 1.4829 0.2549 1.4829 1.2177
No log 11.2 112 1.6834 0.2424 1.6834 1.2975
No log 11.4 114 1.4216 0.2331 1.4216 1.1923
No log 11.6 116 1.1585 0.3784 1.1585 1.0763
No log 11.8 118 1.0321 0.3451 1.0321 1.0159
No log 12.0 120 1.0322 0.2835 1.0322 1.0160
No log 12.2 122 1.0475 0.2936 1.0475 1.0235
No log 12.4 124 1.2994 0.1652 1.2994 1.1399
No log 12.6 126 1.5075 0.1967 1.5075 1.2278
No log 12.8 128 1.4200 0.1587 1.4200 1.1916
No log 13.0 130 1.1501 0.3152 1.1501 1.0724
No log 13.2 132 1.0226 0.3462 1.0226 1.0112
No log 13.4 134 1.0388 0.2674 1.0388 1.0192
No log 13.6 136 1.0046 0.2887 1.0046 1.0023
No log 13.8 138 0.9880 0.3502 0.9880 0.9940
No log 14.0 140 1.4261 0.2274 1.4261 1.1942
No log 14.2 142 1.7619 0.2155 1.7619 1.3274
No log 14.4 144 1.6462 0.1952 1.6462 1.2831
No log 14.6 146 1.2664 0.2607 1.2664 1.1254
No log 14.8 148 1.0097 0.3848 1.0097 1.0048
No log 15.0 150 0.9851 0.2939 0.9851 0.9925
No log 15.2 152 0.9889 0.3451 0.9889 0.9944
No log 15.4 154 1.1157 0.4214 1.1157 1.0563
No log 15.6 156 1.4143 0.1279 1.4143 1.1892
No log 15.8 158 1.5698 0.1854 1.5698 1.2529
No log 16.0 160 1.4908 0.1680 1.4908 1.2210
No log 16.2 162 1.4052 0.1084 1.4052 1.1854
No log 16.4 164 1.3059 0.1884 1.3059 1.1428
No log 16.6 166 1.2250 0.3030 1.2250 1.1068
No log 16.8 168 1.2374 0.3471 1.2374 1.1124
No log 17.0 170 1.1944 0.3144 1.1944 1.0929
No log 17.2 172 1.2374 0.3339 1.2374 1.1124
No log 17.4 174 1.4131 0.1916 1.4131 1.1888
No log 17.6 176 1.4658 0.1765 1.4658 1.2107
No log 17.8 178 1.2646 0.3261 1.2646 1.1245
No log 18.0 180 1.0491 0.3519 1.0491 1.0242
No log 18.2 182 0.9754 0.3942 0.9754 0.9876
No log 18.4 184 0.9545 0.3942 0.9545 0.9770
No log 18.6 186 0.9607 0.4347 0.9607 0.9801
No log 18.8 188 0.9419 0.4310 0.9419 0.9705
No log 19.0 190 0.9270 0.4408 0.9270 0.9628
No log 19.2 192 0.9812 0.4630 0.9812 0.9906
No log 19.4 194 1.0579 0.4815 1.0579 1.0286
No log 19.6 196 1.0868 0.4526 1.0868 1.0425
No log 19.8 198 1.0551 0.4500 1.0551 1.0272
No log 20.0 200 1.0761 0.4500 1.0761 1.0374
No log 20.2 202 0.9433 0.4723 0.9433 0.9712
No log 20.4 204 0.8742 0.4236 0.8742 0.9350
No log 20.6 206 0.8649 0.3914 0.8649 0.9300
No log 20.8 208 0.8870 0.4578 0.8870 0.9418
No log 21.0 210 0.9482 0.4542 0.9482 0.9738
No log 21.2 212 1.0303 0.4537 1.0303 1.0150
No log 21.4 214 0.9671 0.4676 0.9671 0.9834
No log 21.6 216 0.9209 0.4641 0.9209 0.9596
No log 21.8 218 0.9151 0.4507 0.9151 0.9566
No log 22.0 220 0.9308 0.4273 0.9308 0.9648
No log 22.2 222 1.0792 0.4287 1.0792 1.0389
No log 22.4 224 1.3730 0.2245 1.3730 1.1718
No log 22.6 226 1.3895 0.2165 1.3895 1.1788
No log 22.8 228 1.1353 0.4240 1.1353 1.0655
No log 23.0 230 0.8846 0.4 0.8846 0.9405
No log 23.2 232 0.8525 0.3960 0.8525 0.9233
No log 23.4 234 0.8625 0.4598 0.8625 0.9287
No log 23.6 236 0.9965 0.4473 0.9965 0.9982
No log 23.8 238 1.1504 0.4022 1.1504 1.0726
No log 24.0 240 1.1637 0.3936 1.1637 1.0787
No log 24.2 242 1.0749 0.4989 1.0749 1.0368
No log 24.4 244 0.9532 0.4507 0.9532 0.9763
No log 24.6 246 0.9367 0.3356 0.9367 0.9678
No log 24.8 248 0.9363 0.3356 0.9363 0.9677
No log 25.0 250 0.9395 0.3859 0.9395 0.9693
No log 25.2 252 0.9939 0.4214 0.9939 0.9969
No log 25.4 254 1.0969 0.4263 1.0969 1.0473
No log 25.6 256 1.0944 0.3822 1.0944 1.0461
No log 25.8 258 1.0335 0.4098 1.0335 1.0166
No log 26.0 260 0.9322 0.5139 0.9322 0.9655
No log 26.2 262 0.9131 0.4375 0.9131 0.9556
No log 26.4 264 0.9407 0.4888 0.9407 0.9699
No log 26.6 266 1.0016 0.4842 1.0016 1.0008
No log 26.8 268 0.9906 0.4787 0.9906 0.9953
No log 27.0 270 0.9402 0.4237 0.9402 0.9697
No log 27.2 272 0.9122 0.4157 0.9122 0.9551
No log 27.4 274 0.9310 0.4051 0.9310 0.9649
No log 27.6 276 0.9684 0.4624 0.9684 0.9841
No log 27.8 278 0.9793 0.4650 0.9793 0.9896
No log 28.0 280 0.9409 0.4136 0.9409 0.9700
No log 28.2 282 0.9210 0.4136 0.9210 0.9597
No log 28.4 284 0.9294 0.4408 0.9294 0.9640
No log 28.6 286 0.9246 0.4408 0.9246 0.9616
No log 28.8 288 0.9646 0.4907 0.9646 0.9821
No log 29.0 290 1.0451 0.4577 1.0451 1.0223
No log 29.2 292 1.0750 0.4484 1.0750 1.0368
No log 29.4 294 1.0733 0.4484 1.0733 1.0360
No log 29.6 296 0.9928 0.4641 0.9928 0.9964
No log 29.8 298 0.9520 0.4375 0.9520 0.9757
No log 30.0 300 0.9532 0.4341 0.9532 0.9763
No log 30.2 302 0.9941 0.5139 0.9941 0.9970
No log 30.4 304 0.9720 0.4728 0.9720 0.9859
No log 30.6 306 0.9232 0.4450 0.9232 0.9609
No log 30.8 308 0.9274 0.4728 0.9274 0.9630
No log 31.0 310 1.0408 0.4516 1.0408 1.0202
No log 31.2 312 1.2667 0.2436 1.2667 1.1255
No log 31.4 314 1.2675 0.2436 1.2675 1.1258
No log 31.6 316 1.1097 0.3140 1.1097 1.0534
No log 31.8 318 0.9286 0.4728 0.9286 0.9636
No log 32.0 320 0.8744 0.4511 0.8744 0.9351
No log 32.2 322 0.8801 0.4371 0.8801 0.9381
No log 32.4 324 0.9132 0.4833 0.9132 0.9556
No log 32.6 326 1.0196 0.4509 1.0196 1.0097
No log 32.8 328 1.1231 0.3574 1.1231 1.0598
No log 33.0 330 1.1409 0.3435 1.1409 1.0681
No log 33.2 332 1.0493 0.4214 1.0493 1.0244
No log 33.4 334 0.9514 0.4236 0.9514 0.9754
No log 33.6 336 0.9426 0.4549 0.9426 0.9709
No log 33.8 338 0.9223 0.4157 0.9223 0.9604
No log 34.0 340 0.9506 0.4717 0.9506 0.9750
No log 34.2 342 1.0709 0.4096 1.0709 1.0348
No log 34.4 344 1.2458 0.2436 1.2458 1.1162
No log 34.6 346 1.1844 0.2962 1.1844 1.0883
No log 34.8 348 1.0427 0.4321 1.0427 1.0211
No log 35.0 350 0.9001 0.4741 0.9001 0.9487
No log 35.2 352 0.8487 0.4888 0.8487 0.9213
No log 35.4 354 0.8382 0.4220 0.8382 0.9155
No log 35.6 356 0.8404 0.3974 0.8404 0.9167
No log 35.8 358 0.8444 0.3663 0.8444 0.9189
No log 36.0 360 0.8471 0.4681 0.8471 0.9204
No log 36.2 362 0.8447 0.4681 0.8447 0.9191
No log 36.4 364 0.8357 0.4635 0.8357 0.9142
No log 36.6 366 0.8434 0.4079 0.8434 0.9183
No log 36.8 368 0.8487 0.4527 0.8487 0.9213
No log 37.0 370 0.8729 0.4719 0.8729 0.9343
No log 37.2 372 0.9291 0.4469 0.9291 0.9639
No log 37.4 374 0.9772 0.4191 0.9772 0.9885
No log 37.6 376 0.9445 0.4078 0.9445 0.9719
No log 37.8 378 0.8933 0.4337 0.8933 0.9451
No log 38.0 380 0.8841 0.3404 0.8841 0.9403
No log 38.2 382 0.8788 0.3685 0.8788 0.9374
No log 38.4 384 0.8718 0.4363 0.8718 0.9337
No log 38.6 386 0.8891 0.4299 0.8891 0.9429
No log 38.8 388 0.9248 0.3985 0.9248 0.9617
No log 39.0 390 0.9492 0.4596 0.9492 0.9743
No log 39.2 392 1.0323 0.4061 1.0323 1.0160
No log 39.4 394 1.0585 0.4191 1.0585 1.0288
No log 39.6 396 1.0652 0.4191 1.0652 1.0321
No log 39.8 398 1.0038 0.4155 1.0038 1.0019
No log 40.0 400 1.0065 0.4155 1.0065 1.0032
No log 40.2 402 1.0026 0.4155 1.0026 1.0013
No log 40.4 404 0.9867 0.4076 0.9867 0.9933
No log 40.6 406 0.9324 0.4444 0.9324 0.9656
No log 40.8 408 0.8901 0.4371 0.8901 0.9435
No log 41.0 410 0.8642 0.4367 0.8642 0.9296
No log 41.2 412 0.8475 0.4367 0.8475 0.9206
No log 41.4 414 0.8543 0.3539 0.8543 0.9243
No log 41.6 416 0.8545 0.4428 0.8545 0.9244
No log 41.8 418 0.8568 0.4012 0.8568 0.9256
No log 42.0 420 0.8862 0.4301 0.8862 0.9414
No log 42.2 422 0.9251 0.5132 0.9251 0.9618
No log 42.4 424 0.9043 0.4587 0.9043 0.9510
No log 42.6 426 0.8790 0.4861 0.8790 0.9375
No log 42.8 428 0.8572 0.4861 0.8572 0.9258
No log 43.0 430 0.8478 0.4861 0.8478 0.9208
No log 43.2 432 0.8384 0.4598 0.8384 0.9157
No log 43.4 434 0.8369 0.4472 0.8369 0.9148
No log 43.6 436 0.8422 0.4075 0.8422 0.9177
No log 43.8 438 0.8560 0.4054 0.8560 0.9252
No log 44.0 440 0.8857 0.4752 0.8857 0.9411
No log 44.2 442 0.8974 0.5139 0.8974 0.9473
No log 44.4 444 0.8626 0.4606 0.8626 0.9287
No log 44.6 446 0.8188 0.4334 0.8188 0.9049
No log 44.8 448 0.8103 0.4180 0.8103 0.9002
No log 45.0 450 0.8495 0.4638 0.8495 0.9217
No log 45.2 452 0.8633 0.3847 0.8633 0.9291
No log 45.4 454 0.8608 0.4393 0.8608 0.9278
No log 45.6 456 0.8338 0.4398 0.8338 0.9131
No log 45.8 458 0.8257 0.4072 0.8257 0.9087
No log 46.0 460 0.8411 0.4411 0.8411 0.9171
No log 46.2 462 0.8375 0.4728 0.8375 0.9151
No log 46.4 464 0.8056 0.4903 0.8056 0.8976
No log 46.6 466 0.7914 0.4180 0.7914 0.8896
No log 46.8 468 0.8074 0.4526 0.8074 0.8985
No log 47.0 470 0.8089 0.4526 0.8089 0.8994
No log 47.2 472 0.8186 0.4286 0.8186 0.9048
No log 47.4 474 0.8128 0.4286 0.8128 0.9015
No log 47.6 476 0.8009 0.4579 0.8009 0.8949
No log 47.8 478 0.8094 0.4903 0.8094 0.8997
No log 48.0 480 0.8567 0.4868 0.8567 0.9256
No log 48.2 482 0.8791 0.4868 0.8791 0.9376
No log 48.4 484 0.8532 0.4728 0.8532 0.9237
No log 48.6 486 0.8202 0.4439 0.8202 0.9057
No log 48.8 488 0.8042 0.4784 0.8042 0.8968
No log 49.0 490 0.7977 0.4784 0.7977 0.8931
No log 49.2 492 0.7978 0.4784 0.7978 0.8932
No log 49.4 494 0.7960 0.4645 0.7960 0.8922
No log 49.6 496 0.8021 0.4479 0.8021 0.8956
No log 49.8 498 0.7945 0.4479 0.7945 0.8914
0.2535 50.0 500 0.7955 0.4479 0.7955 0.8919
0.2535 50.2 502 0.7970 0.4115 0.7970 0.8928
0.2535 50.4 504 0.7885 0.4760 0.7885 0.8880
0.2535 50.6 506 0.7783 0.4748 0.7783 0.8822
0.2535 50.8 508 0.7667 0.4748 0.7667 0.8756
0.2535 51.0 510 0.7716 0.4671 0.7716 0.8784
0.2535 51.2 512 0.7876 0.4334 0.7876 0.8875
0.2535 51.4 514 0.8178 0.4157 0.8178 0.9043
0.2535 51.6 516 0.8509 0.4557 0.8509 0.9224
0.2535 51.8 518 0.8710 0.4557 0.8710 0.9333
0.2535 52.0 520 0.8954 0.4557 0.8954 0.9463
0.2535 52.2 522 0.9225 0.4315 0.9225 0.9605
0.2535 52.4 524 0.9008 0.4483 0.9008 0.9491
0.2535 52.6 526 0.8922 0.4094 0.8922 0.9446
0.2535 52.8 528 0.8677 0.4518 0.8677 0.9315
0.2535 53.0 530 0.8520 0.3970 0.8520 0.9230
0.2535 53.2 532 0.8491 0.4428 0.8491 0.9215
0.2535 53.4 534 0.8564 0.4324 0.8564 0.9254
0.2535 53.6 536 0.8769 0.4056 0.8769 0.9365
0.2535 53.8 538 0.9362 0.4570 0.9362 0.9676
0.2535 54.0 540 1.0353 0.3783 1.0353 1.0175
0.2535 54.2 542 1.0580 0.3913 1.0580 1.0286
0.2535 54.4 544 0.9995 0.4227 0.9995 0.9997
0.2535 54.6 546 0.9062 0.4310 0.9062 0.9520
0.2535 54.8 548 0.8761 0.4444 0.8761 0.9360

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k3_task2_organization

Finetuned
(4019)
this model