ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k8_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8158
  • Qwk: -0.0686
  • Mse: 0.8158
  • Rmse: 0.9032

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0488 2 3.5423 -0.0154 3.5423 1.8821
No log 0.0976 4 1.8229 0.0213 1.8229 1.3502
No log 0.1463 6 1.4498 0.0 1.4498 1.2041
No log 0.1951 8 0.9423 -0.0316 0.9423 0.9707
No log 0.2439 10 0.8447 -0.0823 0.8447 0.9191
No log 0.2927 12 1.2237 0.0604 1.2237 1.1062
No log 0.3415 14 0.9227 0.0207 0.9227 0.9606
No log 0.3902 16 0.7258 0.0759 0.7258 0.8519
No log 0.4390 18 0.7933 0.0512 0.7933 0.8907
No log 0.4878 20 0.9857 0.0623 0.9857 0.9928
No log 0.5366 22 1.0405 -0.0423 1.0405 1.0201
No log 0.5854 24 1.0262 -0.0987 1.0262 1.0130
No log 0.6341 26 0.8952 0.0017 0.8952 0.9462
No log 0.6829 28 0.8068 0.0225 0.8068 0.8982
No log 0.7317 30 0.9742 -0.1253 0.9742 0.9870
No log 0.7805 32 1.2602 0.0041 1.2602 1.1226
No log 0.8293 34 1.1629 0.0566 1.1629 1.0784
No log 0.8780 36 0.9727 0.0084 0.9727 0.9862
No log 0.9268 38 0.9285 -0.0806 0.9285 0.9636
No log 0.9756 40 0.9869 0.0069 0.9869 0.9934
No log 1.0244 42 1.1949 0.0155 1.1949 1.0931
No log 1.0732 44 1.0275 0.0764 1.0275 1.0136
No log 1.1220 46 1.2343 0.0535 1.2343 1.1110
No log 1.1707 48 1.1250 -0.0237 1.1250 1.0607
No log 1.2195 50 0.9911 0.0448 0.9911 0.9956
No log 1.2683 52 0.9310 0.0541 0.9310 0.9649
No log 1.3171 54 1.1348 0.0525 1.1348 1.0653
No log 1.3659 56 1.1923 0.0027 1.1923 1.0919
No log 1.4146 58 0.9066 -0.0313 0.9066 0.9522
No log 1.4634 60 1.5071 0.0112 1.5071 1.2276
No log 1.5122 62 1.5473 0.0601 1.5473 1.2439
No log 1.5610 64 0.9769 -0.0007 0.9769 0.9884
No log 1.6098 66 0.9294 0.0134 0.9294 0.9640
No log 1.6585 68 0.9563 0.0064 0.9563 0.9779
No log 1.7073 70 1.2006 -0.0539 1.2006 1.0957
No log 1.7561 72 1.4031 -0.0374 1.4031 1.1845
No log 1.8049 74 1.0807 -0.0110 1.0807 1.0396
No log 1.8537 76 1.1539 0.0451 1.1539 1.0742
No log 1.9024 78 1.3512 -0.0305 1.3512 1.1624
No log 1.9512 80 1.0445 0.0028 1.0445 1.0220
No log 2.0 82 1.0789 -0.0211 1.0789 1.0387
No log 2.0488 84 1.4486 -0.0456 1.4486 1.2036
No log 2.0976 86 1.1973 -0.0987 1.1973 1.0942
No log 2.1463 88 0.8500 -0.1572 0.8500 0.9220
No log 2.1951 90 0.8898 -0.0195 0.8898 0.9433
No log 2.2439 92 0.8379 0.0028 0.8379 0.9154
No log 2.2927 94 0.8634 0.0421 0.8634 0.9292
No log 2.3415 96 0.9465 -0.0029 0.9465 0.9729
No log 2.3902 98 0.9026 -0.0630 0.9026 0.9501
No log 2.4390 100 0.9076 0.1277 0.9076 0.9527
No log 2.4878 102 0.8851 0.0608 0.8851 0.9408
No log 2.5366 104 0.9698 -0.0068 0.9698 0.9848
No log 2.5854 106 0.8184 0.1962 0.8184 0.9046
No log 2.6341 108 0.8019 0.2443 0.8019 0.8955
No log 2.6829 110 0.8208 0.0586 0.8208 0.9060
No log 2.7317 112 0.8269 0.1181 0.8269 0.9093
No log 2.7805 114 0.8605 0.1179 0.8605 0.9276
No log 2.8293 116 0.8623 0.1539 0.8623 0.9286
No log 2.8780 118 0.8569 0.2181 0.8569 0.9257
No log 2.9268 120 0.8580 0.1304 0.8580 0.9263
No log 2.9756 122 0.8642 0.1187 0.8642 0.9297
No log 3.0244 124 0.8741 0.1251 0.8741 0.9349
No log 3.0732 126 0.8648 0.1255 0.8648 0.9300
No log 3.1220 128 0.9226 0.0696 0.9226 0.9605
No log 3.1707 130 0.8709 0.1426 0.8709 0.9332
No log 3.2195 132 0.9100 -0.0008 0.9100 0.9540
No log 3.2683 134 0.9828 -0.0823 0.9828 0.9914
No log 3.3171 136 0.7871 0.0357 0.7871 0.8872
No log 3.3659 138 0.8096 0.1827 0.8096 0.8998
No log 3.4146 140 0.8035 0.0791 0.8035 0.8964
No log 3.4634 142 1.0235 -0.0114 1.0235 1.0117
No log 3.5122 144 0.9234 -0.0799 0.9234 0.9610
No log 3.5610 146 0.7512 0.0922 0.7512 0.8667
No log 3.6098 148 0.7493 0.0375 0.7493 0.8656
No log 3.6585 150 0.8349 -0.1126 0.8349 0.9137
No log 3.7073 152 0.8292 -0.1126 0.8292 0.9106
No log 3.7561 154 0.8071 0.0828 0.8071 0.8984
No log 3.8049 156 0.8434 -0.0209 0.8434 0.9184
No log 3.8537 158 0.8614 -0.0209 0.8614 0.9281
No log 3.9024 160 0.8293 0.0670 0.8293 0.9106
No log 3.9512 162 0.7992 0.2275 0.7992 0.8940
No log 4.0 164 0.8045 0.1687 0.8045 0.8970
No log 4.0488 166 0.8266 0.0771 0.8266 0.9092
No log 4.0976 168 0.9008 -0.0251 0.9008 0.9491
No log 4.1463 170 0.9355 -0.0669 0.9355 0.9672
No log 4.1951 172 0.9795 0.0040 0.9795 0.9897
No log 4.2439 174 0.8256 0.0432 0.8256 0.9086
No log 4.2927 176 0.8150 0.0028 0.8150 0.9028
No log 4.3415 178 0.8110 -0.0195 0.8110 0.9005
No log 4.3902 180 0.8128 -0.0195 0.8128 0.9016
No log 4.4390 182 0.8280 -0.1131 0.8280 0.9100
No log 4.4878 184 0.8677 -0.0699 0.8677 0.9315
No log 4.5366 186 0.8381 -0.1006 0.8381 0.9155
No log 4.5854 188 0.7999 -0.0027 0.7999 0.8944
No log 4.6341 190 0.7680 -0.0628 0.7680 0.8763
No log 4.6829 192 0.8200 0.0017 0.8200 0.9056
No log 4.7317 194 0.7852 -0.0778 0.7852 0.8861
No log 4.7805 196 0.7541 0.1029 0.7541 0.8684
No log 4.8293 198 0.7754 0.1029 0.7754 0.8806
No log 4.8780 200 0.8006 0.0068 0.8006 0.8948
No log 4.9268 202 1.1292 -0.0331 1.1292 1.0627
No log 4.9756 204 1.0813 0.0353 1.0813 1.0399
No log 5.0244 206 0.8201 0.0670 0.8201 0.9056
No log 5.0732 208 0.8966 0.1685 0.8966 0.9469
No log 5.1220 210 0.8490 0.0123 0.8490 0.9214
No log 5.1707 212 0.9070 0.0040 0.9070 0.9524
No log 5.2195 214 0.9839 -0.0122 0.9839 0.9919
No log 5.2683 216 0.8438 -0.0295 0.8438 0.9186
No log 5.3171 218 0.7627 0.0375 0.7627 0.8733
No log 5.3659 220 0.7736 0.0191 0.7736 0.8795
No log 5.4146 222 0.8697 0.0316 0.8697 0.9326
No log 5.4634 224 0.8283 -0.0390 0.8283 0.9101
No log 5.5122 226 0.7523 0.1379 0.7523 0.8673
No log 5.5610 228 0.7854 -0.0264 0.7854 0.8862
No log 5.6098 230 0.8401 0.0043 0.8401 0.9165
No log 5.6585 232 0.8277 -0.0351 0.8277 0.9098
No log 5.7073 234 0.7785 0.1404 0.7785 0.8823
No log 5.7561 236 0.7742 0.1404 0.7742 0.8799
No log 5.8049 238 0.7533 0.1379 0.7533 0.8679
No log 5.8537 240 0.9202 -0.0425 0.9202 0.9593
No log 5.9024 242 0.8137 0.0099 0.8137 0.9021
No log 5.9512 244 0.6901 0.0914 0.6901 0.8307
No log 6.0 246 0.7256 -0.1067 0.7256 0.8518
No log 6.0488 248 0.7174 -0.0033 0.7174 0.8470
No log 6.0976 250 0.7651 0.0914 0.7651 0.8747
No log 6.1463 252 0.9623 -0.0837 0.9623 0.9809
No log 6.1951 254 0.9777 -0.0097 0.9777 0.9888
No log 6.2439 256 0.9278 -0.0806 0.9278 0.9632
No log 6.2927 258 0.7290 0.0914 0.7290 0.8538
No log 6.3415 260 0.8008 0.1036 0.8008 0.8949
No log 6.3902 262 0.7641 0.1408 0.7641 0.8741
No log 6.4390 264 0.7048 0.0414 0.7048 0.8395
No log 6.4878 266 0.7518 0.0588 0.7518 0.8671
No log 6.5366 268 0.7285 0.1371 0.7285 0.8535
No log 6.5854 270 0.7296 0.1354 0.7296 0.8542
No log 6.6341 272 0.7439 0.0488 0.7439 0.8625
No log 6.6829 274 0.7624 0.1371 0.7624 0.8732
No log 6.7317 276 0.9019 -0.0424 0.9019 0.9497
No log 6.7805 278 0.8479 0.0490 0.8479 0.9208
No log 6.8293 280 0.8149 0.1778 0.8149 0.9027
No log 6.8780 282 0.8629 0.0581 0.8629 0.9289
No log 6.9268 284 0.8401 -0.0699 0.8401 0.9165
No log 6.9756 286 0.8585 -0.0283 0.8585 0.9265
No log 7.0244 288 0.8391 -0.0699 0.8391 0.9160
No log 7.0732 290 0.8091 0.0821 0.8091 0.8995
No log 7.1220 292 0.7999 0.0981 0.7999 0.8944
No log 7.1707 294 0.7605 0.0863 0.7605 0.8721
No log 7.2195 296 0.8355 -0.0355 0.8355 0.9140
No log 7.2683 298 0.8490 -0.0788 0.8490 0.9214
No log 7.3171 300 0.8270 -0.0240 0.8270 0.9094
No log 7.3659 302 0.8399 -0.0699 0.8399 0.9164
No log 7.4146 304 0.8156 0.0257 0.8156 0.9031
No log 7.4634 306 0.8777 -0.0425 0.8777 0.9368
No log 7.5122 308 0.8964 -0.0504 0.8964 0.9468
No log 7.5610 310 0.7832 0.0257 0.7832 0.8850
No log 7.6098 312 0.7584 0.0 0.7584 0.8709
No log 7.6585 314 0.7573 0.0 0.7573 0.8702
No log 7.7073 316 0.7555 -0.0195 0.7555 0.8692
No log 7.7561 318 0.8237 0.0409 0.8237 0.9076
No log 7.8049 320 0.7582 0.1148 0.7582 0.8707
No log 7.8537 322 0.7586 -0.0560 0.7586 0.8710
No log 7.9024 324 0.7967 -0.0406 0.7967 0.8926
No log 7.9512 326 0.7532 -0.0595 0.7532 0.8679
No log 8.0 328 0.8576 0.0442 0.8576 0.9261
No log 8.0488 330 0.9990 0.0609 0.9990 0.9995
No log 8.0976 332 0.9016 -0.0735 0.9016 0.9495
No log 8.1463 334 0.8569 0.0749 0.8569 0.9257
No log 8.1951 336 0.8613 -0.0218 0.8613 0.9281
No log 8.2439 338 1.0023 0.0250 1.0023 1.0011
No log 8.2927 340 1.1829 0.0026 1.1829 1.0876
No log 8.3415 342 1.0482 -0.0211 1.0482 1.0238
No log 8.3902 344 0.8791 -0.0767 0.8791 0.9376
No log 8.4390 346 0.8307 -0.0118 0.8307 0.9114
No log 8.4878 348 0.8904 0.0559 0.8904 0.9436
No log 8.5366 350 0.8575 0.1240 0.8575 0.9260
No log 8.5854 352 0.9066 -0.0408 0.9066 0.9522
No log 8.6341 354 1.0654 0.0576 1.0654 1.0322
No log 8.6829 356 1.0158 0.0576 1.0158 1.0079
No log 8.7317 358 0.8482 0.0442 0.8482 0.9210
No log 8.7805 360 0.7345 0.1254 0.7345 0.8570
No log 8.8293 362 0.7033 0.1371 0.7033 0.8386
No log 8.8780 364 0.6934 0.1371 0.6934 0.8327
No log 8.9268 366 0.7193 0.2258 0.7193 0.8481
No log 8.9756 368 0.8381 0.0476 0.8381 0.9155
No log 9.0244 370 0.8254 0.0476 0.8254 0.9085
No log 9.0732 372 0.7592 0.0723 0.7592 0.8713
No log 9.1220 374 0.7546 0.1311 0.7546 0.8687
No log 9.1707 376 0.7922 0.1047 0.7922 0.8900
No log 9.2195 378 0.8088 -0.0336 0.8088 0.8994
No log 9.2683 380 0.7887 0.0303 0.7887 0.8881
No log 9.3171 382 0.8012 -0.0262 0.8012 0.8951
No log 9.3659 384 0.8001 0.0123 0.8001 0.8945
No log 9.4146 386 0.7981 0.0600 0.7981 0.8934
No log 9.4634 388 0.8488 0.0017 0.8488 0.9213
No log 9.5122 390 0.8169 0.0476 0.8169 0.9038
No log 9.5610 392 0.7967 0.0549 0.7967 0.8926
No log 9.6098 394 0.7537 0.1259 0.7537 0.8681
No log 9.6585 396 0.7573 0.0869 0.7573 0.8702
No log 9.7073 398 0.7795 0.0869 0.7795 0.8829
No log 9.7561 400 0.8201 -0.0283 0.8201 0.9056
No log 9.8049 402 0.8369 -0.0723 0.8369 0.9148
No log 9.8537 404 0.9279 -0.0033 0.9279 0.9633
No log 9.9024 406 0.9015 -0.0033 0.9015 0.9495
No log 9.9512 408 0.8223 0.0123 0.8223 0.9068
No log 10.0 410 0.7849 0.0768 0.7849 0.8859
No log 10.0488 412 0.7873 -0.0252 0.7873 0.8873
No log 10.0976 414 0.8821 0.0755 0.8821 0.9392
No log 10.1463 416 0.8680 0.0392 0.8680 0.9317
No log 10.1951 418 0.7893 0.0247 0.7893 0.8884
No log 10.2439 420 0.7711 0.0303 0.7711 0.8781
No log 10.2927 422 0.7823 0.0303 0.7823 0.8845
No log 10.3415 424 0.8898 0.0017 0.8898 0.9433
No log 10.3902 426 1.0032 0.0545 1.0032 1.0016
No log 10.4390 428 0.9384 -0.0056 0.9384 0.9687
No log 10.4878 430 0.9087 0.0017 0.9087 0.9533
No log 10.5366 432 0.8689 -0.0336 0.8689 0.9321
No log 10.5854 434 0.8442 -0.0699 0.8442 0.9188
No log 10.6341 436 0.8847 -0.0322 0.8847 0.9406
No log 10.6829 438 1.1086 0.0486 1.1086 1.0529
No log 10.7317 440 1.2084 -0.0013 1.2084 1.0993
No log 10.7805 442 1.0705 0.0046 1.0705 1.0346
No log 10.8293 444 0.8795 -0.0054 0.8795 0.9378
No log 10.8780 446 0.7898 0.0123 0.7898 0.8887
No log 10.9268 448 0.7780 0.1096 0.7780 0.8820
No log 10.9756 450 0.7821 0.0041 0.7821 0.8844
No log 11.0244 452 0.7773 0.0490 0.7773 0.8817
No log 11.0732 454 0.7718 0.0600 0.7718 0.8785
No log 11.1220 456 0.7744 0.0690 0.7744 0.8800
No log 11.1707 458 0.7733 0.0690 0.7733 0.8794
No log 11.2195 460 0.7838 0.0442 0.7838 0.8853
No log 11.2683 462 0.7546 0.1565 0.7546 0.8687
No log 11.3171 464 0.7358 0.1202 0.7358 0.8578
No log 11.3659 466 0.7196 0.0338 0.7196 0.8483
No log 11.4146 468 0.7385 0.1202 0.7385 0.8594
No log 11.4634 470 0.8241 0.0442 0.8241 0.9078
No log 11.5122 472 0.8383 0.0409 0.8383 0.9156
No log 11.5610 474 0.7807 0.0588 0.7807 0.8836
No log 11.6098 476 0.7902 0.0549 0.7902 0.8889
No log 11.6585 478 0.7962 0.0236 0.7962 0.8923
No log 11.7073 480 0.8105 -0.0283 0.8105 0.9003
No log 11.7561 482 0.7942 0.0236 0.7942 0.8912
No log 11.8049 484 0.8069 -0.0336 0.8069 0.8983
No log 11.8537 486 0.8023 0.0043 0.8023 0.8957
No log 11.9024 488 0.7598 0.0670 0.7598 0.8717
No log 11.9512 490 0.7789 0.0670 0.7789 0.8825
No log 12.0 492 0.8931 -0.0079 0.8931 0.9450
No log 12.0488 494 0.9392 0.0260 0.9392 0.9691
No log 12.0976 496 0.9224 0.0287 0.9224 0.9604
No log 12.1463 498 0.8151 -0.1140 0.8151 0.9028
0.2695 12.1951 500 0.8030 0.0338 0.8030 0.8961
0.2695 12.2439 502 0.8030 0.0776 0.8030 0.8961
0.2695 12.2927 504 0.8875 -0.0425 0.8875 0.9421
0.2695 12.3415 506 1.0157 0.0609 1.0157 1.0078
0.2695 12.3902 508 1.0132 0.0609 1.0132 1.0066
0.2695 12.4390 510 0.9132 0.0287 0.9132 0.9556
0.2695 12.4878 512 0.7729 0.0549 0.7729 0.8792
0.2695 12.5366 514 0.7287 -0.0578 0.7287 0.8536
0.2695 12.5854 516 0.7465 -0.0541 0.7465 0.8640
0.2695 12.6341 518 0.7126 -0.0560 0.7126 0.8442
0.2695 12.6829 520 0.7241 0.2180 0.7241 0.8509
0.2695 12.7317 522 0.7811 0.0442 0.7811 0.8838
0.2695 12.7805 524 0.7633 0.0476 0.7633 0.8737
0.2695 12.8293 526 0.7144 0.0814 0.7144 0.8452
0.2695 12.8780 528 0.7089 0.0395 0.7089 0.8419
0.2695 12.9268 530 0.7069 0.0355 0.7069 0.8408
0.2695 12.9756 532 0.7340 0.1758 0.7340 0.8567
0.2695 13.0244 534 0.7544 0.1254 0.7544 0.8685
0.2695 13.0732 536 0.8050 0.0509 0.8050 0.8972
0.2695 13.1220 538 0.8095 0.0944 0.8095 0.8997
0.2695 13.1707 540 0.7812 0.0776 0.7812 0.8839
0.2695 13.2195 542 0.9261 -0.0056 0.9261 0.9624
0.2695 13.2683 544 1.1765 -0.0385 1.1765 1.0847
0.2695 13.3171 546 1.1620 -0.0385 1.1620 1.0779
0.2695 13.3659 548 0.9521 0.0287 0.9521 0.9757
0.2695 13.4146 550 0.8158 -0.0686 0.8158 0.9032

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k8_task3_organization

Finetuned
(4019)
this model