ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run2_AugV5_k7_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0657
  • Qwk: 0.5355
  • Mse: 1.0657
  • Rmse: 1.0323

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0571 2 5.3430 -0.0131 5.3430 2.3115
No log 0.1143 4 3.1272 0.0649 3.1272 1.7684
No log 0.1714 6 2.6556 -0.1003 2.6556 1.6296
No log 0.2286 8 2.4195 -0.1156 2.4195 1.5555
No log 0.2857 10 1.7140 0.0267 1.7140 1.3092
No log 0.3429 12 1.2703 0.1423 1.2703 1.1271
No log 0.4 14 1.2489 0.2203 1.2489 1.1175
No log 0.4571 16 1.1863 0.2751 1.1863 1.0892
No log 0.5143 18 1.2975 0.1746 1.2975 1.1391
No log 0.5714 20 1.3932 0.1370 1.3932 1.1803
No log 0.6286 22 1.3680 0.1204 1.3680 1.1696
No log 0.6857 24 1.2413 0.3295 1.2413 1.1141
No log 0.7429 26 1.1591 0.3779 1.1591 1.0766
No log 0.8 28 1.2529 0.2372 1.2529 1.1193
No log 0.8571 30 1.1474 0.3339 1.1474 1.0712
No log 0.9143 32 1.1505 0.3121 1.1505 1.0726
No log 0.9714 34 1.1191 0.3563 1.1191 1.0579
No log 1.0286 36 1.1045 0.3838 1.1045 1.0510
No log 1.0857 38 1.0767 0.3972 1.0767 1.0376
No log 1.1429 40 1.0334 0.3862 1.0334 1.0166
No log 1.2 42 0.8723 0.4907 0.8723 0.9340
No log 1.2571 44 0.8736 0.5391 0.8736 0.9347
No log 1.3143 46 0.8995 0.5487 0.8995 0.9484
No log 1.3714 48 0.9249 0.5679 0.9249 0.9617
No log 1.4286 50 0.9743 0.5092 0.9743 0.9871
No log 1.4857 52 0.9736 0.4893 0.9736 0.9867
No log 1.5429 54 0.9967 0.4736 0.9967 0.9984
No log 1.6 56 0.9438 0.5147 0.9438 0.9715
No log 1.6571 58 1.0154 0.5401 1.0154 1.0077
No log 1.7143 60 0.9127 0.5038 0.9127 0.9554
No log 1.7714 62 1.0786 0.5033 1.0786 1.0385
No log 1.8286 64 1.0134 0.5134 1.0134 1.0067
No log 1.8857 66 0.8451 0.5707 0.8451 0.9193
No log 1.9429 68 0.9047 0.5867 0.9047 0.9512
No log 2.0 70 0.8410 0.6429 0.8410 0.9171
No log 2.0571 72 0.8599 0.5858 0.8599 0.9273
No log 2.1143 74 1.0333 0.5201 1.0333 1.0165
No log 2.1714 76 0.9665 0.5218 0.9665 0.9831
No log 2.2286 78 0.8244 0.5851 0.8244 0.9080
No log 2.2857 80 0.8443 0.5752 0.8443 0.9189
No log 2.3429 82 0.8400 0.5657 0.8400 0.9165
No log 2.4 84 0.9719 0.5372 0.9719 0.9858
No log 2.4571 86 0.9467 0.5528 0.9467 0.9730
No log 2.5143 88 0.7803 0.6361 0.7803 0.8834
No log 2.5714 90 0.7549 0.6035 0.7549 0.8688
No log 2.6286 92 0.8418 0.5549 0.8418 0.9175
No log 2.6857 94 0.7118 0.6245 0.7118 0.8437
No log 2.7429 96 0.9979 0.5331 0.9979 0.9989
No log 2.8 98 1.3117 0.4883 1.3117 1.1453
No log 2.8571 100 1.0906 0.5105 1.0906 1.0443
No log 2.9143 102 0.6861 0.6388 0.6861 0.8283
No log 2.9714 104 0.7704 0.5976 0.7704 0.8777
No log 3.0286 106 0.7716 0.6420 0.7716 0.8784
No log 3.0857 108 0.6576 0.6805 0.6576 0.8109
No log 3.1429 110 0.8588 0.5971 0.8588 0.9267
No log 3.2 112 1.0898 0.5213 1.0898 1.0439
No log 3.2571 114 1.1132 0.5205 1.1132 1.0551
No log 3.3143 116 1.0380 0.5333 1.0380 1.0188
No log 3.3714 118 1.0357 0.5401 1.0357 1.0177
No log 3.4286 120 0.9567 0.6132 0.9567 0.9781
No log 3.4857 122 1.0190 0.5522 1.0190 1.0095
No log 3.5429 124 1.2592 0.4356 1.2592 1.1222
No log 3.6 126 1.1839 0.4543 1.1839 1.0881
No log 3.6571 128 1.0883 0.4830 1.0883 1.0432
No log 3.7143 130 1.0148 0.5338 1.0148 1.0074
No log 3.7714 132 0.7750 0.6389 0.7750 0.8803
No log 3.8286 134 0.7688 0.6161 0.7688 0.8768
No log 3.8857 136 0.7777 0.6601 0.7777 0.8819
No log 3.9429 138 0.9604 0.6227 0.9604 0.9800
No log 4.0 140 1.2622 0.5311 1.2622 1.1235
No log 4.0571 142 1.2558 0.5144 1.2558 1.1206
No log 4.1143 144 1.0652 0.5312 1.0652 1.0321
No log 4.1714 146 0.8674 0.6355 0.8674 0.9313
No log 4.2286 148 0.7753 0.6502 0.7753 0.8805
No log 4.2857 150 0.8393 0.6358 0.8393 0.9161
No log 4.3429 152 0.8959 0.6019 0.8959 0.9465
No log 4.4 154 0.8972 0.5988 0.8972 0.9472
No log 4.4571 156 0.7496 0.6873 0.7496 0.8658
No log 4.5143 158 0.7267 0.6988 0.7267 0.8524
No log 4.5714 160 0.7081 0.6854 0.7081 0.8415
No log 4.6286 162 0.7194 0.6824 0.7194 0.8481
No log 4.6857 164 0.8391 0.6738 0.8391 0.9160
No log 4.7429 166 1.1286 0.5640 1.1286 1.0623
No log 4.8 168 1.0770 0.5831 1.0770 1.0378
No log 4.8571 170 0.8046 0.6916 0.8046 0.8970
No log 4.9143 172 0.7048 0.6692 0.7048 0.8395
No log 4.9714 174 0.7162 0.6755 0.7162 0.8463
No log 5.0286 176 0.7766 0.6634 0.7766 0.8812
No log 5.0857 178 0.9326 0.6175 0.9326 0.9657
No log 5.1429 180 1.0926 0.5442 1.0926 1.0453
No log 5.2 182 0.9450 0.6135 0.9450 0.9721
No log 5.2571 184 0.7525 0.6745 0.7525 0.8674
No log 5.3143 186 0.7507 0.6302 0.7507 0.8664
No log 5.3714 188 0.7302 0.6167 0.7302 0.8545
No log 5.4286 190 0.7342 0.6394 0.7342 0.8568
No log 5.4857 192 0.8565 0.6562 0.8565 0.9255
No log 5.5429 194 0.9119 0.5911 0.9119 0.9550
No log 5.6 196 0.9576 0.5911 0.9576 0.9786
No log 5.6571 198 0.8575 0.6347 0.8575 0.9260
No log 5.7143 200 0.7727 0.6947 0.7727 0.8790
No log 5.7714 202 0.7947 0.6674 0.7947 0.8915
No log 5.8286 204 0.9209 0.6373 0.9209 0.9597
No log 5.8857 206 1.0566 0.6122 1.0566 1.0279
No log 5.9429 208 0.9640 0.6463 0.9640 0.9819
No log 6.0 210 0.8122 0.6719 0.8122 0.9012
No log 6.0571 212 0.7881 0.6924 0.7881 0.8877
No log 6.1143 214 0.8804 0.6246 0.8804 0.9383
No log 6.1714 216 0.9973 0.5694 0.9973 0.9986
No log 6.2286 218 0.9968 0.5608 0.9968 0.9984
No log 6.2857 220 0.8269 0.6476 0.8269 0.9093
No log 6.3429 222 0.7584 0.6718 0.7584 0.8708
No log 6.4 224 0.6644 0.6765 0.6644 0.8151
No log 6.4571 226 0.6248 0.6804 0.6248 0.7905
No log 6.5143 228 0.7130 0.6969 0.7130 0.8444
No log 6.5714 230 0.9473 0.6137 0.9473 0.9733
No log 6.6286 232 0.9530 0.6280 0.9530 0.9762
No log 6.6857 234 0.7536 0.7182 0.7536 0.8681
No log 6.7429 236 0.7443 0.7166 0.7443 0.8627
No log 6.8 238 0.8401 0.6413 0.8401 0.9166
No log 6.8571 240 0.9515 0.6228 0.9515 0.9754
No log 6.9143 242 0.8301 0.6694 0.8301 0.9111
No log 6.9714 244 0.7428 0.6982 0.7428 0.8619
No log 7.0286 246 0.8660 0.6178 0.8660 0.9306
No log 7.0857 248 0.9069 0.5964 0.9069 0.9523
No log 7.1429 250 0.9427 0.6036 0.9427 0.9709
No log 7.2 252 0.8597 0.6288 0.8597 0.9272
No log 7.2571 254 0.8272 0.6288 0.8272 0.9095
No log 7.3143 256 0.9297 0.6086 0.9297 0.9642
No log 7.3714 258 0.8701 0.6293 0.8701 0.9328
No log 7.4286 260 0.7681 0.6944 0.7681 0.8764
No log 7.4857 262 0.7007 0.7263 0.7007 0.8371
No log 7.5429 264 0.6556 0.7099 0.6556 0.8097
No log 7.6 266 0.6917 0.7006 0.6917 0.8317
No log 7.6571 268 0.8437 0.6470 0.8437 0.9185
No log 7.7143 270 0.8517 0.6517 0.8517 0.9229
No log 7.7714 272 0.7675 0.7046 0.7675 0.8761
No log 7.8286 274 0.7646 0.7054 0.7646 0.8744
No log 7.8857 276 0.7847 0.6628 0.7847 0.8858
No log 7.9429 278 0.8038 0.6628 0.8038 0.8965
No log 8.0 280 0.7983 0.6527 0.7983 0.8935
No log 8.0571 282 0.7692 0.6788 0.7692 0.8770
No log 8.1143 284 0.7306 0.6965 0.7306 0.8548
No log 8.1714 286 0.7067 0.6590 0.7067 0.8406
No log 8.2286 288 0.7578 0.6885 0.7578 0.8705
No log 8.2857 290 0.8385 0.6512 0.8385 0.9157
No log 8.3429 292 0.9729 0.5842 0.9729 0.9864
No log 8.4 294 0.8527 0.6420 0.8527 0.9234
No log 8.4571 296 0.7301 0.6967 0.7301 0.8545
No log 8.5143 298 0.6768 0.6643 0.6768 0.8227
No log 8.5714 300 0.6640 0.6516 0.6640 0.8149
No log 8.6286 302 0.6898 0.6664 0.6898 0.8306
No log 8.6857 304 0.8909 0.6159 0.8909 0.9439
No log 8.7429 306 1.1835 0.4966 1.1835 1.0879
No log 8.8 308 1.2569 0.4960 1.2569 1.1211
No log 8.8571 310 1.1361 0.5189 1.1361 1.0659
No log 8.9143 312 1.0565 0.5399 1.0565 1.0279
No log 8.9714 314 0.8676 0.6551 0.8676 0.9315
No log 9.0286 316 0.7602 0.6899 0.7602 0.8719
No log 9.0857 318 0.7732 0.6707 0.7732 0.8793
No log 9.1429 320 0.8802 0.5899 0.8802 0.9382
No log 9.2 322 1.0857 0.5209 1.0857 1.0419
No log 9.2571 324 1.1553 0.5265 1.1553 1.0748
No log 9.3143 326 1.0771 0.5189 1.0771 1.0378
No log 9.3714 328 1.0802 0.5209 1.0802 1.0393
No log 9.4286 330 0.9214 0.5783 0.9214 0.9599
No log 9.4857 332 0.8058 0.6211 0.8058 0.8977
No log 9.5429 334 0.7409 0.6878 0.7409 0.8607
No log 9.6 336 0.7973 0.6509 0.7973 0.8929
No log 9.6571 338 0.9476 0.5867 0.9476 0.9735
No log 9.7143 340 0.8953 0.5806 0.8953 0.9462
No log 9.7714 342 0.7900 0.6858 0.7900 0.8888
No log 9.8286 344 0.7239 0.6774 0.7239 0.8508
No log 9.8857 346 0.7151 0.6791 0.7151 0.8457
No log 9.9429 348 0.7722 0.6819 0.7722 0.8787
No log 10.0 350 0.9402 0.6286 0.9402 0.9696
No log 10.0571 352 1.0030 0.6179 1.0030 1.0015
No log 10.1143 354 0.8716 0.6648 0.8716 0.9336
No log 10.1714 356 0.7326 0.7014 0.7326 0.8559
No log 10.2286 358 0.7226 0.6905 0.7226 0.8501
No log 10.2857 360 0.7907 0.6880 0.7907 0.8892
No log 10.3429 362 0.9463 0.5667 0.9463 0.9728
No log 10.4 364 1.0349 0.5381 1.0349 1.0173
No log 10.4571 366 0.9543 0.5932 0.9543 0.9769
No log 10.5143 368 0.8436 0.6396 0.8436 0.9185
No log 10.5714 370 0.8043 0.6596 0.8043 0.8968
No log 10.6286 372 0.7737 0.7008 0.7737 0.8796
No log 10.6857 374 0.8094 0.6618 0.8094 0.8996
No log 10.7429 376 0.7938 0.6666 0.7938 0.8909
No log 10.8 378 0.7080 0.7159 0.7080 0.8414
No log 10.8571 380 0.6661 0.6835 0.6661 0.8162
No log 10.9143 382 0.6601 0.6835 0.6601 0.8125
No log 10.9714 384 0.6558 0.6619 0.6558 0.8098
No log 11.0286 386 0.6875 0.7120 0.6875 0.8291
No log 11.0857 388 0.7488 0.6509 0.7488 0.8653
No log 11.1429 390 0.8062 0.6638 0.8062 0.8979
No log 11.2 392 0.7661 0.6570 0.7661 0.8753
No log 11.2571 394 0.7196 0.7086 0.7196 0.8483
No log 11.3143 396 0.7788 0.6609 0.7788 0.8825
No log 11.3714 398 0.8422 0.6489 0.8422 0.9177
No log 11.4286 400 0.9245 0.6167 0.9245 0.9615
No log 11.4857 402 0.8394 0.6395 0.8394 0.9162
No log 11.5429 404 0.8095 0.6505 0.8095 0.8997
No log 11.6 406 0.7775 0.6481 0.7775 0.8818
No log 11.6571 408 0.8326 0.6582 0.8326 0.9125
No log 11.7143 410 0.7396 0.7004 0.7396 0.8600
No log 11.7714 412 0.6857 0.7117 0.6857 0.8280
No log 11.8286 414 0.7590 0.7058 0.7590 0.8712
No log 11.8857 416 0.9415 0.6233 0.9415 0.9703
No log 11.9429 418 1.0937 0.6062 1.0937 1.0458
No log 12.0 420 1.0981 0.5797 1.0981 1.0479
No log 12.0571 422 1.0543 0.5637 1.0543 1.0268
No log 12.1143 424 1.0139 0.5377 1.0139 1.0069
No log 12.1714 426 0.9686 0.5629 0.9686 0.9842
No log 12.2286 428 1.0128 0.5138 1.0128 1.0064
No log 12.2857 430 1.0498 0.5128 1.0498 1.0246
No log 12.3429 432 0.8833 0.6268 0.8833 0.9398
No log 12.4 434 0.7097 0.6949 0.7097 0.8424
No log 12.4571 436 0.6709 0.6879 0.6709 0.8191
No log 12.5143 438 0.6722 0.6978 0.6722 0.8198
No log 12.5714 440 0.7612 0.7090 0.7612 0.8725
No log 12.6286 442 0.9488 0.6221 0.9488 0.9741
No log 12.6857 444 0.9146 0.6440 0.9146 0.9563
No log 12.7429 446 0.7522 0.6964 0.7522 0.8673
No log 12.8 448 0.6023 0.6598 0.6023 0.7761
No log 12.8571 450 0.6168 0.6413 0.6168 0.7854
No log 12.9143 452 0.6096 0.6681 0.6096 0.7808
No log 12.9714 454 0.6104 0.6865 0.6104 0.7813
No log 13.0286 456 0.6612 0.7128 0.6612 0.8131
No log 13.0857 458 0.6447 0.7036 0.6447 0.8029
No log 13.1429 460 0.6339 0.7079 0.6339 0.7962
No log 13.2 462 0.6410 0.6216 0.6410 0.8006
No log 13.2571 464 0.6306 0.6425 0.6306 0.7941
No log 13.3143 466 0.6213 0.6835 0.6213 0.7882
No log 13.3714 468 0.6497 0.7204 0.6497 0.8061
No log 13.4286 470 0.7284 0.6667 0.7284 0.8535
No log 13.4857 472 0.7837 0.6530 0.7837 0.8853
No log 13.5429 474 0.7644 0.6530 0.7644 0.8743
No log 13.6 476 0.7373 0.6657 0.7373 0.8586
No log 13.6571 478 0.7246 0.6857 0.7246 0.8512
No log 13.7143 480 0.6883 0.6970 0.6883 0.8297
No log 13.7714 482 0.6774 0.7007 0.6774 0.8230
No log 13.8286 484 0.7164 0.7116 0.7164 0.8464
No log 13.8857 486 0.7260 0.7116 0.7260 0.8521
No log 13.9429 488 0.7837 0.6726 0.7837 0.8852
No log 14.0 490 0.8038 0.6715 0.8038 0.8965
No log 14.0571 492 0.8259 0.6629 0.8259 0.9088
No log 14.1143 494 0.7970 0.6973 0.7970 0.8928
No log 14.1714 496 0.7024 0.7072 0.7024 0.8381
No log 14.2286 498 0.7095 0.6971 0.7095 0.8423
0.3637 14.2857 500 0.7234 0.6954 0.7234 0.8505
0.3637 14.3429 502 0.7055 0.7079 0.7055 0.8399
0.3637 14.4 504 0.7237 0.7030 0.7237 0.8507
0.3637 14.4571 506 0.7712 0.6960 0.7712 0.8782
0.3637 14.5143 508 0.7300 0.6960 0.7300 0.8544
0.3637 14.5714 510 0.6766 0.7263 0.6766 0.8226
0.3637 14.6286 512 0.6835 0.7247 0.6835 0.8268
0.3637 14.6857 514 0.7758 0.6851 0.7758 0.8808
0.3637 14.7429 516 0.8551 0.6773 0.8551 0.9247
0.3637 14.8 518 0.8200 0.6915 0.8200 0.9055
0.3637 14.8571 520 0.7410 0.6865 0.7410 0.8608
0.3637 14.9143 522 0.6627 0.6997 0.6627 0.8140
0.3637 14.9714 524 0.6722 0.7049 0.6722 0.8199
0.3637 15.0286 526 0.7373 0.6982 0.7373 0.8587
0.3637 15.0857 528 0.8253 0.7017 0.8253 0.9085
0.3637 15.1429 530 0.8180 0.6814 0.8180 0.9044
0.3637 15.2 532 0.7457 0.6816 0.7457 0.8636
0.3637 15.2571 534 0.7131 0.7012 0.7131 0.8445
0.3637 15.3143 536 0.7075 0.7152 0.7075 0.8411
0.3637 15.3714 538 0.6597 0.7086 0.6597 0.8122
0.3637 15.4286 540 0.6856 0.7224 0.6856 0.8280
0.3637 15.4857 542 0.7557 0.6901 0.7557 0.8693
0.3637 15.5429 544 0.9779 0.6161 0.9779 0.9889
0.3637 15.6 546 1.1443 0.5428 1.1443 1.0697
0.3637 15.6571 548 1.1605 0.5411 1.1605 1.0773
0.3637 15.7143 550 1.0657 0.5355 1.0657 1.0323

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run2_AugV5_k7_task1_organization

Finetuned
(4023)
this model