ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k7_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6290
  • Qwk: 0.5315
  • Mse: 0.6290
  • Rmse: 0.7931

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0833 2 4.0787 0.0130 4.0787 2.0196
No log 0.1667 4 2.5344 -0.0274 2.5344 1.5920
No log 0.25 6 1.4940 0.0460 1.4940 1.2223
No log 0.3333 8 1.1651 0.1901 1.1651 1.0794
No log 0.4167 10 1.0271 0.2239 1.0271 1.0134
No log 0.5 12 1.0403 0.1263 1.0403 1.0199
No log 0.5833 14 0.9673 0.2770 0.9673 0.9835
No log 0.6667 16 1.1427 0.1618 1.1427 1.0690
No log 0.75 18 1.7266 0.1450 1.7266 1.3140
No log 0.8333 20 1.8829 0.1039 1.8829 1.3722
No log 0.9167 22 1.6161 0.1084 1.6161 1.2713
No log 1.0 24 1.2964 0.0627 1.2964 1.1386
No log 1.0833 26 1.1049 0.2094 1.1049 1.0512
No log 1.1667 28 1.0236 0.2640 1.0236 1.0117
No log 1.25 30 0.9542 0.3340 0.9542 0.9768
No log 1.3333 32 0.9419 0.3340 0.9419 0.9705
No log 1.4167 34 1.0354 0.2752 1.0354 1.0175
No log 1.5 36 1.0456 0.3110 1.0456 1.0226
No log 1.5833 38 0.8864 0.4042 0.8864 0.9415
No log 1.6667 40 0.8439 0.4024 0.8439 0.9186
No log 1.75 42 0.8664 0.3571 0.8664 0.9308
No log 1.8333 44 0.7751 0.4489 0.7751 0.8804
No log 1.9167 46 0.8253 0.4591 0.8253 0.9084
No log 2.0 48 0.9717 0.2523 0.9717 0.9858
No log 2.0833 50 0.8308 0.4715 0.8308 0.9115
No log 2.1667 52 0.8189 0.3719 0.8189 0.9049
No log 2.25 54 0.9263 0.4541 0.9263 0.9624
No log 2.3333 56 0.9162 0.4056 0.9162 0.9572
No log 2.4167 58 0.7710 0.4416 0.7710 0.8780
No log 2.5 60 0.9178 0.4341 0.9178 0.9580
No log 2.5833 62 0.9077 0.4341 0.9077 0.9528
No log 2.6667 64 0.7305 0.4903 0.7305 0.8547
No log 2.75 66 0.8584 0.3590 0.8584 0.9265
No log 2.8333 68 1.0132 0.3433 1.0132 1.0066
No log 2.9167 70 0.8499 0.3497 0.8499 0.9219
No log 3.0 72 0.6393 0.6046 0.6393 0.7996
No log 3.0833 74 0.7353 0.5342 0.7353 0.8575
No log 3.1667 76 0.6865 0.6081 0.6865 0.8286
No log 3.25 78 0.6899 0.5672 0.6899 0.8306
No log 3.3333 80 1.1352 0.4228 1.1352 1.0655
No log 3.4167 82 1.2227 0.3642 1.2227 1.1058
No log 3.5 84 0.9113 0.4425 0.9113 0.9546
No log 3.5833 86 0.6981 0.6067 0.6981 0.8355
No log 3.6667 88 0.6689 0.5682 0.6689 0.8179
No log 3.75 90 0.6965 0.6188 0.6965 0.8345
No log 3.8333 92 0.6806 0.6196 0.6806 0.8250
No log 3.9167 94 0.6863 0.5474 0.6863 0.8285
No log 4.0 96 0.6539 0.6275 0.6539 0.8086
No log 4.0833 98 0.6603 0.6087 0.6603 0.8126
No log 4.1667 100 0.6594 0.6520 0.6594 0.8120
No log 4.25 102 0.6948 0.5416 0.6948 0.8336
No log 4.3333 104 0.7291 0.5492 0.7291 0.8539
No log 4.4167 106 0.6684 0.6782 0.6684 0.8176
No log 4.5 108 0.6997 0.5934 0.6997 0.8365
No log 4.5833 110 0.7608 0.4914 0.7608 0.8723
No log 4.6667 112 0.8697 0.4982 0.8697 0.9326
No log 4.75 114 0.8055 0.5387 0.8055 0.8975
No log 4.8333 116 0.8496 0.5188 0.8496 0.9217
No log 4.9167 118 0.8474 0.5082 0.8474 0.9206
No log 5.0 120 0.6529 0.5849 0.6529 0.8080
No log 5.0833 122 0.6183 0.4841 0.6183 0.7863
No log 5.1667 124 0.7701 0.4947 0.7701 0.8775
No log 5.25 126 0.8023 0.5172 0.8023 0.8957
No log 5.3333 128 0.6545 0.5018 0.6545 0.8090
No log 5.4167 130 0.5844 0.7013 0.5844 0.7644
No log 5.5 132 0.7598 0.5713 0.7598 0.8717
No log 5.5833 134 0.6622 0.6560 0.6622 0.8138
No log 5.6667 136 0.6014 0.6646 0.6014 0.7755
No log 5.75 138 0.8653 0.5541 0.8653 0.9302
No log 5.8333 140 0.8845 0.5830 0.8845 0.9405
No log 5.9167 142 0.6925 0.5759 0.6925 0.8322
No log 6.0 144 0.5871 0.6067 0.5871 0.7662
No log 6.0833 146 0.5906 0.6470 0.5906 0.7685
No log 6.1667 148 0.6380 0.5841 0.6380 0.7987
No log 6.25 150 0.7508 0.5234 0.7508 0.8665
No log 6.3333 152 0.7047 0.5446 0.7047 0.8394
No log 6.4167 154 0.6253 0.5856 0.6253 0.7908
No log 6.5 156 0.6440 0.6315 0.6440 0.8025
No log 6.5833 158 0.6232 0.6432 0.6232 0.7894
No log 6.6667 160 0.6195 0.5759 0.6195 0.7871
No log 6.75 162 0.7385 0.5235 0.7385 0.8594
No log 6.8333 164 0.7121 0.5336 0.7121 0.8439
No log 6.9167 166 0.6064 0.6517 0.6064 0.7787
No log 7.0 168 0.6540 0.6179 0.6540 0.8087
No log 7.0833 170 0.7633 0.5668 0.7633 0.8737
No log 7.1667 172 0.7121 0.5486 0.7121 0.8439
No log 7.25 174 0.6276 0.6680 0.6276 0.7922
No log 7.3333 176 0.6230 0.6788 0.6230 0.7893
No log 7.4167 178 0.6386 0.6219 0.6386 0.7991
No log 7.5 180 0.6417 0.5373 0.6417 0.8011
No log 7.5833 182 0.6786 0.4984 0.6786 0.8238
No log 7.6667 184 0.8297 0.4508 0.8297 0.9109
No log 7.75 186 0.7592 0.4700 0.7592 0.8713
No log 7.8333 188 0.6235 0.6057 0.6235 0.7896
No log 7.9167 190 0.6767 0.5942 0.6767 0.8226
No log 8.0 192 0.6785 0.5677 0.6785 0.8237
No log 8.0833 194 0.6305 0.7184 0.6305 0.7940
No log 8.1667 196 0.6800 0.6341 0.6800 0.8246
No log 8.25 198 0.6452 0.7184 0.6452 0.8032
No log 8.3333 200 0.6514 0.6446 0.6514 0.8071
No log 8.4167 202 0.6621 0.5833 0.6621 0.8137
No log 8.5 204 0.6214 0.6165 0.6214 0.7883
No log 8.5833 206 0.6360 0.5069 0.6360 0.7975
No log 8.6667 208 0.7050 0.4738 0.7050 0.8396
No log 8.75 210 0.7117 0.4606 0.7117 0.8436
No log 8.8333 212 0.6827 0.4968 0.6827 0.8263
No log 8.9167 214 0.6446 0.5635 0.6446 0.8028
No log 9.0 216 0.6475 0.5565 0.6475 0.8047
No log 9.0833 218 0.6371 0.5678 0.6371 0.7982
No log 9.1667 220 0.6621 0.6246 0.6621 0.8137
No log 9.25 222 0.8015 0.5835 0.8015 0.8953
No log 9.3333 224 0.7862 0.5553 0.7862 0.8867
No log 9.4167 226 0.6703 0.5315 0.6703 0.8187
No log 9.5 228 0.6399 0.5386 0.6399 0.8000
No log 9.5833 230 0.6357 0.5302 0.6357 0.7973
No log 9.6667 232 0.6368 0.5302 0.6368 0.7980
No log 9.75 234 0.6275 0.6054 0.6275 0.7922
No log 9.8333 236 0.6222 0.6054 0.6222 0.7888
No log 9.9167 238 0.6165 0.6272 0.6165 0.7852
No log 10.0 240 0.6017 0.6616 0.6017 0.7757
No log 10.0833 242 0.6025 0.6396 0.6025 0.7762
No log 10.1667 244 0.5923 0.6311 0.5923 0.7696
No log 10.25 246 0.6045 0.6322 0.6045 0.7775
No log 10.3333 248 0.6200 0.5415 0.6200 0.7874
No log 10.4167 250 0.6228 0.5647 0.6228 0.7892
No log 10.5 252 0.6186 0.6241 0.6186 0.7865
No log 10.5833 254 0.6872 0.5666 0.6872 0.8290
No log 10.6667 256 0.6829 0.5560 0.6829 0.8264
No log 10.75 258 0.6422 0.6371 0.6422 0.8014
No log 10.8333 260 0.6335 0.6244 0.6335 0.7959
No log 10.9167 262 0.6492 0.5796 0.6492 0.8057
No log 11.0 264 0.6599 0.5808 0.6599 0.8123
No log 11.0833 266 0.6591 0.5582 0.6591 0.8118
No log 11.1667 268 0.6398 0.5582 0.6398 0.7999
No log 11.25 270 0.6175 0.6035 0.6175 0.7858
No log 11.3333 272 0.6157 0.6364 0.6157 0.7847
No log 11.4167 274 0.6062 0.6606 0.6062 0.7786
No log 11.5 276 0.6028 0.6903 0.6028 0.7764
No log 11.5833 278 0.6074 0.6364 0.6074 0.7794
No log 11.6667 280 0.6082 0.6186 0.6082 0.7799
No log 11.75 282 0.6058 0.6536 0.6058 0.7784
No log 11.8333 284 0.6257 0.7027 0.6257 0.7910
No log 11.9167 286 0.6312 0.7287 0.6312 0.7945
No log 12.0 288 0.6088 0.6350 0.6088 0.7802
No log 12.0833 290 0.6374 0.5439 0.6374 0.7984
No log 12.1667 292 0.6408 0.5439 0.6408 0.8005
No log 12.25 294 0.6208 0.6374 0.6208 0.7879
No log 12.3333 296 0.6498 0.6042 0.6498 0.8061
No log 12.4167 298 0.6578 0.6042 0.6578 0.8111
No log 12.5 300 0.6049 0.6835 0.6049 0.7778
No log 12.5833 302 0.6072 0.5879 0.6072 0.7792
No log 12.6667 304 0.5889 0.6725 0.5889 0.7674
No log 12.75 306 0.6119 0.6167 0.6119 0.7823
No log 12.8333 308 0.6230 0.6240 0.6230 0.7893
No log 12.9167 310 0.6462 0.6683 0.6462 0.8039
No log 13.0 312 0.6236 0.6529 0.6236 0.7897
No log 13.0833 314 0.5959 0.6365 0.5959 0.7719
No log 13.1667 316 0.6053 0.6242 0.6053 0.7780
No log 13.25 318 0.6201 0.5869 0.6201 0.7875
No log 13.3333 320 0.6285 0.6680 0.6285 0.7928
No log 13.4167 322 0.6664 0.6328 0.6664 0.8163
No log 13.5 324 0.6319 0.6177 0.6319 0.7950
No log 13.5833 326 0.6333 0.5409 0.6333 0.7958
No log 13.6667 328 0.6981 0.5873 0.6981 0.8355
No log 13.75 330 0.6789 0.5849 0.6789 0.8239
No log 13.8333 332 0.6194 0.5808 0.6194 0.7870
No log 13.9167 334 0.6401 0.6045 0.6401 0.8000
No log 14.0 336 0.6503 0.6091 0.6503 0.8064
No log 14.0833 338 0.6164 0.6813 0.6164 0.7851
No log 14.1667 340 0.6115 0.6325 0.6115 0.7820
No log 14.25 342 0.6183 0.6234 0.6183 0.7863
No log 14.3333 344 0.5959 0.6259 0.5959 0.7719
No log 14.4167 346 0.5842 0.6470 0.5842 0.7643
No log 14.5 348 0.5835 0.6753 0.5835 0.7639
No log 14.5833 350 0.5792 0.6753 0.5792 0.7610
No log 14.6667 352 0.5689 0.6470 0.5689 0.7543
No log 14.75 354 0.5996 0.5759 0.5996 0.7744
No log 14.8333 356 0.6141 0.5759 0.6141 0.7836
No log 14.9167 358 0.5786 0.6292 0.5786 0.7606
No log 15.0 360 0.5945 0.6252 0.5945 0.7710
No log 15.0833 362 0.6537 0.5822 0.6537 0.8085
No log 15.1667 364 0.6264 0.5975 0.6264 0.7914
No log 15.25 366 0.5821 0.6339 0.5821 0.7629
No log 15.3333 368 0.6188 0.6559 0.6188 0.7867
No log 15.4167 370 0.7057 0.6048 0.7057 0.8401
No log 15.5 372 0.6746 0.6108 0.6746 0.8213
No log 15.5833 374 0.6331 0.6386 0.6331 0.7957
No log 15.6667 376 0.5927 0.6339 0.5927 0.7699
No log 15.75 378 0.5989 0.6397 0.5989 0.7739
No log 15.8333 380 0.5983 0.5847 0.5983 0.7735
No log 15.9167 382 0.6147 0.5783 0.6147 0.7840
No log 16.0 384 0.6480 0.5337 0.6480 0.8050
No log 16.0833 386 0.6591 0.5451 0.6591 0.8119
No log 16.1667 388 0.6337 0.6214 0.6337 0.7961
No log 16.25 390 0.5960 0.6869 0.5960 0.7720
No log 16.3333 392 0.5984 0.6319 0.5984 0.7735
No log 16.4167 394 0.6023 0.6493 0.6023 0.7761
No log 16.5 396 0.5973 0.6319 0.5973 0.7729
No log 16.5833 398 0.5998 0.6430 0.5998 0.7745
No log 16.6667 400 0.5979 0.6745 0.5979 0.7732
No log 16.75 402 0.5965 0.6649 0.5965 0.7723
No log 16.8333 404 0.6014 0.6397 0.6014 0.7755
No log 16.9167 406 0.6031 0.6397 0.6031 0.7766
No log 17.0 408 0.5979 0.6249 0.5979 0.7732
No log 17.0833 410 0.6009 0.6349 0.6009 0.7752
No log 17.1667 412 0.5932 0.6535 0.5932 0.7702
No log 17.25 414 0.5812 0.6349 0.5812 0.7623
No log 17.3333 416 0.5724 0.6681 0.5724 0.7566
No log 17.4167 418 0.5882 0.6815 0.5882 0.7669
No log 17.5 420 0.6005 0.6500 0.6005 0.7749
No log 17.5833 422 0.6022 0.6500 0.6022 0.7760
No log 17.6667 424 0.5841 0.6508 0.5841 0.7643
No log 17.75 426 0.5796 0.6508 0.5796 0.7613
No log 17.8333 428 0.5820 0.6164 0.5820 0.7629
No log 17.9167 430 0.5820 0.6479 0.5820 0.7629
No log 18.0 432 0.5784 0.6770 0.5784 0.7605
No log 18.0833 434 0.5746 0.7026 0.5746 0.7580
No log 18.1667 436 0.5917 0.6997 0.5917 0.7693
No log 18.25 438 0.6078 0.6997 0.6078 0.7796
No log 18.3333 440 0.5874 0.6997 0.5874 0.7664
No log 18.4167 442 0.5829 0.6822 0.5829 0.7634
No log 18.5 444 0.5926 0.6498 0.5926 0.7698
No log 18.5833 446 0.6254 0.5561 0.6254 0.7909
No log 18.6667 448 0.6808 0.5676 0.6808 0.8251
No log 18.75 450 0.7127 0.5770 0.7127 0.8442
No log 18.8333 452 0.6717 0.5738 0.6717 0.8196
No log 18.9167 454 0.6177 0.6535 0.6177 0.7859
No log 19.0 456 0.6072 0.6822 0.6072 0.7793
No log 19.0833 458 0.6149 0.6822 0.6149 0.7842
No log 19.1667 460 0.6084 0.6861 0.6084 0.7800
No log 19.25 462 0.6702 0.6098 0.6702 0.8187
No log 19.3333 464 0.8135 0.5759 0.8135 0.9019
No log 19.4167 466 0.8905 0.5484 0.8905 0.9437
No log 19.5 468 0.8182 0.5759 0.8182 0.9046
No log 19.5833 470 0.6881 0.5748 0.6881 0.8295
No log 19.6667 472 0.6253 0.5945 0.6253 0.7907
No log 19.75 474 0.6194 0.6286 0.6194 0.7870
No log 19.8333 476 0.6221 0.6488 0.6221 0.7887
No log 19.9167 478 0.6203 0.6575 0.6203 0.7876
No log 20.0 480 0.6329 0.5930 0.6329 0.7955
No log 20.0833 482 0.6701 0.5862 0.6701 0.8186
No log 20.1667 484 0.6799 0.5849 0.6799 0.8245
No log 20.25 486 0.6596 0.5826 0.6596 0.8121
No log 20.3333 488 0.6275 0.5930 0.6275 0.7922
No log 20.4167 490 0.6191 0.5950 0.6191 0.7868
No log 20.5 492 0.6290 0.5862 0.6290 0.7931
No log 20.5833 494 0.6317 0.5746 0.6317 0.7948
No log 20.6667 496 0.6300 0.5972 0.6300 0.7937
No log 20.75 498 0.6229 0.6259 0.6229 0.7892
0.2479 20.8333 500 0.6203 0.6246 0.6203 0.7876
0.2479 20.9167 502 0.6270 0.5915 0.6270 0.7918
0.2479 21.0 504 0.6166 0.6107 0.6166 0.7852
0.2479 21.0833 506 0.6166 0.6408 0.6166 0.7853
0.2479 21.1667 508 0.6286 0.6594 0.6286 0.7928
0.2479 21.25 510 0.6324 0.6408 0.6324 0.7952
0.2479 21.3333 512 0.6236 0.6440 0.6236 0.7897
0.2479 21.4167 514 0.6137 0.6246 0.6137 0.7834
0.2479 21.5 516 0.6301 0.5737 0.6301 0.7938
0.2479 21.5833 518 0.6662 0.5450 0.6662 0.8162
0.2479 21.6667 520 0.6626 0.5214 0.6626 0.8140
0.2479 21.75 522 0.6290 0.5315 0.6290 0.7931

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k7_task5_organization

Finetuned
(4019)
this model