ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k6_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8519
  • Qwk: 0.6711
  • Mse: 0.8519
  • Rmse: 0.9230

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0714 2 6.8483 0.0242 6.8483 2.6169
No log 0.1429 4 4.2553 0.0702 4.2553 2.0628
No log 0.2143 6 3.0992 0.0494 3.0992 1.7604
No log 0.2857 8 2.3857 0.0758 2.3857 1.5446
No log 0.3571 10 2.3365 -0.0496 2.3365 1.5286
No log 0.4286 12 1.7178 0.1538 1.7178 1.3107
No log 0.5 14 2.1385 0.1538 2.1385 1.4624
No log 0.5714 16 3.1436 0.1059 3.1436 1.7730
No log 0.6429 18 3.1489 0.1078 3.1489 1.7745
No log 0.7143 20 2.5794 0.1104 2.5794 1.6060
No log 0.7857 22 2.1193 0.2466 2.1193 1.4558
No log 0.8571 24 2.0308 0.3288 2.0308 1.4251
No log 0.9286 26 1.8235 0.3459 1.8235 1.3504
No log 1.0 28 1.6878 0.4127 1.6878 1.2992
No log 1.0714 30 1.5663 0.4160 1.5663 1.2515
No log 1.1429 32 1.7135 0.3803 1.7135 1.3090
No log 1.2143 34 2.4057 0.2000 2.4057 1.5510
No log 1.2857 36 3.1586 0.2105 3.1586 1.7772
No log 1.3571 38 3.2935 0.2374 3.2935 1.8148
No log 1.4286 40 2.4176 0.3005 2.4176 1.5549
No log 1.5 42 1.4300 0.5405 1.4300 1.1958
No log 1.5714 44 1.3056 0.5 1.3056 1.1426
No log 1.6429 46 1.4939 0.5135 1.4939 1.2223
No log 1.7143 48 1.7000 0.4516 1.7000 1.3039
No log 1.7857 50 1.7580 0.4076 1.7580 1.3259
No log 1.8571 52 1.7545 0.3822 1.7545 1.3246
No log 1.9286 54 1.3104 0.5775 1.3104 1.1447
No log 2.0 56 1.0248 0.5857 1.0248 1.0123
No log 2.0714 58 0.9287 0.6143 0.9287 0.9637
No log 2.1429 60 0.8782 0.7172 0.8782 0.9371
No log 2.2143 62 0.8689 0.6944 0.8689 0.9322
No log 2.2857 64 0.9753 0.6259 0.9753 0.9876
No log 2.3571 66 1.5911 0.5638 1.5911 1.2614
No log 2.4286 68 2.0574 0.4306 2.0574 1.4344
No log 2.5 70 1.8307 0.5700 1.8307 1.3530
No log 2.5714 72 1.1568 0.6358 1.1568 1.0755
No log 2.6429 74 0.9008 0.7425 0.9008 0.9491
No log 2.7143 76 0.9285 0.7701 0.9285 0.9636
No log 2.7857 78 1.1918 0.6383 1.1918 1.0917
No log 2.8571 80 1.6383 0.5631 1.6383 1.2799
No log 2.9286 82 1.9032 0.5164 1.9032 1.3796
No log 3.0 84 1.5576 0.5756 1.5576 1.2480
No log 3.0714 86 0.9723 0.6744 0.9723 0.9861
No log 3.1429 88 0.8037 0.7162 0.8037 0.8965
No log 3.2143 90 0.8239 0.6757 0.8239 0.9077
No log 3.2857 92 0.8418 0.6849 0.8418 0.9175
No log 3.3571 94 0.8943 0.6434 0.8943 0.9457
No log 3.4286 96 1.2668 0.5616 1.2668 1.1255
No log 3.5 98 1.3191 0.5405 1.3191 1.1485
No log 3.5714 100 1.2024 0.5714 1.2024 1.0965
No log 3.6429 102 0.9286 0.6667 0.9286 0.9636
No log 3.7143 104 0.7043 0.7347 0.7043 0.8392
No log 3.7857 106 0.6984 0.6933 0.6984 0.8357
No log 3.8571 108 0.7986 0.6712 0.7986 0.8937
No log 3.9286 110 0.7605 0.6993 0.7605 0.8721
No log 4.0 112 0.8476 0.7051 0.8476 0.9207
No log 4.0714 114 1.1007 0.6135 1.1007 1.0491
No log 4.1429 116 1.2791 0.6111 1.2791 1.1310
No log 4.2143 118 1.1547 0.5935 1.1547 1.0746
No log 4.2857 120 1.0334 0.5985 1.0334 1.0165
No log 4.3571 122 1.0444 0.5985 1.0444 1.0219
No log 4.4286 124 1.1124 0.5985 1.1124 1.0547
No log 4.5 126 1.0447 0.5985 1.0447 1.0221
No log 4.5714 128 1.0337 0.6143 1.0337 1.0167
No log 4.6429 130 1.1096 0.6351 1.1096 1.0534
No log 4.7143 132 0.9678 0.6577 0.9678 0.9838
No log 4.7857 134 0.7798 0.7162 0.7798 0.8831
No log 4.8571 136 0.6373 0.7467 0.6373 0.7983
No log 4.9286 138 0.6208 0.7632 0.6208 0.7879
No log 5.0 140 0.6685 0.7662 0.6685 0.8176
No log 5.0714 142 0.8938 0.7135 0.8938 0.9454
No log 5.1429 144 1.0089 0.6514 1.0089 1.0044
No log 5.2143 146 1.0834 0.6316 1.0834 1.0409
No log 5.2857 148 0.9138 0.6968 0.9138 0.9559
No log 5.3571 150 0.6818 0.7763 0.6818 0.8257
No log 5.4286 152 0.6955 0.7273 0.6955 0.8340
No log 5.5 154 0.7379 0.7075 0.7379 0.8590
No log 5.5714 156 0.8085 0.6483 0.8085 0.8992
No log 5.6429 158 0.8182 0.6528 0.8182 0.9045
No log 5.7143 160 0.8205 0.6377 0.8205 0.9058
No log 5.7857 162 0.8915 0.6107 0.8915 0.9442
No log 5.8571 164 0.9326 0.5954 0.9326 0.9657
No log 5.9286 166 0.9215 0.5693 0.9215 0.9600
No log 6.0 168 1.0441 0.6069 1.0441 1.0218
No log 6.0714 170 1.4206 0.5989 1.4206 1.1919
No log 6.1429 172 1.4585 0.6 1.4585 1.2077
No log 6.2143 174 1.0576 0.6514 1.0576 1.0284
No log 6.2857 176 0.8328 0.7219 0.8328 0.9126
No log 6.3571 178 0.7027 0.7613 0.7027 0.8382
No log 6.4286 180 0.7071 0.7682 0.7071 0.8409
No log 6.5 182 0.7062 0.6986 0.7062 0.8404
No log 6.5714 184 0.7558 0.7248 0.7558 0.8693
No log 6.6429 186 0.7974 0.6573 0.7974 0.8930
No log 6.7143 188 0.7792 0.6713 0.7792 0.8827
No log 6.7857 190 0.7128 0.7034 0.7128 0.8443
No log 6.8571 192 0.6143 0.7703 0.6143 0.7838
No log 6.9286 194 0.6475 0.8176 0.6475 0.8047
No log 7.0 196 0.9502 0.7213 0.9502 0.9748
No log 7.0714 198 1.3218 0.5990 1.3218 1.1497
No log 7.1429 200 1.1758 0.6211 1.1758 1.0843
No log 7.2143 202 0.8082 0.7425 0.8082 0.8990
No log 7.2857 204 0.7117 0.7778 0.7117 0.8437
No log 7.3571 206 0.6946 0.7368 0.6946 0.8334
No log 7.4286 208 0.7827 0.7702 0.7827 0.8847
No log 7.5 210 0.8767 0.6747 0.8767 0.9363
No log 7.5714 212 0.8272 0.7308 0.8272 0.9095
No log 7.6429 214 0.6988 0.7027 0.6988 0.8359
No log 7.7143 216 0.6826 0.7222 0.6826 0.8262
No log 7.7857 218 0.7329 0.6761 0.7329 0.8561
No log 7.8571 220 0.8396 0.6043 0.8396 0.9163
No log 7.9286 222 0.8157 0.6232 0.8157 0.9032
No log 8.0 224 0.7752 0.6866 0.7752 0.8805
No log 8.0714 226 0.8457 0.6615 0.8457 0.9196
No log 8.1429 228 0.8256 0.6870 0.8256 0.9086
No log 8.2143 230 0.7420 0.7376 0.7420 0.8614
No log 8.2857 232 0.7530 0.6475 0.7530 0.8677
No log 8.3571 234 0.8607 0.6752 0.8607 0.9278
No log 8.4286 236 1.0144 0.6707 1.0144 1.0072
No log 8.5 238 0.9194 0.6914 0.9194 0.9588
No log 8.5714 240 0.7308 0.6434 0.7308 0.8549
No log 8.6429 242 0.6458 0.7586 0.6458 0.8036
No log 8.7143 244 0.6682 0.7448 0.6682 0.8174
No log 8.7857 246 0.6765 0.75 0.6765 0.8225
No log 8.8571 248 0.7626 0.6528 0.7626 0.8732
No log 8.9286 250 0.9629 0.7125 0.9629 0.9813
No log 9.0 252 1.0433 0.7030 1.0433 1.0214
No log 9.0714 254 0.9069 0.7105 0.9069 0.9523
No log 9.1429 256 0.7432 0.6986 0.7432 0.8621
No log 9.2143 258 0.6975 0.6986 0.6975 0.8352
No log 9.2857 260 0.7864 0.7123 0.7864 0.8868
No log 9.3571 262 0.9959 0.6386 0.9959 0.9979
No log 9.4286 264 1.0890 0.6235 1.0890 1.0436
No log 9.5 266 1.1029 0.6341 1.1029 1.0502
No log 9.5714 268 1.1899 0.65 1.1899 1.0908
No log 9.6429 270 1.1206 0.6533 1.1206 1.0586
No log 9.7143 272 0.9350 0.6043 0.9350 0.9670
No log 9.7857 274 0.8765 0.6620 0.8765 0.9362
No log 9.8571 276 0.7771 0.6383 0.7771 0.8815
No log 9.9286 278 0.7444 0.6620 0.7444 0.8628
No log 10.0 280 0.7500 0.6620 0.7500 0.8660
No log 10.0714 282 0.8414 0.6883 0.8414 0.9173
No log 10.1429 284 0.8214 0.6759 0.8214 0.9063
No log 10.2143 286 0.8196 0.6712 0.8196 0.9053
No log 10.2857 288 0.8185 0.6712 0.8185 0.9047
No log 10.3571 290 0.8785 0.6849 0.8785 0.9373
No log 10.4286 292 0.9487 0.6667 0.9487 0.9740
No log 10.5 294 1.1240 0.6503 1.1240 1.0602
No log 10.5714 296 1.3009 0.6136 1.3009 1.1406
No log 10.6429 298 1.3535 0.6034 1.3535 1.1634
No log 10.7143 300 1.3385 0.6034 1.3385 1.1569
No log 10.7857 302 1.2932 0.6034 1.2932 1.1372
No log 10.8571 304 0.9762 0.6585 0.9762 0.9880
No log 10.9286 306 0.9019 0.7105 0.9019 0.9497
No log 11.0 308 1.0581 0.6133 1.0581 1.0287
No log 11.0714 310 1.2683 0.6452 1.2683 1.1262
No log 11.1429 312 1.1949 0.5874 1.1949 1.0931
No log 11.2143 314 1.0309 0.5899 1.0309 1.0153
No log 11.2857 316 0.8234 0.6377 0.8234 0.9074
No log 11.3571 318 0.7315 0.6763 0.7315 0.8553
No log 11.4286 320 0.7477 0.6619 0.7477 0.8647
No log 11.5 322 0.8298 0.6571 0.8298 0.9110
No log 11.5714 324 0.7900 0.6667 0.7900 0.8888
No log 11.6429 326 0.8108 0.6667 0.8108 0.9005
No log 11.7143 328 0.8679 0.6667 0.8679 0.9316
No log 11.7857 330 0.8143 0.6667 0.8143 0.9024
No log 11.8571 332 0.7500 0.6619 0.7500 0.8660
No log 11.9286 334 0.7371 0.6950 0.7371 0.8586
No log 12.0 336 0.7725 0.6571 0.7725 0.8789
No log 12.0714 338 0.8648 0.6241 0.8648 0.9300
No log 12.1429 340 0.9705 0.6207 0.9705 0.9852
No log 12.2143 342 1.1560 0.6369 1.1560 1.0752
No log 12.2857 344 1.0753 0.6543 1.0753 1.0370
No log 12.3571 346 0.8438 0.6933 0.8438 0.9186
No log 12.4286 348 0.7227 0.7027 0.7227 0.8501
No log 12.5 350 0.7609 0.7436 0.7609 0.8723
No log 12.5714 352 0.7111 0.7333 0.7111 0.8433
No log 12.6429 354 0.7426 0.7237 0.7426 0.8617
No log 12.7143 356 0.8979 0.7329 0.8979 0.9476
No log 12.7857 358 0.9576 0.7125 0.9576 0.9786
No log 12.8571 360 0.7990 0.7105 0.7990 0.8939
No log 12.9286 362 0.8032 0.7105 0.8032 0.8962
No log 13.0 364 0.8459 0.6887 0.8459 0.9197
No log 13.0714 366 0.6871 0.7383 0.6871 0.8289
No log 13.1429 368 0.5879 0.7397 0.5879 0.7667
No log 13.2143 370 0.6186 0.7429 0.6186 0.7865
No log 13.2857 372 0.6346 0.7432 0.6346 0.7966
No log 13.3571 374 0.6686 0.7162 0.6686 0.8177
No log 13.4286 376 0.7485 0.7020 0.7485 0.8652
No log 13.5 378 0.8893 0.7329 0.8893 0.9430
No log 13.5714 380 0.9455 0.6914 0.9455 0.9724
No log 13.6429 382 0.8442 0.7453 0.8442 0.9188
No log 13.7143 384 0.7141 0.7662 0.7141 0.8451
No log 13.7857 386 0.6414 0.7333 0.6414 0.8008
No log 13.8571 388 0.6300 0.7260 0.6300 0.7937
No log 13.9286 390 0.6699 0.7320 0.6699 0.8185
No log 14.0 392 0.7683 0.7531 0.7683 0.8765
No log 14.0714 394 0.7479 0.7261 0.7479 0.8648
No log 14.1429 396 0.6806 0.7211 0.6806 0.8250
No log 14.2143 398 0.7208 0.7123 0.7208 0.8490
No log 14.2857 400 0.7855 0.7075 0.7855 0.8863
No log 14.3571 402 0.8799 0.6389 0.8799 0.9380
No log 14.4286 404 0.8759 0.6389 0.8759 0.9359
No log 14.5 406 0.8243 0.7162 0.8243 0.9079
No log 14.5714 408 0.8165 0.6986 0.8165 0.9036
No log 14.6429 410 0.7823 0.7432 0.7823 0.8845
No log 14.7143 412 0.8339 0.7114 0.8339 0.9132
No log 14.7857 414 1.0806 0.6154 1.0806 1.0395
No log 14.8571 416 1.3318 0.5614 1.3318 1.1541
No log 14.9286 418 1.2321 0.5890 1.2321 1.1100
No log 15.0 420 1.0503 0.5986 1.0503 1.0248
No log 15.0714 422 1.0069 0.6069 1.0069 1.0034
No log 15.1429 424 0.9370 0.6939 0.9370 0.9680
No log 15.2143 426 0.9540 0.6483 0.9540 0.9767
No log 15.2857 428 1.0004 0.5816 1.0004 1.0002
No log 15.3571 430 1.0388 0.6014 1.0388 1.0192
No log 15.4286 432 1.1750 0.5906 1.1750 1.0840
No log 15.5 434 1.2679 0.5906 1.2679 1.1260
No log 15.5714 436 1.1770 0.5906 1.1770 1.0849
No log 15.6429 438 0.9554 0.6575 0.9554 0.9775
No log 15.7143 440 0.7457 0.6986 0.7457 0.8636
No log 15.7857 442 0.6959 0.7297 0.6959 0.8342
No log 15.8571 444 0.7117 0.7297 0.7117 0.8436
No log 15.9286 446 0.8284 0.7607 0.8284 0.9102
No log 16.0 448 1.0277 0.6424 1.0277 1.0138
No log 16.0714 450 1.1787 0.6127 1.1787 1.0857
No log 16.1429 452 1.1040 0.6235 1.1040 1.0507
No log 16.2143 454 0.9267 0.6962 0.9267 0.9627
No log 16.2857 456 0.7431 0.6944 0.7431 0.8621
No log 16.3571 458 0.6743 0.7397 0.6743 0.8212
No log 16.4286 460 0.6808 0.7347 0.6808 0.8251
No log 16.5 462 0.7444 0.7547 0.7444 0.8628
No log 16.5714 464 0.8993 0.7241 0.8993 0.9483
No log 16.6429 466 1.0118 0.6893 1.0118 1.0059
No log 16.7143 468 1.0443 0.6552 1.0443 1.0219
No log 16.7857 470 1.1560 0.6433 1.1560 1.0752
No log 16.8571 472 1.1823 0.6463 1.1823 1.0873
No log 16.9286 474 1.0635 0.6497 1.0635 1.0313
No log 17.0 476 0.9272 0.6839 0.9272 0.9629
No log 17.0714 478 0.8677 0.7261 0.8677 0.9315
No log 17.1429 480 0.9658 0.7037 0.9658 0.9828
No log 17.2143 482 1.2380 0.6034 1.2380 1.1126
No log 17.2857 484 1.3543 0.6196 1.3543 1.1638
No log 17.3571 486 1.2722 0.6188 1.2722 1.1279
No log 17.4286 488 0.9888 0.6627 0.9888 0.9944
No log 17.5 490 0.7958 0.7027 0.7958 0.8921
No log 17.5714 492 0.7636 0.7297 0.7636 0.8739
No log 17.6429 494 0.8253 0.6528 0.8253 0.9085
No log 17.7143 496 0.8740 0.6294 0.8740 0.9349
No log 17.7857 498 0.8456 0.6434 0.8456 0.9196
0.3494 17.8571 500 0.8008 0.6853 0.8008 0.8949
0.3494 17.9286 502 0.7370 0.7347 0.7370 0.8585
0.3494 18.0 504 0.7771 0.7347 0.7771 0.8815
0.3494 18.0714 506 0.8528 0.6483 0.8528 0.9235
0.3494 18.1429 508 1.0156 0.6443 1.0156 1.0078
0.3494 18.2143 510 1.1386 0.6 1.1386 1.0671
0.3494 18.2857 512 1.0605 0.6234 1.0605 1.0298
0.3494 18.3571 514 0.8519 0.6711 0.8519 0.9230

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k6_task1_organization

Finetuned
(4019)
this model