ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run3_AugV5_k13_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6196
  • Qwk: 0.6848
  • Mse: 0.6196
  • Rmse: 0.7871

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0317 2 5.2414 0.0134 5.2414 2.2894
No log 0.0635 4 3.3122 0.0659 3.3122 1.8199
No log 0.0952 6 2.3654 -0.0557 2.3654 1.5380
No log 0.1270 8 2.6147 -0.1560 2.6147 1.6170
No log 0.1587 10 2.0225 -0.0757 2.0225 1.4221
No log 0.1905 12 1.4172 0.0327 1.4172 1.1904
No log 0.2222 14 1.2942 0.1459 1.2942 1.1376
No log 0.2540 16 1.2887 0.1581 1.2887 1.1352
No log 0.2857 18 1.4483 0.0820 1.4483 1.2034
No log 0.3175 20 1.8181 0.0747 1.8181 1.3484
No log 0.3492 22 2.2737 0.1204 2.2737 1.5079
No log 0.3810 24 2.0601 0.1284 2.0601 1.4353
No log 0.4127 26 1.5401 0.1457 1.5401 1.2410
No log 0.4444 28 1.2189 0.2803 1.2189 1.1040
No log 0.4762 30 1.1539 0.2889 1.1539 1.0742
No log 0.5079 32 1.1971 0.2985 1.1971 1.0941
No log 0.5397 34 1.5240 0.2334 1.5240 1.2345
No log 0.5714 36 2.0399 0.2070 2.0399 1.4283
No log 0.6032 38 1.6439 0.2724 1.6439 1.2822
No log 0.6349 40 1.2398 0.3096 1.2398 1.1135
No log 0.6667 42 1.1509 0.2987 1.1509 1.0728
No log 0.6984 44 1.2371 0.2866 1.2371 1.1123
No log 0.7302 46 1.5570 0.2246 1.5570 1.2478
No log 0.7619 48 2.1905 0.1820 2.1905 1.4800
No log 0.7937 50 2.1969 0.1984 2.1969 1.4822
No log 0.8254 52 2.7083 0.1017 2.7083 1.6457
No log 0.8571 54 2.3789 0.1391 2.3789 1.5424
No log 0.8889 56 1.4644 0.2048 1.4644 1.2101
No log 0.9206 58 1.1202 0.2929 1.1202 1.0584
No log 0.9524 60 1.1945 0.2424 1.1945 1.0929
No log 0.9841 62 1.1572 0.2536 1.1572 1.0757
No log 1.0159 64 1.1331 0.2676 1.1331 1.0645
No log 1.0476 66 1.3130 0.1411 1.3130 1.1459
No log 1.0794 68 1.7790 0.1622 1.7790 1.3338
No log 1.1111 70 1.8647 0.1639 1.8647 1.3655
No log 1.1429 72 1.8666 0.1602 1.8666 1.3663
No log 1.1746 74 1.6028 0.1705 1.6028 1.2660
No log 1.2063 76 1.3553 0.2113 1.3553 1.1642
No log 1.2381 78 1.2825 0.1989 1.2825 1.1325
No log 1.2698 80 1.2189 0.2136 1.2189 1.1040
No log 1.3016 82 1.3744 0.2428 1.3744 1.1723
No log 1.3333 84 1.8164 0.2481 1.8164 1.3477
No log 1.3651 86 1.9306 0.2859 1.9306 1.3895
No log 1.3968 88 1.5811 0.2841 1.5811 1.2574
No log 1.4286 90 1.2157 0.3257 1.2157 1.1026
No log 1.4603 92 1.1511 0.3930 1.1511 1.0729
No log 1.4921 94 1.1278 0.3884 1.1278 1.0620
No log 1.5238 96 1.0737 0.3934 1.0737 1.0362
No log 1.5556 98 1.0532 0.4433 1.0532 1.0262
No log 1.5873 100 1.0971 0.4734 1.0971 1.0474
No log 1.6190 102 1.4189 0.3312 1.4189 1.1912
No log 1.6508 104 1.8234 0.3173 1.8234 1.3503
No log 1.6825 106 1.8873 0.3143 1.8873 1.3738
No log 1.7143 108 1.9295 0.3143 1.9295 1.3891
No log 1.7460 110 2.1085 0.2768 2.1085 1.4521
No log 1.7778 112 1.7330 0.3114 1.7330 1.3164
No log 1.8095 114 1.2445 0.3345 1.2445 1.1156
No log 1.8413 116 0.9606 0.4932 0.9606 0.9801
No log 1.8730 118 0.8864 0.5185 0.8864 0.9415
No log 1.9048 120 0.9679 0.4731 0.9679 0.9838
No log 1.9365 122 1.2407 0.3963 1.2407 1.1139
No log 1.9683 124 1.2425 0.3758 1.2425 1.1147
No log 2.0 126 0.9806 0.4226 0.9806 0.9902
No log 2.0317 128 0.8215 0.4825 0.8215 0.9064
No log 2.0635 130 0.7830 0.4970 0.7830 0.8849
No log 2.0952 132 0.7578 0.5850 0.7578 0.8705
No log 2.1270 134 0.8191 0.5767 0.8191 0.9050
No log 2.1587 136 0.9226 0.5538 0.9226 0.9605
No log 2.1905 138 1.0146 0.5139 1.0146 1.0073
No log 2.2222 140 0.8667 0.5554 0.8667 0.9310
No log 2.2540 142 0.8183 0.5775 0.8183 0.9046
No log 2.2857 144 0.7865 0.5662 0.7865 0.8868
No log 2.3175 146 0.7851 0.5874 0.7851 0.8861
No log 2.3492 148 0.8496 0.5620 0.8496 0.9217
No log 2.3810 150 0.7940 0.5862 0.7940 0.8910
No log 2.4127 152 0.7951 0.5755 0.7951 0.8917
No log 2.4444 154 0.7391 0.6095 0.7391 0.8597
No log 2.4762 156 0.7606 0.6389 0.7606 0.8721
No log 2.5079 158 0.8564 0.6040 0.8564 0.9254
No log 2.5397 160 0.8172 0.6063 0.8172 0.9040
No log 2.5714 162 0.7308 0.5984 0.7308 0.8549
No log 2.6032 164 0.7276 0.5920 0.7276 0.8530
No log 2.6349 166 0.7228 0.5958 0.7228 0.8502
No log 2.6667 168 0.7194 0.6235 0.7194 0.8482
No log 2.6984 170 0.8228 0.6217 0.8229 0.9071
No log 2.7302 172 1.2662 0.5102 1.2662 1.1253
No log 2.7619 174 1.4675 0.4509 1.4675 1.2114
No log 2.7937 176 1.0979 0.5438 1.0979 1.0478
No log 2.8254 178 0.7107 0.6500 0.7107 0.8430
No log 2.8571 180 0.8104 0.6392 0.8104 0.9002
No log 2.8889 182 0.8500 0.6130 0.8500 0.9219
No log 2.9206 184 0.7858 0.6376 0.7858 0.8865
No log 2.9524 186 0.7518 0.6486 0.7518 0.8671
No log 2.9841 188 0.7736 0.6566 0.7736 0.8796
No log 3.0159 190 0.7415 0.6535 0.7415 0.8611
No log 3.0476 192 0.7553 0.6623 0.7553 0.8691
No log 3.0794 194 0.8238 0.6364 0.8238 0.9077
No log 3.1111 196 0.7646 0.6363 0.7646 0.8744
No log 3.1429 198 0.7292 0.5938 0.7292 0.8539
No log 3.1746 200 0.8311 0.5535 0.8311 0.9117
No log 3.2063 202 0.8349 0.5796 0.8349 0.9137
No log 3.2381 204 0.6943 0.6595 0.6943 0.8332
No log 3.2698 206 0.7668 0.6557 0.7668 0.8757
No log 3.3016 208 0.8168 0.6559 0.8168 0.9038
No log 3.3333 210 0.7250 0.6887 0.7250 0.8514
No log 3.3651 212 0.7709 0.6717 0.7709 0.8780
No log 3.3968 214 0.8219 0.6690 0.8219 0.9066
No log 3.4286 216 0.7064 0.6724 0.7064 0.8405
No log 3.4603 218 0.7319 0.6837 0.7319 0.8555
No log 3.4921 220 0.8161 0.6007 0.8161 0.9034
No log 3.5238 222 0.9106 0.6147 0.9106 0.9542
No log 3.5556 224 0.9164 0.6085 0.9164 0.9573
No log 3.5873 226 0.7966 0.5814 0.7966 0.8925
No log 3.6190 228 0.7488 0.5392 0.7488 0.8653
No log 3.6508 230 0.7645 0.5444 0.7645 0.8744
No log 3.6825 232 0.8321 0.5895 0.8321 0.9122
No log 3.7143 234 0.9477 0.5704 0.9477 0.9735
No log 3.7460 236 0.9822 0.5496 0.9822 0.9911
No log 3.7778 238 0.8760 0.6088 0.8760 0.9359
No log 3.8095 240 0.8209 0.6217 0.8209 0.9061
No log 3.8413 242 0.7598 0.6321 0.7598 0.8717
No log 3.8730 244 0.7390 0.6729 0.7390 0.8597
No log 3.9048 246 0.6890 0.7123 0.6890 0.8301
No log 3.9365 248 0.6828 0.7117 0.6828 0.8263
No log 3.9683 250 0.6645 0.6899 0.6645 0.8152
No log 4.0 252 0.6560 0.6434 0.6560 0.8100
No log 4.0317 254 0.6502 0.5548 0.6502 0.8063
No log 4.0635 256 0.6686 0.5765 0.6686 0.8177
No log 4.0952 258 0.6738 0.5506 0.6738 0.8209
No log 4.1270 260 0.6616 0.5953 0.6616 0.8134
No log 4.1587 262 0.6573 0.6324 0.6573 0.8108
No log 4.1905 264 0.6772 0.6452 0.6772 0.8229
No log 4.2222 266 0.6576 0.6858 0.6576 0.8109
No log 4.2540 268 0.6878 0.6813 0.6878 0.8294
No log 4.2857 270 0.6950 0.6855 0.6950 0.8337
No log 4.3175 272 0.7179 0.6731 0.7179 0.8473
No log 4.3492 274 0.7984 0.6404 0.7984 0.8935
No log 4.3810 276 0.9903 0.5529 0.9903 0.9951
No log 4.4127 278 1.1230 0.5401 1.1230 1.0597
No log 4.4444 280 1.1479 0.5220 1.1479 1.0714
No log 4.4762 282 0.9160 0.6033 0.9160 0.9571
No log 4.5079 284 0.7308 0.6440 0.7308 0.8549
No log 4.5397 286 0.6439 0.6630 0.6439 0.8024
No log 4.5714 288 0.6410 0.6863 0.6410 0.8006
No log 4.6032 290 0.6883 0.6822 0.6883 0.8296
No log 4.6349 292 0.7120 0.6804 0.7120 0.8438
No log 4.6667 294 0.7109 0.6883 0.7109 0.8432
No log 4.6984 296 0.6417 0.7267 0.6417 0.8010
No log 4.7302 298 0.6677 0.6893 0.6677 0.8171
No log 4.7619 300 0.7052 0.6476 0.7052 0.8398
No log 4.7937 302 0.6625 0.6897 0.6625 0.8140
No log 4.8254 304 0.6779 0.7040 0.6779 0.8234
No log 4.8571 306 0.7150 0.6867 0.7150 0.8456
No log 4.8889 308 0.6919 0.6853 0.6919 0.8318
No log 4.9206 310 0.6826 0.6633 0.6826 0.8262
No log 4.9524 312 0.7174 0.6503 0.7174 0.8470
No log 4.9841 314 0.7226 0.6473 0.7226 0.8500
No log 5.0159 316 0.7180 0.6578 0.7180 0.8473
No log 5.0476 318 0.7459 0.6467 0.7459 0.8637
No log 5.0794 320 0.8267 0.6262 0.8267 0.9092
No log 5.1111 322 0.8135 0.6389 0.8135 0.9020
No log 5.1429 324 0.7795 0.6459 0.7795 0.8829
No log 5.1746 326 0.6998 0.6841 0.6998 0.8365
No log 5.2063 328 0.6782 0.6339 0.6782 0.8236
No log 5.2381 330 0.6766 0.6339 0.6766 0.8226
No log 5.2698 332 0.6906 0.5669 0.6906 0.8310
No log 5.3016 334 0.6822 0.6074 0.6822 0.8259
No log 5.3333 336 0.7463 0.6422 0.7463 0.8639
No log 5.3651 338 0.8757 0.6136 0.8757 0.9358
No log 5.3968 340 0.8437 0.6112 0.8437 0.9185
No log 5.4286 342 0.7101 0.6764 0.7101 0.8427
No log 5.4603 344 0.6678 0.6890 0.6678 0.8172
No log 5.4921 346 0.6640 0.6897 0.6640 0.8149
No log 5.5238 348 0.7126 0.7063 0.7126 0.8442
No log 5.5556 350 0.7524 0.6799 0.7524 0.8674
No log 5.5873 352 0.7226 0.6971 0.7226 0.8501
No log 5.6190 354 0.6640 0.6885 0.6640 0.8149
No log 5.6508 356 0.6674 0.6562 0.6674 0.8170
No log 5.6825 358 0.6691 0.6801 0.6691 0.8180
No log 5.7143 360 0.7171 0.6852 0.7171 0.8468
No log 5.7460 362 0.7938 0.6610 0.7938 0.8909
No log 5.7778 364 0.7221 0.6850 0.7221 0.8498
No log 5.8095 366 0.6785 0.6835 0.6785 0.8237
No log 5.8413 368 0.6818 0.6869 0.6818 0.8257
No log 5.8730 370 0.6989 0.6752 0.6989 0.8360
No log 5.9048 372 0.7260 0.6755 0.7260 0.8521
No log 5.9365 374 0.8020 0.6014 0.8020 0.8956
No log 5.9683 376 0.8502 0.5913 0.8502 0.9220
No log 6.0 378 0.8352 0.5913 0.8352 0.9139
No log 6.0317 380 0.7635 0.6359 0.7635 0.8738
No log 6.0635 382 0.7501 0.6635 0.7501 0.8661
No log 6.0952 384 0.7750 0.6453 0.7750 0.8803
No log 6.1270 386 0.8881 0.5995 0.8881 0.9424
No log 6.1587 388 1.0510 0.5518 1.0510 1.0252
No log 6.1905 390 1.0455 0.5707 1.0455 1.0225
No log 6.2222 392 0.9005 0.5741 0.9005 0.9489
No log 6.2540 394 0.7551 0.6535 0.7551 0.8690
No log 6.2857 396 0.7411 0.6554 0.7411 0.8609
No log 6.3175 398 0.7928 0.6147 0.7928 0.8904
No log 6.3492 400 0.8515 0.6104 0.8515 0.9228
No log 6.3810 402 0.8495 0.6194 0.8495 0.9217
No log 6.4127 404 0.7477 0.6176 0.7477 0.8647
No log 6.4444 406 0.7272 0.6349 0.7272 0.8528
No log 6.4762 408 0.7317 0.6011 0.7317 0.8554
No log 6.5079 410 0.7686 0.5952 0.7686 0.8767
No log 6.5397 412 0.8309 0.6430 0.8309 0.9116
No log 6.5714 414 0.7952 0.5995 0.7952 0.8917
No log 6.6032 416 0.7442 0.5842 0.7442 0.8627
No log 6.6349 418 0.7325 0.5641 0.7325 0.8558
No log 6.6667 420 0.7235 0.5716 0.7235 0.8506
No log 6.6984 422 0.7063 0.6329 0.7063 0.8404
No log 6.7302 424 0.7267 0.6403 0.7267 0.8524
No log 6.7619 426 0.7366 0.6523 0.7366 0.8582
No log 6.7937 428 0.7522 0.6646 0.7522 0.8673
No log 6.8254 430 0.7684 0.6134 0.7684 0.8766
No log 6.8571 432 0.7338 0.6390 0.7338 0.8566
No log 6.8889 434 0.7589 0.6322 0.7589 0.8711
No log 6.9206 436 0.8456 0.6324 0.8456 0.9196
No log 6.9524 438 0.8951 0.5951 0.8951 0.9461
No log 6.9841 440 0.9044 0.6085 0.9044 0.9510
No log 7.0159 442 0.8563 0.5993 0.8563 0.9254
No log 7.0476 444 0.8319 0.5964 0.8319 0.9121
No log 7.0794 446 0.7715 0.6099 0.7715 0.8783
No log 7.1111 448 0.7248 0.6179 0.7248 0.8513
No log 7.1429 450 0.6834 0.6408 0.6834 0.8267
No log 7.1746 452 0.7058 0.6341 0.7058 0.8401
No log 7.2063 454 0.6870 0.6918 0.6870 0.8289
No log 7.2381 456 0.7104 0.6614 0.7104 0.8428
No log 7.2698 458 0.7389 0.6447 0.7389 0.8596
No log 7.3016 460 0.7379 0.6293 0.7379 0.8590
No log 7.3333 462 0.7590 0.6028 0.7590 0.8712
No log 7.3651 464 0.7119 0.6059 0.7119 0.8437
No log 7.3968 466 0.7033 0.6094 0.7033 0.8386
No log 7.4286 468 0.6774 0.5924 0.6774 0.8230
No log 7.4603 470 0.6728 0.6066 0.6728 0.8202
No log 7.4921 472 0.7485 0.6657 0.7485 0.8651
No log 7.5238 474 0.7936 0.6434 0.7936 0.8908
No log 7.5556 476 0.6825 0.6585 0.6825 0.8262
No log 7.5873 478 0.6254 0.6789 0.6254 0.7908
No log 7.6190 480 0.6227 0.6758 0.6227 0.7891
No log 7.6508 482 0.6411 0.6842 0.6411 0.8007
No log 7.6825 484 0.6719 0.6768 0.6719 0.8197
No log 7.7143 486 0.6983 0.6624 0.6983 0.8356
No log 7.7460 488 0.7829 0.6088 0.7829 0.8848
No log 7.7778 490 0.7345 0.6488 0.7345 0.8570
No log 7.8095 492 0.6332 0.7041 0.6332 0.7957
No log 7.8413 494 0.6455 0.6888 0.6455 0.8034
No log 7.8730 496 0.7517 0.6328 0.7517 0.8670
No log 7.9048 498 0.7136 0.6768 0.7136 0.8448
0.5203 7.9365 500 0.6180 0.7012 0.6180 0.7861
0.5203 7.9683 502 0.7003 0.6675 0.7003 0.8368
0.5203 8.0 504 0.8190 0.6225 0.8190 0.9050
0.5203 8.0317 506 0.7981 0.6478 0.7981 0.8934
0.5203 8.0635 508 0.6719 0.6645 0.6719 0.8197
0.5203 8.0952 510 0.6196 0.6848 0.6196 0.7871

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run3_AugV5_k13_task1_organization

Finetuned
(4023)
this model