ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k11_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5951
  • Qwk: 0.5328
  • Mse: 0.5951
  • Rmse: 0.7714

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0526 2 2.5143 -0.0646 2.5143 1.5857
No log 0.1053 4 1.2715 0.1277 1.2715 1.1276
No log 0.1579 6 0.8541 0.0955 0.8541 0.9242
No log 0.2105 8 0.8060 0.2008 0.8060 0.8978
No log 0.2632 10 0.7222 0.2132 0.7222 0.8498
No log 0.3158 12 0.7386 0.2352 0.7386 0.8594
No log 0.3684 14 0.7781 0.3234 0.7781 0.8821
No log 0.4211 16 0.7218 0.2471 0.7218 0.8496
No log 0.4737 18 0.7661 0.1372 0.7661 0.8753
No log 0.5263 20 0.8303 0.2494 0.8303 0.9112
No log 0.5789 22 0.7542 0.1372 0.7542 0.8685
No log 0.6316 24 0.7509 0.0 0.7509 0.8666
No log 0.6842 26 0.7540 0.0 0.7540 0.8684
No log 0.7368 28 0.7544 0.0 0.7544 0.8686
No log 0.7895 30 0.7925 0.0078 0.7925 0.8902
No log 0.8421 32 0.8495 0.0535 0.8495 0.9217
No log 0.8947 34 1.0061 0.0964 1.0061 1.0031
No log 0.9474 36 1.0236 0.2008 1.0236 1.0117
No log 1.0 38 0.9635 0.2381 0.9635 0.9816
No log 1.0526 40 0.8930 0.2109 0.8930 0.9450
No log 1.1053 42 0.8100 0.0495 0.8100 0.9000
No log 1.1579 44 0.7615 0.1277 0.7615 0.8726
No log 1.2105 46 0.7783 0.1315 0.7783 0.8822
No log 1.2632 48 0.8478 0.2526 0.8478 0.9208
No log 1.3158 50 0.8507 0.2156 0.8507 0.9223
No log 1.3684 52 0.8295 0.1372 0.8295 0.9108
No log 1.4211 54 0.8341 0.1372 0.8341 0.9133
No log 1.4737 56 0.8246 0.1372 0.8246 0.9081
No log 1.5263 58 0.8027 0.0481 0.8027 0.8959
No log 1.5789 60 0.7908 0.0937 0.7908 0.8893
No log 1.6316 62 0.7404 0.0 0.7404 0.8605
No log 1.6842 64 0.7141 0.0840 0.7141 0.8451
No log 1.7368 66 0.6981 0.0840 0.6981 0.8355
No log 1.7895 68 0.6701 0.1674 0.6701 0.8186
No log 1.8421 70 0.6572 0.2046 0.6572 0.8107
No log 1.8947 72 0.6639 0.2046 0.6639 0.8148
No log 1.9474 74 0.6681 0.3184 0.6681 0.8174
No log 2.0 76 0.6950 0.4139 0.6950 0.8336
No log 2.0526 78 0.7685 0.4152 0.7685 0.8766
No log 2.1053 80 0.7362 0.4395 0.7362 0.8580
No log 2.1579 82 0.6671 0.4859 0.6671 0.8168
No log 2.2105 84 0.6060 0.3717 0.6060 0.7784
No log 2.2632 86 0.5826 0.4729 0.5826 0.7633
No log 2.3158 88 0.6242 0.4007 0.6242 0.7901
No log 2.3684 90 0.7379 0.4283 0.7379 0.8590
No log 2.4211 92 0.9075 0.2875 0.9075 0.9526
No log 2.4737 94 0.9642 0.2784 0.9642 0.9819
No log 2.5263 96 0.9137 0.2977 0.9137 0.9559
No log 2.5789 98 0.7715 0.1770 0.7715 0.8784
No log 2.6316 100 0.6976 0.0393 0.6976 0.8352
No log 2.6842 102 0.7341 0.2537 0.7341 0.8568
No log 2.7368 104 0.7770 0.2345 0.7770 0.8814
No log 2.7895 106 0.7495 0.2857 0.7495 0.8657
No log 2.8421 108 0.7811 0.1263 0.7811 0.8838
No log 2.8947 110 1.0514 0.0627 1.0514 1.0254
No log 2.9474 112 1.2884 0.0763 1.2884 1.1351
No log 3.0 114 1.3801 0.1001 1.3801 1.1748
No log 3.0526 116 1.2179 0.1514 1.2179 1.1036
No log 3.1053 118 0.8944 0.1553 0.8944 0.9457
No log 3.1579 120 0.7451 0.2642 0.7451 0.8632
No log 3.2105 122 0.7201 0.3729 0.7201 0.8486
No log 3.2632 124 0.7525 0.3999 0.7525 0.8675
No log 3.3158 126 0.7820 0.3002 0.7820 0.8843
No log 3.3684 128 0.9196 0.2273 0.9196 0.9590
No log 3.4211 130 0.9541 0.2514 0.9541 0.9768
No log 3.4737 132 0.8703 0.2557 0.8703 0.9329
No log 3.5263 134 0.6693 0.5112 0.6693 0.8181
No log 3.5789 136 0.5657 0.5479 0.5657 0.7521
No log 3.6316 138 0.5707 0.5195 0.5707 0.7555
No log 3.6842 140 0.6448 0.5247 0.6448 0.8030
No log 3.7368 142 0.7286 0.4765 0.7286 0.8536
No log 3.7895 144 0.7285 0.4986 0.7285 0.8535
No log 3.8421 146 0.6718 0.4747 0.6718 0.8197
No log 3.8947 148 0.6087 0.5822 0.6087 0.7802
No log 3.9474 150 0.5975 0.5893 0.5975 0.7730
No log 4.0 152 0.5878 0.5893 0.5878 0.7667
No log 4.0526 154 0.5836 0.6173 0.5836 0.7640
No log 4.1053 156 0.6050 0.5943 0.6050 0.7778
No log 4.1579 158 0.6940 0.5100 0.6940 0.8331
No log 4.2105 160 0.6423 0.5262 0.6423 0.8015
No log 4.2632 162 0.5834 0.6222 0.5834 0.7638
No log 4.3158 164 0.6925 0.4298 0.6925 0.8322
No log 4.3684 166 0.8174 0.4016 0.8174 0.9041
No log 4.4211 168 0.7475 0.4350 0.7475 0.8646
No log 4.4737 170 0.5759 0.5267 0.5759 0.7589
No log 4.5263 172 0.5997 0.4239 0.5997 0.7744
No log 4.5789 174 0.7121 0.3665 0.7121 0.8439
No log 4.6316 176 0.6303 0.3770 0.6303 0.7939
No log 4.6842 178 0.5816 0.5248 0.5816 0.7626
No log 4.7368 180 0.6549 0.5178 0.6549 0.8093
No log 4.7895 182 0.7052 0.4289 0.7052 0.8398
No log 4.8421 184 0.7277 0.4032 0.7277 0.8531
No log 4.8947 186 0.6919 0.4686 0.6919 0.8318
No log 4.9474 188 0.5745 0.6247 0.5745 0.7580
No log 5.0 190 0.5611 0.6247 0.5611 0.7491
No log 5.0526 192 0.5963 0.6104 0.5963 0.7722
No log 5.1053 194 0.7655 0.3015 0.7655 0.8749
No log 5.1579 196 0.9030 0.2981 0.9030 0.9503
No log 5.2105 198 0.8672 0.3009 0.8672 0.9312
No log 5.2632 200 0.7111 0.4096 0.7111 0.8433
No log 5.3158 202 0.5563 0.6409 0.5563 0.7459
No log 5.3684 204 0.5341 0.5625 0.5341 0.7308
No log 5.4211 206 0.5384 0.6222 0.5384 0.7338
No log 5.4737 208 0.6314 0.4598 0.6314 0.7946
No log 5.5263 210 0.8034 0.4016 0.8034 0.8963
No log 5.5789 212 0.9973 0.3717 0.9973 0.9987
No log 5.6316 214 1.1059 0.1913 1.1059 1.0516
No log 5.6842 216 1.0042 0.3292 1.0042 1.0021
No log 5.7368 218 0.7482 0.4651 0.7482 0.8650
No log 5.7895 220 0.5597 0.6308 0.5597 0.7481
No log 5.8421 222 0.5466 0.5812 0.5466 0.7394
No log 5.8947 224 0.5559 0.6396 0.5559 0.7456
No log 5.9474 226 0.6277 0.5455 0.6277 0.7923
No log 6.0 228 0.6425 0.4949 0.6425 0.8015
No log 6.0526 230 0.6431 0.5455 0.6431 0.8019
No log 6.1053 232 0.5847 0.5406 0.5847 0.7647
No log 6.1579 234 0.5769 0.4992 0.5769 0.7595
No log 6.2105 236 0.5807 0.4771 0.5807 0.7620
No log 6.2632 238 0.6106 0.5041 0.6106 0.7814
No log 6.3158 240 0.6169 0.4766 0.6169 0.7854
No log 6.3684 242 0.5877 0.6289 0.5877 0.7666
No log 6.4211 244 0.5936 0.6355 0.5936 0.7705
No log 6.4737 246 0.5753 0.5827 0.5753 0.7585
No log 6.5263 248 0.5636 0.6121 0.5636 0.7507
No log 6.5789 250 0.5842 0.5483 0.5842 0.7643
No log 6.6316 252 0.5869 0.5483 0.5869 0.7661
No log 6.6842 254 0.5568 0.5687 0.5568 0.7462
No log 6.7368 256 0.5419 0.5617 0.5419 0.7361
No log 6.7895 258 0.5404 0.5831 0.5404 0.7351
No log 6.8421 260 0.5498 0.6183 0.5498 0.7415
No log 6.8947 262 0.5693 0.6034 0.5693 0.7545
No log 6.9474 264 0.5441 0.6395 0.5441 0.7376
No log 7.0 266 0.5561 0.5580 0.5561 0.7457
No log 7.0526 268 0.5582 0.5525 0.5582 0.7471
No log 7.1053 270 0.5354 0.6046 0.5354 0.7317
No log 7.1579 272 0.7512 0.4204 0.7512 0.8667
No log 7.2105 274 0.9135 0.3671 0.9135 0.9558
No log 7.2632 276 0.7755 0.4116 0.7755 0.8806
No log 7.3158 278 0.5842 0.6390 0.5842 0.7643
No log 7.3684 280 0.5085 0.6245 0.5085 0.7131
No log 7.4211 282 0.5159 0.6210 0.5159 0.7183
No log 7.4737 284 0.5924 0.4842 0.5924 0.7697
No log 7.5263 286 0.6158 0.4842 0.6158 0.7847
No log 7.5789 288 0.5610 0.5512 0.5610 0.7490
No log 7.6316 290 0.5217 0.6271 0.5217 0.7223
No log 7.6842 292 0.5179 0.6377 0.5179 0.7196
No log 7.7368 294 0.5414 0.5528 0.5414 0.7358
No log 7.7895 296 0.5503 0.5840 0.5503 0.7418
No log 7.8421 298 0.5510 0.6484 0.5510 0.7423
No log 7.8947 300 0.5925 0.5873 0.5925 0.7697
No log 7.9474 302 0.6365 0.5252 0.6365 0.7978
No log 8.0 304 0.5866 0.5636 0.5866 0.7659
No log 8.0526 306 0.5397 0.6344 0.5397 0.7346
No log 8.1053 308 0.5435 0.5657 0.5435 0.7372
No log 8.1579 310 0.5492 0.5657 0.5492 0.7410
No log 8.2105 312 0.5592 0.5065 0.5592 0.7478
No log 8.2632 314 0.5610 0.5028 0.5610 0.7490
No log 8.3158 316 0.5487 0.6183 0.5487 0.7408
No log 8.3684 318 0.5264 0.6636 0.5264 0.7255
No log 8.4211 320 0.5275 0.6339 0.5275 0.7263
No log 8.4737 322 0.5315 0.6650 0.5315 0.7290
No log 8.5263 324 0.6080 0.5251 0.6080 0.7797
No log 8.5789 326 0.6813 0.4756 0.6813 0.8254
No log 8.6316 328 0.6213 0.5696 0.6213 0.7882
No log 8.6842 330 0.5282 0.6667 0.5282 0.7268
No log 8.7368 332 0.5183 0.6455 0.5183 0.7199
No log 8.7895 334 0.5201 0.6339 0.5201 0.7212
No log 8.8421 336 0.5241 0.6339 0.5241 0.7240
No log 8.8947 338 0.5299 0.6747 0.5299 0.7280
No log 8.9474 340 0.5414 0.6087 0.5414 0.7358
No log 9.0 342 0.5347 0.6371 0.5347 0.7312
No log 9.0526 344 0.5312 0.6286 0.5312 0.7289
No log 9.1053 346 0.5310 0.5936 0.5310 0.7287
No log 9.1579 348 0.5260 0.6286 0.5260 0.7253
No log 9.2105 350 0.5384 0.6197 0.5384 0.7338
No log 9.2632 352 0.5470 0.6295 0.5470 0.7396
No log 9.3158 354 0.5157 0.6068 0.5157 0.7181
No log 9.3684 356 0.5041 0.6156 0.5041 0.7100
No log 9.4211 358 0.5048 0.6156 0.5048 0.7105
No log 9.4737 360 0.5203 0.6068 0.5203 0.7213
No log 9.5263 362 0.6179 0.5266 0.6179 0.7861
No log 9.5789 364 0.6280 0.5002 0.6280 0.7925
No log 9.6316 366 0.5847 0.5538 0.5847 0.7646
No log 9.6842 368 0.5241 0.6269 0.5241 0.7239
No log 9.7368 370 0.5088 0.5767 0.5088 0.7133
No log 9.7895 372 0.5246 0.5422 0.5246 0.7243
No log 9.8421 374 0.5117 0.6137 0.5117 0.7153
No log 9.8947 376 0.4998 0.5992 0.4998 0.7069
No log 9.9474 378 0.4999 0.6200 0.4999 0.7071
No log 10.0 380 0.4984 0.6215 0.4984 0.7060
No log 10.0526 382 0.4782 0.7055 0.4782 0.6915
No log 10.1053 384 0.4715 0.7055 0.4715 0.6867
No log 10.1579 386 0.4728 0.7055 0.4728 0.6876
No log 10.2105 388 0.4758 0.6724 0.4758 0.6898
No log 10.2632 390 0.4735 0.6620 0.4735 0.6881
No log 10.3158 392 0.4819 0.6770 0.4819 0.6942
No log 10.3684 394 0.5917 0.5122 0.5917 0.7692
No log 10.4211 396 0.6931 0.4350 0.6931 0.8325
No log 10.4737 398 0.6492 0.4584 0.6492 0.8058
No log 10.5263 400 0.5341 0.6367 0.5341 0.7308
No log 10.5789 402 0.5222 0.6455 0.5222 0.7226
No log 10.6316 404 0.6119 0.4969 0.6119 0.7823
No log 10.6842 406 0.6323 0.4633 0.6323 0.7952
No log 10.7368 408 0.5530 0.5195 0.5530 0.7436
No log 10.7895 410 0.4843 0.6554 0.4843 0.6959
No log 10.8421 412 0.4889 0.6782 0.4889 0.6992
No log 10.8947 414 0.5012 0.6235 0.5012 0.7080
No log 10.9474 416 0.4833 0.7236 0.4833 0.6952
No log 11.0 418 0.4547 0.7032 0.4547 0.6743
No log 11.0526 420 0.4803 0.6251 0.4803 0.6930
No log 11.1053 422 0.4969 0.6058 0.4969 0.7049
No log 11.1579 424 0.4690 0.6431 0.4690 0.6848
No log 11.2105 426 0.4706 0.6667 0.4706 0.6860
No log 11.2632 428 0.4945 0.6414 0.4945 0.7032
No log 11.3158 430 0.4781 0.6282 0.4781 0.6915
No log 11.3684 432 0.4663 0.6636 0.4663 0.6828
No log 11.4211 434 0.5281 0.5166 0.5281 0.7267
No log 11.4737 436 0.5728 0.4979 0.5728 0.7569
No log 11.5263 438 0.5410 0.5758 0.5410 0.7356
No log 11.5789 440 0.4846 0.6730 0.4846 0.6961
No log 11.6316 442 0.4821 0.7114 0.4821 0.6943
No log 11.6842 444 0.4948 0.6295 0.4948 0.7034
No log 11.7368 446 0.4887 0.6295 0.4887 0.6991
No log 11.7895 448 0.4733 0.6866 0.4733 0.6880
No log 11.8421 450 0.4595 0.6929 0.4595 0.6779
No log 11.8947 452 0.4653 0.6830 0.4653 0.6821
No log 11.9474 454 0.4634 0.6830 0.4634 0.6807
No log 12.0 456 0.4739 0.7052 0.4739 0.6884
No log 12.0526 458 0.4947 0.7319 0.4947 0.7033
No log 12.1053 460 0.4811 0.7052 0.4811 0.6936
No log 12.1579 462 0.4730 0.6723 0.4730 0.6878
No log 12.2105 464 0.4803 0.6699 0.4803 0.6930
No log 12.2632 466 0.4842 0.6372 0.4842 0.6958
No log 12.3158 468 0.4737 0.6985 0.4737 0.6882
No log 12.3684 470 0.4667 0.7158 0.4667 0.6832
No log 12.4211 472 0.4893 0.6598 0.4893 0.6995
No log 12.4737 474 0.5353 0.5603 0.5353 0.7316
No log 12.5263 476 0.5661 0.5251 0.5661 0.7524
No log 12.5789 478 0.5413 0.5160 0.5413 0.7358
No log 12.6316 480 0.4884 0.6279 0.4884 0.6989
No log 12.6842 482 0.4789 0.6830 0.4789 0.6921
No log 12.7368 484 0.4948 0.5836 0.4948 0.7034
No log 12.7895 486 0.4996 0.5467 0.4996 0.7068
No log 12.8421 488 0.5019 0.5648 0.5019 0.7085
No log 12.8947 490 0.5170 0.5493 0.5170 0.7190
No log 12.9474 492 0.5612 0.4908 0.5612 0.7492
No log 13.0 494 0.5772 0.4886 0.5772 0.7597
No log 13.0526 496 0.5695 0.5095 0.5695 0.7546
No log 13.1053 498 0.5509 0.6156 0.5509 0.7422
0.3108 13.1579 500 0.5263 0.5929 0.5263 0.7255
0.3108 13.2105 502 0.5086 0.5902 0.5086 0.7132
0.3108 13.2632 504 0.5036 0.5970 0.5036 0.7097
0.3108 13.3158 506 0.5067 0.5731 0.5067 0.7118
0.3108 13.3684 508 0.5127 0.4958 0.5127 0.7160
0.3108 13.4211 510 0.4954 0.5731 0.4954 0.7038
0.3108 13.4737 512 0.4856 0.6170 0.4856 0.6969
0.3108 13.5263 514 0.4777 0.6739 0.4777 0.6911
0.3108 13.5789 516 0.4733 0.6365 0.4733 0.6879
0.3108 13.6316 518 0.4716 0.6839 0.4716 0.6867
0.3108 13.6842 520 0.4910 0.6683 0.4910 0.7007
0.3108 13.7368 522 0.5578 0.5528 0.5578 0.7468
0.3108 13.7895 524 0.6270 0.4943 0.6270 0.7918
0.3108 13.8421 526 0.6440 0.4884 0.6440 0.8025
0.3108 13.8947 528 0.6301 0.4884 0.6301 0.7938
0.3108 13.9474 530 0.5951 0.5328 0.5951 0.7714

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k11_task7_organization

Finetuned
(4019)
this model