ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k7_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5637
  • Qwk: 0.6578
  • Mse: 0.5637
  • Rmse: 0.7508

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1 2 4.0764 0.0130 4.0764 2.0190
No log 0.2 4 2.0564 -0.0252 2.0564 1.4340
No log 0.3 6 1.4209 0.0030 1.4209 1.1920
No log 0.4 8 1.0602 0.2716 1.0602 1.0297
No log 0.5 10 1.2760 0.0374 1.2760 1.1296
No log 0.6 12 1.4109 -0.0627 1.4109 1.1878
No log 0.7 14 1.2724 -0.0477 1.2724 1.1280
No log 0.8 16 1.1484 0.1028 1.1484 1.0716
No log 0.9 18 1.0733 0.2068 1.0733 1.0360
No log 1.0 20 1.0495 0.2236 1.0495 1.0245
No log 1.1 22 1.0321 0.2236 1.0321 1.0159
No log 1.2 24 0.9954 0.2697 0.9954 0.9977
No log 1.3 26 0.9358 0.3243 0.9358 0.9674
No log 1.4 28 0.9000 0.3688 0.9000 0.9487
No log 1.5 30 0.9182 0.4109 0.9182 0.9582
No log 1.6 32 0.8474 0.4585 0.8474 0.9205
No log 1.7 34 0.8921 0.3037 0.8921 0.9445
No log 1.8 36 1.1198 0.2497 1.1198 1.0582
No log 1.9 38 1.5948 -0.1370 1.5948 1.2628
No log 2.0 40 1.5142 -0.1247 1.5142 1.2305
No log 2.1 42 1.0373 0.2467 1.0373 1.0185
No log 2.2 44 0.7922 0.4218 0.7922 0.8900
No log 2.3 46 0.7808 0.4420 0.7808 0.8836
No log 2.4 48 0.7535 0.4568 0.7535 0.8680
No log 2.5 50 0.7504 0.4953 0.7504 0.8662
No log 2.6 52 0.6959 0.5235 0.6959 0.8342
No log 2.7 54 0.7635 0.5487 0.7635 0.8738
No log 2.8 56 1.0131 0.3706 1.0131 1.0065
No log 2.9 58 1.0251 0.4140 1.0251 1.0125
No log 3.0 60 0.9034 0.4681 0.9034 0.9505
No log 3.1 62 0.7010 0.4616 0.7010 0.8373
No log 3.2 64 0.6195 0.6363 0.6195 0.7871
No log 3.3 66 0.6177 0.6207 0.6177 0.7859
No log 3.4 68 0.6122 0.6690 0.6122 0.7824
No log 3.5 70 0.6054 0.6866 0.6054 0.7781
No log 3.6 72 0.5906 0.6869 0.5906 0.7685
No log 3.7 74 0.5826 0.7455 0.5826 0.7633
No log 3.8 76 0.5776 0.7404 0.5776 0.7600
No log 3.9 78 0.6178 0.6644 0.6178 0.7860
No log 4.0 80 0.5713 0.7193 0.5713 0.7559
No log 4.1 82 0.5571 0.7005 0.5571 0.7464
No log 4.2 84 0.5568 0.7399 0.5568 0.7462
No log 4.3 86 0.5813 0.7304 0.5813 0.7624
No log 4.4 88 0.5992 0.7395 0.5992 0.7741
No log 4.5 90 0.5646 0.7177 0.5646 0.7514
No log 4.6 92 0.6011 0.6210 0.6011 0.7753
No log 4.7 94 0.6215 0.6753 0.6215 0.7884
No log 4.8 96 0.5613 0.6751 0.5613 0.7492
No log 4.9 98 0.5921 0.6886 0.5921 0.7695
No log 5.0 100 0.5858 0.7396 0.5858 0.7654
No log 5.1 102 0.5766 0.7246 0.5766 0.7593
No log 5.2 104 0.5887 0.6733 0.5887 0.7673
No log 5.3 106 0.5937 0.6445 0.5937 0.7705
No log 5.4 108 0.6361 0.6296 0.6361 0.7975
No log 5.5 110 0.6987 0.6447 0.6987 0.8359
No log 5.6 112 0.6528 0.6363 0.6528 0.8080
No log 5.7 114 0.5694 0.6883 0.5694 0.7546
No log 5.8 116 0.5639 0.7199 0.5639 0.7509
No log 5.9 118 0.6260 0.6404 0.6260 0.7912
No log 6.0 120 0.6427 0.6507 0.6427 0.8017
No log 6.1 122 0.6769 0.6492 0.6769 0.8227
No log 6.2 124 0.5774 0.6695 0.5774 0.7599
No log 6.3 126 0.5924 0.7061 0.5924 0.7697
No log 6.4 128 0.6308 0.7067 0.6308 0.7942
No log 6.5 130 0.6291 0.6505 0.6291 0.7932
No log 6.6 132 0.6357 0.6064 0.6357 0.7973
No log 6.7 134 0.6040 0.6502 0.6040 0.7772
No log 6.8 136 0.6200 0.6748 0.6200 0.7874
No log 6.9 138 0.6581 0.5964 0.6581 0.8113
No log 7.0 140 0.5903 0.6582 0.5903 0.7683
No log 7.1 142 0.5489 0.7147 0.5489 0.7409
No log 7.2 144 0.5638 0.6919 0.5638 0.7508
No log 7.3 146 0.6462 0.6404 0.6462 0.8039
No log 7.4 148 0.6732 0.6404 0.6732 0.8205
No log 7.5 150 0.5848 0.6543 0.5848 0.7647
No log 7.6 152 0.5601 0.6797 0.5601 0.7484
No log 7.7 154 0.5593 0.6998 0.5593 0.7479
No log 7.8 156 0.5549 0.6674 0.5549 0.7449
No log 7.9 158 0.8021 0.5713 0.8021 0.8956
No log 8.0 160 1.0109 0.5205 1.0109 1.0054
No log 8.1 162 0.9526 0.5771 0.9526 0.9760
No log 8.2 164 0.7017 0.6432 0.7017 0.8377
No log 8.3 166 0.5866 0.6438 0.5866 0.7659
No log 8.4 168 0.5897 0.7203 0.5897 0.7679
No log 8.5 170 0.6210 0.5994 0.6210 0.7880
No log 8.6 172 0.6562 0.6091 0.6562 0.8101
No log 8.7 174 0.6280 0.6256 0.6280 0.7925
No log 8.8 176 0.6262 0.7081 0.6262 0.7913
No log 8.9 178 0.6813 0.5912 0.6813 0.8254
No log 9.0 180 0.6982 0.5912 0.6982 0.8356
No log 9.1 182 0.6813 0.5912 0.6813 0.8254
No log 9.2 184 0.6180 0.6728 0.6180 0.7861
No log 9.3 186 0.5876 0.6262 0.5876 0.7665
No log 9.4 188 0.5699 0.6370 0.5699 0.7549
No log 9.5 190 0.5541 0.6908 0.5541 0.7444
No log 9.6 192 0.5670 0.6647 0.5670 0.7530
No log 9.7 194 0.5546 0.6990 0.5546 0.7447
No log 9.8 196 0.5317 0.7360 0.5317 0.7292
No log 9.9 198 0.5231 0.7272 0.5231 0.7232
No log 10.0 200 0.5390 0.6812 0.5390 0.7342
No log 10.1 202 0.5506 0.6854 0.5506 0.7420
No log 10.2 204 0.5306 0.7033 0.5306 0.7284
No log 10.3 206 0.5336 0.6954 0.5336 0.7305
No log 10.4 208 0.5164 0.7026 0.5164 0.7186
No log 10.5 210 0.5277 0.6240 0.5277 0.7264
No log 10.6 212 0.6989 0.6967 0.6989 0.8360
No log 10.7 214 0.7580 0.6744 0.7580 0.8707
No log 10.8 216 0.6183 0.6780 0.6183 0.7863
No log 10.9 218 0.5083 0.6764 0.5083 0.7130
No log 11.0 220 0.5199 0.7182 0.5199 0.7211
No log 11.1 222 0.5359 0.7384 0.5359 0.7321
No log 11.2 224 0.5121 0.7026 0.5121 0.7156
No log 11.3 226 0.5094 0.7064 0.5094 0.7137
No log 11.4 228 0.5159 0.7178 0.5159 0.7183
No log 11.5 230 0.5434 0.7259 0.5434 0.7371
No log 11.6 232 0.5534 0.7301 0.5534 0.7439
No log 11.7 234 0.5514 0.7450 0.5514 0.7425
No log 11.8 236 0.5388 0.7196 0.5388 0.7340
No log 11.9 238 0.5365 0.6938 0.5365 0.7325
No log 12.0 240 0.5477 0.6843 0.5477 0.7401
No log 12.1 242 0.5601 0.6525 0.5601 0.7484
No log 12.2 244 0.6086 0.6374 0.6086 0.7801
No log 12.3 246 0.7399 0.6069 0.7399 0.8601
No log 12.4 248 0.7695 0.6289 0.7695 0.8772
No log 12.5 250 0.6515 0.6737 0.6515 0.8072
No log 12.6 252 0.5463 0.7061 0.5463 0.7391
No log 12.7 254 0.5349 0.7266 0.5349 0.7314
No log 12.8 256 0.5408 0.7289 0.5408 0.7354
No log 12.9 258 0.5752 0.7191 0.5752 0.7584
No log 13.0 260 0.6254 0.6906 0.6254 0.7908
No log 13.1 262 0.7058 0.6578 0.7058 0.8401
No log 13.2 264 0.6641 0.7161 0.6641 0.8149
No log 13.3 266 0.5791 0.6476 0.5791 0.7610
No log 13.4 268 0.5964 0.6803 0.5964 0.7723
No log 13.5 270 0.5865 0.6703 0.5865 0.7658
No log 13.6 272 0.5739 0.6479 0.5739 0.7575
No log 13.7 274 0.5635 0.6327 0.5635 0.7507
No log 13.8 276 0.5611 0.7001 0.5611 0.7491
No log 13.9 278 0.5622 0.6796 0.5622 0.7498
No log 14.0 280 0.5856 0.5681 0.5856 0.7653
No log 14.1 282 0.6410 0.5471 0.6410 0.8006
No log 14.2 284 0.6481 0.5351 0.6481 0.8051
No log 14.3 286 0.5877 0.5905 0.5877 0.7666
No log 14.4 288 0.5321 0.6911 0.5321 0.7295
No log 14.5 290 0.5183 0.7450 0.5183 0.7199
No log 14.6 292 0.5099 0.6888 0.5099 0.7141
No log 14.7 294 0.5207 0.7267 0.5207 0.7216
No log 14.8 296 0.5955 0.6065 0.5955 0.7717
No log 14.9 298 0.6387 0.5770 0.6387 0.7992
No log 15.0 300 0.5836 0.6516 0.5836 0.7639
No log 15.1 302 0.5258 0.6724 0.5258 0.7251
No log 15.2 304 0.5991 0.6740 0.5991 0.7740
No log 15.3 306 0.6655 0.6170 0.6655 0.8158
No log 15.4 308 0.6281 0.6906 0.6281 0.7925
No log 15.5 310 0.5810 0.6814 0.5810 0.7622
No log 15.6 312 0.6130 0.6377 0.6130 0.7829
No log 15.7 314 0.7048 0.6344 0.7048 0.8395
No log 15.8 316 0.6632 0.6277 0.6632 0.8144
No log 15.9 318 0.5780 0.6237 0.5780 0.7603
No log 16.0 320 0.5546 0.6814 0.5546 0.7447
No log 16.1 322 0.5768 0.6762 0.5768 0.7594
No log 16.2 324 0.5747 0.6620 0.5747 0.7581
No log 16.3 326 0.5579 0.6425 0.5579 0.7469
No log 16.4 328 0.5460 0.6425 0.5460 0.7389
No log 16.5 330 0.5419 0.6715 0.5419 0.7362
No log 16.6 332 0.5416 0.6354 0.5416 0.7359
No log 16.7 334 0.5572 0.6354 0.5572 0.7465
No log 16.8 336 0.5571 0.6354 0.5571 0.7464
No log 16.9 338 0.5570 0.6460 0.5570 0.7463
No log 17.0 340 0.5613 0.6788 0.5613 0.7492
No log 17.1 342 0.5588 0.6460 0.5588 0.7475
No log 17.2 344 0.5674 0.6400 0.5674 0.7532
No log 17.3 346 0.5671 0.6292 0.5671 0.7531
No log 17.4 348 0.5511 0.6690 0.5511 0.7423
No log 17.5 350 0.5598 0.6499 0.5598 0.7482
No log 17.6 352 0.5619 0.6620 0.5619 0.7496
No log 17.7 354 0.5501 0.6690 0.5501 0.7417
No log 17.8 356 0.5785 0.6337 0.5785 0.7606
No log 17.9 358 0.5646 0.6419 0.5646 0.7514
No log 18.0 360 0.5242 0.7193 0.5242 0.7240
No log 18.1 362 0.5251 0.7132 0.5251 0.7247
No log 18.2 364 0.5324 0.7139 0.5324 0.7296
No log 18.3 366 0.5374 0.7097 0.5374 0.7331
No log 18.4 368 0.5415 0.7056 0.5415 0.7359
No log 18.5 370 0.5281 0.7132 0.5281 0.7267
No log 18.6 372 0.5261 0.7245 0.5261 0.7253
No log 18.7 374 0.5128 0.7139 0.5128 0.7161
No log 18.8 376 0.4968 0.7139 0.4968 0.7049
No log 18.9 378 0.4882 0.7034 0.4882 0.6987
No log 19.0 380 0.5024 0.7301 0.5024 0.7088
No log 19.1 382 0.5110 0.7592 0.5110 0.7148
No log 19.2 384 0.5306 0.7632 0.5306 0.7284
No log 19.3 386 0.5171 0.7592 0.5171 0.7191
No log 19.4 388 0.5278 0.7685 0.5278 0.7265
No log 19.5 390 0.5182 0.7501 0.5182 0.7199
No log 19.6 392 0.4986 0.7082 0.4986 0.7061
No log 19.7 394 0.5066 0.7245 0.5066 0.7118
No log 19.8 396 0.5422 0.6028 0.5422 0.7363
No log 19.9 398 0.5501 0.5928 0.5501 0.7417
No log 20.0 400 0.5271 0.6322 0.5271 0.7260
No log 20.1 402 0.4895 0.7355 0.4895 0.6996
No log 20.2 404 0.5097 0.7196 0.5097 0.7139
No log 20.3 406 0.5578 0.7051 0.5578 0.7469
No log 20.4 408 0.5397 0.6779 0.5397 0.7346
No log 20.5 410 0.5175 0.7132 0.5175 0.7194
No log 20.6 412 0.5510 0.5999 0.5510 0.7423
No log 20.7 414 0.5603 0.5999 0.5603 0.7485
No log 20.8 416 0.5598 0.6161 0.5598 0.7482
No log 20.9 418 0.5441 0.6476 0.5441 0.7376
No log 21.0 420 0.5295 0.6768 0.5295 0.7277
No log 21.1 422 0.5337 0.6796 0.5337 0.7306
No log 21.2 424 0.5507 0.6611 0.5507 0.7421
No log 21.3 426 0.5549 0.6611 0.5549 0.7449
No log 21.4 428 0.5169 0.6903 0.5169 0.7190
No log 21.5 430 0.5049 0.7193 0.5049 0.7105
No log 21.6 432 0.5013 0.7193 0.5013 0.7080
No log 21.7 434 0.5022 0.7193 0.5022 0.7087
No log 21.8 436 0.5015 0.7193 0.5015 0.7082
No log 21.9 438 0.5007 0.7355 0.5007 0.7076
No log 22.0 440 0.5092 0.7129 0.5092 0.7136
No log 22.1 442 0.5368 0.7133 0.5368 0.7327
No log 22.2 444 0.5283 0.7392 0.5283 0.7269
No log 22.3 446 0.4984 0.7351 0.4984 0.7060
No log 22.4 448 0.4944 0.7303 0.4944 0.7031
No log 22.5 450 0.5043 0.7351 0.5043 0.7101
No log 22.6 452 0.5150 0.7081 0.5150 0.7177
No log 22.7 454 0.5062 0.7458 0.5062 0.7115
No log 22.8 456 0.5054 0.6919 0.5054 0.7109
No log 22.9 458 0.5105 0.6919 0.5105 0.7145
No log 23.0 460 0.4995 0.7193 0.4995 0.7067
No log 23.1 462 0.4952 0.7456 0.4952 0.7037
No log 23.2 464 0.4965 0.7360 0.4965 0.7046
No log 23.3 466 0.4955 0.7077 0.4955 0.7039
No log 23.4 468 0.4931 0.6764 0.4931 0.7022
No log 23.5 470 0.5245 0.6544 0.5245 0.7243
No log 23.6 472 0.5493 0.6497 0.5493 0.7411
No log 23.7 474 0.5254 0.6312 0.5254 0.7249
No log 23.8 476 0.4952 0.6919 0.4952 0.7037
No log 23.9 478 0.4837 0.7143 0.4837 0.6955
No log 24.0 480 0.4914 0.7495 0.4914 0.7010
No log 24.1 482 0.4992 0.7189 0.4992 0.7066
No log 24.2 484 0.4977 0.7298 0.4977 0.7055
No log 24.3 486 0.5020 0.7245 0.5020 0.7085
No log 24.4 488 0.5050 0.6882 0.5050 0.7106
No log 24.5 490 0.5075 0.6882 0.5075 0.7124
No log 24.6 492 0.5125 0.6882 0.5125 0.7159
No log 24.7 494 0.5357 0.6778 0.5357 0.7319
No log 24.8 496 0.5726 0.6733 0.5726 0.7567
No log 24.9 498 0.6238 0.5895 0.6238 0.7898
0.2657 25.0 500 0.5885 0.6629 0.5885 0.7671
0.2657 25.1 502 0.5318 0.6882 0.5318 0.7292
0.2657 25.2 504 0.5353 0.7204 0.5353 0.7317
0.2657 25.3 506 0.5507 0.7253 0.5507 0.7421
0.2657 25.4 508 0.5447 0.7148 0.5447 0.7380
0.2657 25.5 510 0.5350 0.7041 0.5350 0.7315
0.2657 25.6 512 0.5337 0.6882 0.5337 0.7306
0.2657 25.7 514 0.5547 0.6871 0.5547 0.7448
0.2657 25.8 516 0.5948 0.6176 0.5948 0.7712
0.2657 25.9 518 0.6545 0.5902 0.6545 0.8090
0.2657 26.0 520 0.6985 0.6259 0.6985 0.8357
0.2657 26.1 522 0.6919 0.6105 0.6919 0.8318
0.2657 26.2 524 0.5966 0.6968 0.5966 0.7724
0.2657 26.3 526 0.5167 0.7001 0.5167 0.7188
0.2657 26.4 528 0.5292 0.6803 0.5292 0.7275
0.2657 26.5 530 0.5900 0.6226 0.5900 0.7681
0.2657 26.6 532 0.5872 0.6226 0.5872 0.7663
0.2657 26.7 534 0.5326 0.6470 0.5326 0.7298
0.2657 26.8 536 0.5161 0.6933 0.5161 0.7184
0.2657 26.9 538 0.5729 0.7167 0.5729 0.7569
0.2657 27.0 540 0.6147 0.6841 0.6147 0.7840
0.2657 27.1 542 0.6058 0.6711 0.6058 0.7783
0.2657 27.2 544 0.5845 0.6869 0.5845 0.7646
0.2657 27.3 546 0.5466 0.6815 0.5466 0.7393
0.2657 27.4 548 0.5399 0.6262 0.5399 0.7347
0.2657 27.5 550 0.5379 0.6322 0.5379 0.7334
0.2657 27.6 552 0.5420 0.6276 0.5420 0.7362
0.2657 27.7 554 0.5549 0.6276 0.5549 0.7449
0.2657 27.8 556 0.5637 0.6578 0.5637 0.7508

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k7_task5_organization

Finetuned
(4023)
this model