ArabicNewSplits6_FineTuningAraBERTFreeze_run2_AugV5_k18_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8552
  • Qwk: 0.6059
  • Mse: 0.8552
  • Rmse: 0.9248

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0444 2 7.8663 -0.0474 7.8663 2.8047
No log 0.0889 4 5.5542 -0.0361 5.5542 2.3567
No log 0.1333 6 3.9606 0.0258 3.9606 1.9901
No log 0.1778 8 3.0034 0.0964 3.0034 1.7330
No log 0.2222 10 2.2876 0.1582 2.2876 1.5125
No log 0.2667 12 1.7455 0.1313 1.7455 1.3212
No log 0.3111 14 1.4522 0.2132 1.4522 1.2051
No log 0.3556 16 1.2678 0.2050 1.2678 1.1260
No log 0.4 18 1.1163 0.3079 1.1163 1.0566
No log 0.4444 20 1.0272 0.3490 1.0272 1.0135
No log 0.4889 22 0.9664 0.4017 0.9664 0.9830
No log 0.5333 24 1.0665 0.4403 1.0665 1.0327
No log 0.5778 26 1.2031 0.3506 1.2031 1.0969
No log 0.6222 28 1.1792 0.3933 1.1792 1.0859
No log 0.6667 30 1.1531 0.3876 1.1531 1.0738
No log 0.7111 32 1.0404 0.4516 1.0404 1.0200
No log 0.7556 34 0.8866 0.5191 0.8866 0.9416
No log 0.8 36 0.8072 0.5493 0.8072 0.8984
No log 0.8444 38 0.7949 0.5708 0.7949 0.8915
No log 0.8889 40 0.8030 0.5567 0.8030 0.8961
No log 0.9333 42 0.9037 0.4874 0.9037 0.9506
No log 0.9778 44 1.1521 0.4527 1.1521 1.0733
No log 1.0222 46 1.3467 0.4109 1.3467 1.1605
No log 1.0667 48 1.3188 0.4165 1.3188 1.1484
No log 1.1111 50 1.0869 0.4727 1.0869 1.0426
No log 1.1556 52 0.8496 0.5635 0.8496 0.9217
No log 1.2 54 0.7877 0.6236 0.7877 0.8875
No log 1.2444 56 0.8374 0.6057 0.8374 0.9151
No log 1.2889 58 0.9548 0.5413 0.9548 0.9771
No log 1.3333 60 1.0481 0.5260 1.0481 1.0238
No log 1.3778 62 1.1655 0.4847 1.1655 1.0796
No log 1.4222 64 1.0606 0.5148 1.0606 1.0298
No log 1.4667 66 0.8977 0.5688 0.8977 0.9475
No log 1.5111 68 0.8579 0.5940 0.8579 0.9262
No log 1.5556 70 0.9193 0.5710 0.9193 0.9588
No log 1.6 72 0.9976 0.5490 0.9976 0.9988
No log 1.6444 74 0.9621 0.5579 0.9621 0.9809
No log 1.6889 76 0.9693 0.5361 0.9694 0.9846
No log 1.7333 78 1.0506 0.5322 1.0506 1.0250
No log 1.7778 80 0.9995 0.5361 0.9995 0.9998
No log 1.8222 82 0.8418 0.6211 0.8418 0.9175
No log 1.8667 84 0.7149 0.6382 0.7149 0.8455
No log 1.9111 86 0.7010 0.6543 0.7010 0.8373
No log 1.9556 88 0.7892 0.6119 0.7892 0.8884
No log 2.0 90 0.9322 0.5773 0.9322 0.9655
No log 2.0444 92 0.9442 0.5607 0.9442 0.9717
No log 2.0889 94 0.8228 0.6164 0.8228 0.9071
No log 2.1333 96 0.6403 0.6821 0.6403 0.8002
No log 2.1778 98 0.6001 0.6917 0.6001 0.7747
No log 2.2222 100 0.6188 0.7029 0.6188 0.7866
No log 2.2667 102 0.5975 0.6845 0.5975 0.7730
No log 2.3111 104 0.6403 0.6707 0.6403 0.8002
No log 2.3556 106 0.7617 0.6302 0.7617 0.8728
No log 2.4 108 0.8343 0.6124 0.8343 0.9134
No log 2.4444 110 0.7950 0.6074 0.7950 0.8917
No log 2.4889 112 0.6790 0.6930 0.6790 0.8240
No log 2.5333 114 0.6429 0.6871 0.6429 0.8018
No log 2.5778 116 0.6512 0.7011 0.6512 0.8069
No log 2.6222 118 0.7067 0.6856 0.7067 0.8407
No log 2.6667 120 0.8775 0.6324 0.8775 0.9368
No log 2.7111 122 1.2507 0.4960 1.2507 1.1183
No log 2.7556 124 1.4299 0.4628 1.4299 1.1958
No log 2.8 126 1.2925 0.4976 1.2925 1.1369
No log 2.8444 128 0.9650 0.6170 0.9650 0.9823
No log 2.8889 130 0.8596 0.6478 0.8596 0.9271
No log 2.9333 132 0.7239 0.7039 0.7239 0.8508
No log 2.9778 134 0.6777 0.6883 0.6777 0.8233
No log 3.0222 136 0.7245 0.6937 0.7245 0.8512
No log 3.0667 138 0.7604 0.6899 0.7604 0.8720
No log 3.1111 140 0.6963 0.7030 0.6963 0.8344
No log 3.1556 142 0.6373 0.6886 0.6373 0.7983
No log 3.2 144 0.6352 0.6886 0.6352 0.7970
No log 3.2444 146 0.6858 0.7158 0.6858 0.8281
No log 3.2889 148 0.7854 0.6742 0.7854 0.8862
No log 3.3333 150 0.7603 0.6883 0.7603 0.8720
No log 3.3778 152 0.7494 0.6883 0.7494 0.8657
No log 3.4222 154 0.7088 0.6985 0.7088 0.8419
No log 3.4667 156 0.7048 0.6916 0.7048 0.8395
No log 3.5111 158 0.7174 0.6696 0.7174 0.8470
No log 3.5556 160 0.7779 0.6708 0.7779 0.8820
No log 3.6 162 0.8317 0.6450 0.8317 0.9120
No log 3.6444 164 0.7869 0.6419 0.7869 0.8871
No log 3.6889 166 0.7086 0.6896 0.7086 0.8418
No log 3.7333 168 0.6820 0.7058 0.6820 0.8259
No log 3.7778 170 0.6904 0.6935 0.6904 0.8309
No log 3.8222 172 0.7857 0.6553 0.7857 0.8864
No log 3.8667 174 0.8271 0.6420 0.8271 0.9095
No log 3.9111 176 0.8242 0.6474 0.8242 0.9079
No log 3.9556 178 0.7235 0.6938 0.7235 0.8506
No log 4.0 180 0.6986 0.7041 0.6986 0.8358
No log 4.0444 182 0.6757 0.6989 0.6757 0.8220
No log 4.0889 184 0.6800 0.7174 0.6800 0.8246
No log 4.1333 186 0.6344 0.7140 0.6344 0.7965
No log 4.1778 188 0.6262 0.7239 0.6262 0.7913
No log 4.2222 190 0.6614 0.7112 0.6614 0.8132
No log 4.2667 192 0.6389 0.7258 0.6389 0.7993
No log 4.3111 194 0.6203 0.7264 0.6203 0.7876
No log 4.3556 196 0.6307 0.7208 0.6307 0.7942
No log 4.4 198 0.6335 0.6965 0.6335 0.7959
No log 4.4444 200 0.6500 0.6965 0.6500 0.8062
No log 4.4889 202 0.6636 0.7089 0.6636 0.8146
No log 4.5333 204 0.6531 0.7062 0.6531 0.8082
No log 4.5778 206 0.6386 0.7185 0.6386 0.7991
No log 4.6222 208 0.6298 0.7141 0.6298 0.7936
No log 4.6667 210 0.6259 0.7015 0.6259 0.7911
No log 4.7111 212 0.6442 0.6971 0.6442 0.8026
No log 4.7556 214 0.6808 0.6967 0.6808 0.8251
No log 4.8 216 0.6440 0.724 0.6440 0.8025
No log 4.8444 218 0.6755 0.7300 0.6755 0.8219
No log 4.8889 220 0.6909 0.7300 0.6909 0.8312
No log 4.9333 222 0.6768 0.7101 0.6768 0.8227
No log 4.9778 224 0.6861 0.7100 0.6861 0.8283
No log 5.0222 226 0.7023 0.7100 0.7023 0.8381
No log 5.0667 228 0.7144 0.6989 0.7144 0.8452
No log 5.1111 230 0.6936 0.7060 0.6936 0.8328
No log 5.1556 232 0.6972 0.7021 0.6972 0.8350
No log 5.2 234 0.7559 0.6476 0.7559 0.8694
No log 5.2444 236 0.8697 0.6370 0.8697 0.9326
No log 5.2889 238 0.8297 0.6258 0.8297 0.9109
No log 5.3333 240 0.7455 0.6350 0.7455 0.8634
No log 5.3778 242 0.6995 0.7155 0.6995 0.8364
No log 5.4222 244 0.7237 0.6514 0.7237 0.8507
No log 5.4667 246 0.7663 0.6418 0.7663 0.8754
No log 5.5111 248 0.7865 0.6508 0.7865 0.8869
No log 5.5556 250 0.7860 0.6396 0.7860 0.8866
No log 5.6 252 0.7672 0.6529 0.7672 0.8759
No log 5.6444 254 0.7032 0.6590 0.7032 0.8386
No log 5.6889 256 0.6712 0.7063 0.6712 0.8193
No log 5.7333 258 0.6613 0.7172 0.6613 0.8132
No log 5.7778 260 0.6940 0.7156 0.6940 0.8330
No log 5.8222 262 0.6851 0.7082 0.6851 0.8277
No log 5.8667 264 0.6936 0.7121 0.6936 0.8328
No log 5.9111 266 0.7108 0.6979 0.7108 0.8431
No log 5.9556 268 0.7232 0.7082 0.7232 0.8504
No log 6.0 270 0.7212 0.6913 0.7212 0.8492
No log 6.0444 272 0.7373 0.6972 0.7373 0.8586
No log 6.0889 274 0.7402 0.6972 0.7402 0.8603
No log 6.1333 276 0.7457 0.6972 0.7457 0.8635
No log 6.1778 278 0.7723 0.6963 0.7723 0.8788
No log 6.2222 280 0.7444 0.6984 0.7444 0.8628
No log 6.2667 282 0.7602 0.7187 0.7602 0.8719
No log 6.3111 284 0.8068 0.6859 0.8068 0.8982
No log 6.3556 286 0.8707 0.6338 0.8707 0.9331
No log 6.4 288 0.8074 0.6859 0.8074 0.8985
No log 6.4444 290 0.7850 0.6912 0.7850 0.8860
No log 6.4889 292 0.7938 0.6919 0.7938 0.8909
No log 6.5333 294 0.7949 0.6819 0.7949 0.8916
No log 6.5778 296 0.7988 0.6757 0.7988 0.8938
No log 6.6222 298 0.8014 0.6647 0.8014 0.8952
No log 6.6667 300 0.7693 0.6930 0.7693 0.8771
No log 6.7111 302 0.7422 0.6906 0.7422 0.8615
No log 6.7556 304 0.7587 0.7047 0.7587 0.8710
No log 6.8 306 0.8154 0.6594 0.8154 0.9030
No log 6.8444 308 0.7499 0.6793 0.7499 0.8660
No log 6.8889 310 0.7419 0.6816 0.7419 0.8613
No log 6.9333 312 0.7744 0.6750 0.7744 0.8800
No log 6.9778 314 0.7204 0.6684 0.7204 0.8488
No log 7.0222 316 0.7056 0.6512 0.7056 0.8400
No log 7.0667 318 0.7347 0.6386 0.7347 0.8571
No log 7.1111 320 0.7246 0.6541 0.7246 0.8512
No log 7.1556 322 0.8145 0.6215 0.8145 0.9025
No log 7.2 324 0.8572 0.5710 0.8572 0.9258
No log 7.2444 326 0.7945 0.6472 0.7945 0.8913
No log 7.2889 328 0.7859 0.6555 0.7859 0.8865
No log 7.3333 330 0.8752 0.6293 0.8752 0.9355
No log 7.3778 332 0.8556 0.6337 0.8556 0.9250
No log 7.4222 334 0.7518 0.6628 0.7518 0.8671
No log 7.4667 336 0.7306 0.6620 0.7306 0.8547
No log 7.5111 338 0.7253 0.6479 0.7253 0.8517
No log 7.5556 340 0.7270 0.6413 0.7270 0.8526
No log 7.6 342 0.7465 0.6294 0.7465 0.8640
No log 7.6444 344 0.7384 0.6362 0.7384 0.8593
No log 7.6889 346 0.7348 0.6602 0.7348 0.8572
No log 7.7333 348 0.7781 0.7091 0.7781 0.8821
No log 7.7778 350 0.8433 0.6413 0.8433 0.9183
No log 7.8222 352 0.8499 0.6289 0.8499 0.9219
No log 7.8667 354 0.7877 0.6913 0.7877 0.8875
No log 7.9111 356 0.7428 0.6609 0.7428 0.8619
No log 7.9556 358 0.7597 0.6325 0.7597 0.8716
No log 8.0 360 0.8104 0.6182 0.8104 0.9002
No log 8.0444 362 0.7613 0.6257 0.7613 0.8725
No log 8.0889 364 0.7443 0.6836 0.7443 0.8627
No log 8.1333 366 0.7702 0.7024 0.7702 0.8776
No log 8.1778 368 0.7371 0.6983 0.7371 0.8585
No log 8.2222 370 0.7527 0.6392 0.7527 0.8676
No log 8.2667 372 0.9235 0.6040 0.9235 0.9610
No log 8.3111 374 0.9325 0.6040 0.9325 0.9657
No log 8.3556 376 0.7954 0.6460 0.7954 0.8918
No log 8.4 378 0.7496 0.6813 0.7496 0.8658
No log 8.4444 380 0.7639 0.6800 0.7639 0.8740
No log 8.4889 382 0.7611 0.6814 0.7611 0.8724
No log 8.5333 384 0.7654 0.6785 0.7654 0.8749
No log 8.5778 386 0.7722 0.6658 0.7722 0.8788
No log 8.6222 388 0.7902 0.6450 0.7902 0.8890
No log 8.6667 390 0.7953 0.6720 0.7953 0.8918
No log 8.7111 392 0.8163 0.6749 0.8163 0.9035
No log 8.7556 394 0.8234 0.6633 0.8234 0.9074
No log 8.8 396 0.8179 0.6587 0.8179 0.9044
No log 8.8444 398 0.8157 0.6587 0.8157 0.9032
No log 8.8889 400 0.8262 0.6673 0.8262 0.9089
No log 8.9333 402 0.8020 0.6549 0.8020 0.8956
No log 8.9778 404 0.7881 0.6809 0.7881 0.8877
No log 9.0222 406 0.7776 0.6845 0.7776 0.8818
No log 9.0667 408 0.7591 0.6809 0.7591 0.8713
No log 9.1111 410 0.7499 0.6508 0.7499 0.8660
No log 9.1556 412 0.7339 0.6700 0.7339 0.8567
No log 9.2 414 0.7130 0.6737 0.7130 0.8444
No log 9.2444 416 0.7071 0.6399 0.7071 0.8409
No log 9.2889 418 0.6901 0.7223 0.6901 0.8307
No log 9.3333 420 0.6903 0.7064 0.6903 0.8309
No log 9.3778 422 0.6958 0.6608 0.6958 0.8342
No log 9.4222 424 0.7052 0.6554 0.7052 0.8398
No log 9.4667 426 0.7152 0.6554 0.7152 0.8457
No log 9.5111 428 0.7311 0.7191 0.7311 0.8551
No log 9.5556 430 0.8187 0.6441 0.8187 0.9048
No log 9.6 432 0.8858 0.6340 0.8858 0.9411
No log 9.6444 434 0.8406 0.6469 0.8406 0.9168
No log 9.6889 436 0.7573 0.6930 0.7573 0.8702
No log 9.7333 438 0.7857 0.6123 0.7857 0.8864
No log 9.7778 440 0.8011 0.6194 0.8011 0.8951
No log 9.8222 442 0.7686 0.6468 0.7686 0.8767
No log 9.8667 444 0.8301 0.6572 0.8301 0.9111
No log 9.9111 446 0.8935 0.6410 0.8935 0.9452
No log 9.9556 448 0.8515 0.6473 0.8515 0.9228
No log 10.0 450 0.7831 0.6587 0.7831 0.8849
No log 10.0444 452 0.7886 0.6377 0.7886 0.8880
No log 10.0889 454 0.7886 0.6242 0.7886 0.8881
No log 10.1333 456 0.7704 0.6384 0.7704 0.8777
No log 10.1778 458 0.7609 0.6577 0.7609 0.8723
No log 10.2222 460 0.7579 0.6577 0.7579 0.8706
No log 10.2667 462 0.7564 0.6656 0.7564 0.8697
No log 10.3111 464 0.7658 0.6719 0.7658 0.8751
No log 10.3556 466 0.7667 0.6776 0.7667 0.8756
No log 10.4 468 0.7607 0.6929 0.7607 0.8722
No log 10.4444 470 0.7593 0.6675 0.7593 0.8714
No log 10.4889 472 0.7655 0.6739 0.7655 0.8749
No log 10.5333 474 0.7844 0.6678 0.7844 0.8857
No log 10.5778 476 0.7870 0.6721 0.7870 0.8871
No log 10.6222 478 0.7872 0.6861 0.7872 0.8872
No log 10.6667 480 0.7775 0.6690 0.7775 0.8817
No log 10.7111 482 0.8070 0.6369 0.8070 0.8983
No log 10.7556 484 0.8086 0.6369 0.8086 0.8992
No log 10.8 486 0.7768 0.6463 0.7768 0.8813
No log 10.8444 488 0.8055 0.6704 0.8055 0.8975
No log 10.8889 490 0.8529 0.6207 0.8529 0.9235
No log 10.9333 492 0.8183 0.6664 0.8183 0.9046
No log 10.9778 494 0.7735 0.6639 0.7735 0.8795
No log 11.0222 496 0.8478 0.6062 0.8478 0.9208
No log 11.0667 498 0.8959 0.6078 0.8959 0.9465
0.5847 11.1111 500 0.8342 0.6090 0.8342 0.9133
0.5847 11.1556 502 0.7747 0.6639 0.7747 0.8802
0.5847 11.2 504 0.7890 0.6366 0.7890 0.8882
0.5847 11.2444 506 0.7886 0.6330 0.7886 0.8880
0.5847 11.2889 508 0.7777 0.6605 0.7777 0.8819
0.5847 11.3333 510 0.7932 0.6391 0.7932 0.8906
0.5847 11.3778 512 0.8079 0.6323 0.8079 0.8989
0.5847 11.4222 514 0.8412 0.6147 0.8412 0.9172
0.5847 11.4667 516 0.8683 0.6193 0.8683 0.9318
0.5847 11.5111 518 0.8191 0.6308 0.8191 0.9050
0.5847 11.5556 520 0.8318 0.6383 0.8318 0.9120
0.5847 11.6 522 0.8999 0.6239 0.8999 0.9486
0.5847 11.6444 524 0.9141 0.6290 0.9141 0.9561
0.5847 11.6889 526 0.8697 0.6175 0.8697 0.9326
0.5847 11.7333 528 0.8567 0.6169 0.8567 0.9256
0.5847 11.7778 530 0.8693 0.6036 0.8693 0.9324
0.5847 11.8222 532 0.8687 0.6057 0.8687 0.9320
0.5847 11.8667 534 0.8606 0.5989 0.8606 0.9277
0.5847 11.9111 536 0.8610 0.5972 0.8610 0.9279
0.5847 11.9556 538 0.8552 0.6059 0.8552 0.9248

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.1+cu121
  • Datasets 3.2.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERTFreeze_run2_AugV5_k18_task1_organization

Finetuned
(4024)
this model