ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k16_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7486
  • Qwk: 0.7170
  • Mse: 0.7486
  • Rmse: 0.8652

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0263 2 6.8139 0.0239 6.8139 2.6104
No log 0.0526 4 4.5495 0.0650 4.5495 2.1330
No log 0.0789 6 3.7296 0.0303 3.7296 1.9312
No log 0.1053 8 2.9966 0.0244 2.9966 1.7311
No log 0.1316 10 2.2571 0.2553 2.2571 1.5024
No log 0.1579 12 2.0233 0.25 2.0233 1.4224
No log 0.1842 14 1.5879 0.3304 1.5879 1.2601
No log 0.2105 16 1.6733 0.3810 1.6733 1.2936
No log 0.2368 18 3.1116 0.1451 3.1116 1.7640
No log 0.2632 20 3.6046 0.1712 3.6046 1.8986
No log 0.2895 22 2.6517 0.1556 2.6517 1.6284
No log 0.3158 24 1.6660 0.3885 1.6660 1.2907
No log 0.3421 26 1.2860 0.4286 1.2860 1.1340
No log 0.3684 28 1.3207 0.4962 1.3207 1.1492
No log 0.3947 30 1.9617 0.2930 1.9617 1.4006
No log 0.4211 32 2.5021 0.2404 2.5021 1.5818
No log 0.4474 34 2.2668 0.2955 2.2668 1.5056
No log 0.4737 36 1.6758 0.3893 1.6758 1.2945
No log 0.5 38 1.1167 0.5846 1.1167 1.0567
No log 0.5263 40 1.2536 0.4516 1.2536 1.1196
No log 0.5526 42 1.1368 0.48 1.1368 1.0662
No log 0.5789 44 1.4131 0.4690 1.4131 1.1888
No log 0.6053 46 2.6385 0.2995 2.6385 1.6244
No log 0.6316 48 3.3692 0.2066 3.3692 1.8355
No log 0.6579 50 3.3026 0.2203 3.3026 1.8173
No log 0.6842 52 2.5877 0.3037 2.5877 1.6086
No log 0.7105 54 1.3816 0.5405 1.3816 1.1754
No log 0.7368 56 0.7991 0.6667 0.7991 0.8939
No log 0.7632 58 0.7764 0.6866 0.7764 0.8812
No log 0.7895 60 0.8364 0.6519 0.8364 0.9145
No log 0.8158 62 1.2244 0.5753 1.2244 1.1065
No log 0.8421 64 1.7457 0.4444 1.7457 1.3212
No log 0.8684 66 1.8116 0.4974 1.8116 1.3460
No log 0.8947 68 1.2047 0.6489 1.2047 1.0976
No log 0.9211 70 1.0679 0.6595 1.0679 1.0334
No log 0.9474 72 1.2999 0.6170 1.2999 1.1401
No log 0.9737 74 1.4072 0.6064 1.4072 1.1863
No log 1.0 76 1.5159 0.5851 1.5159 1.2312
No log 1.0263 78 2.5503 0.3507 2.5503 1.5970
No log 1.0526 80 3.1715 0.2946 3.1715 1.7809
No log 1.0789 82 2.9152 0.3226 2.9152 1.7074
No log 1.1053 84 1.5874 0.5699 1.5874 1.2599
No log 1.1316 86 0.8130 0.6887 0.8130 0.9017
No log 1.1579 88 0.7637 0.6980 0.7637 0.8739
No log 1.1842 90 0.8903 0.6875 0.8903 0.9436
No log 1.2105 92 1.2357 0.6180 1.2357 1.1116
No log 1.2368 94 1.2270 0.5988 1.2270 1.1077
No log 1.2632 96 0.9373 0.7020 0.9373 0.9681
No log 1.2895 98 0.7870 0.6906 0.7870 0.8871
No log 1.3158 100 0.7466 0.6763 0.7466 0.8640
No log 1.3421 102 0.7488 0.6950 0.7488 0.8653
No log 1.3684 104 0.9056 0.7170 0.9056 0.9516
No log 1.3947 106 0.9616 0.6871 0.9616 0.9806
No log 1.4211 108 0.8086 0.725 0.8086 0.8992
No log 1.4474 110 0.6840 0.7651 0.6840 0.8271
No log 1.4737 112 0.6905 0.7643 0.6905 0.8310
No log 1.5 114 0.8594 0.7485 0.8594 0.9271
No log 1.5263 116 1.3639 0.5685 1.3639 1.1679
No log 1.5526 118 1.4912 0.5784 1.4912 1.2212
No log 1.5789 120 1.1959 0.6429 1.1959 1.0936
No log 1.6053 122 0.8448 0.7444 0.8448 0.9191
No log 1.6316 124 0.7570 0.7976 0.7570 0.8700
No log 1.6579 126 0.7326 0.7407 0.7326 0.8559
No log 1.6842 128 0.8219 0.6753 0.8219 0.9066
No log 1.7105 130 1.0404 0.6494 1.0404 1.0200
No log 1.7368 132 1.3898 0.5357 1.3898 1.1789
No log 1.7632 134 1.5291 0.5357 1.5291 1.2366
No log 1.7895 136 1.4643 0.5357 1.4643 1.2101
No log 1.8158 138 1.1692 0.6258 1.1692 1.0813
No log 1.8421 140 0.8400 0.7329 0.8400 0.9165
No log 1.8684 142 0.8576 0.7020 0.8576 0.9261
No log 1.8947 144 0.9619 0.6577 0.9619 0.9807
No log 1.9211 146 0.9284 0.6883 0.9284 0.9635
No log 1.9474 148 0.7065 0.7831 0.7065 0.8405
No log 1.9737 150 0.6285 0.8070 0.6285 0.7928
No log 2.0 152 0.6131 0.8070 0.6131 0.7830
No log 2.0263 154 0.6028 0.8 0.6028 0.7764
No log 2.0526 156 0.7371 0.7470 0.7371 0.8585
No log 2.0789 158 0.7548 0.7296 0.7548 0.8688
No log 2.1053 160 0.7751 0.7179 0.7751 0.8804
No log 2.1316 162 0.9418 0.6375 0.9418 0.9704
No log 2.1579 164 1.1249 0.6467 1.1249 1.0606
No log 2.1842 166 0.9398 0.6494 0.9398 0.9694
No log 2.2105 168 0.8583 0.6358 0.8583 0.9264
No log 2.2368 170 0.9399 0.675 0.9399 0.9695
No log 2.2632 172 1.1434 0.6369 1.1434 1.0693
No log 2.2895 174 1.3900 0.6211 1.3900 1.1790
No log 2.3158 176 1.2718 0.6073 1.2718 1.1278
No log 2.3421 178 0.9185 0.7079 0.9185 0.9584
No log 2.3684 180 0.6695 0.7665 0.6695 0.8182
No log 2.3947 182 0.6883 0.7929 0.6883 0.8296
No log 2.4211 184 0.7584 0.7607 0.7584 0.8709
No log 2.4474 186 0.7674 0.7578 0.7674 0.8760
No log 2.4737 188 0.6604 0.7778 0.6604 0.8126
No log 2.5 190 0.6954 0.7853 0.6954 0.8339
No log 2.5263 192 0.7502 0.7879 0.7502 0.8661
No log 2.5526 194 0.7399 0.7853 0.7399 0.8602
No log 2.5789 196 0.8757 0.7209 0.8757 0.9358
No log 2.6053 198 0.9175 0.6977 0.9175 0.9578
No log 2.6316 200 0.6275 0.8049 0.6275 0.7922
No log 2.6579 202 0.5241 0.7867 0.5241 0.7239
No log 2.6842 204 0.5379 0.7867 0.5379 0.7334
No log 2.7105 206 0.6125 0.8 0.6125 0.7827
No log 2.7368 208 0.6993 0.8114 0.6993 0.8363
No log 2.7632 210 0.6301 0.8 0.6301 0.7938
No log 2.7895 212 0.6380 0.7778 0.6380 0.7988
No log 2.8158 214 0.7005 0.7826 0.7005 0.8369
No log 2.8421 216 0.7166 0.7722 0.7166 0.8465
No log 2.8684 218 0.6716 0.7901 0.6716 0.8195
No log 2.8947 220 0.6435 0.7976 0.6435 0.8022
No log 2.9211 222 0.6211 0.7929 0.6211 0.7881
No log 2.9474 224 0.5827 0.8047 0.5827 0.7633
No log 2.9737 226 0.5650 0.7758 0.5650 0.7516
No log 3.0 228 0.5677 0.825 0.5677 0.7535
No log 3.0263 230 0.6158 0.8293 0.6158 0.7847
No log 3.0526 232 0.6671 0.8 0.6671 0.8168
No log 3.0789 234 0.6541 0.7925 0.6541 0.8088
No log 3.1053 236 0.6995 0.7453 0.6995 0.8363
No log 3.1316 238 0.7361 0.7738 0.7361 0.8580
No log 3.1579 240 0.6938 0.7836 0.6938 0.8330
No log 3.1842 242 0.6787 0.8092 0.6787 0.8238
No log 3.2105 244 0.5687 0.8144 0.5687 0.7541
No log 3.2368 246 0.5955 0.7643 0.5955 0.7717
No log 3.2632 248 0.6973 0.7625 0.6973 0.8351
No log 3.2895 250 0.8179 0.7349 0.8179 0.9044
No log 3.3158 252 0.8524 0.7262 0.8524 0.9233
No log 3.3421 254 0.7201 0.7485 0.7201 0.8486
No log 3.3684 256 0.6254 0.76 0.6254 0.7908
No log 3.3947 258 0.6260 0.7153 0.6260 0.7912
No log 3.4211 260 0.6046 0.7532 0.6046 0.7776
No log 3.4474 262 0.7193 0.7805 0.7193 0.8481
No log 3.4737 264 1.1123 0.6592 1.1123 1.0546
No log 3.5 266 1.2431 0.6203 1.2431 1.1149
No log 3.5263 268 0.9611 0.6813 0.9611 0.9804
No log 3.5526 270 0.6251 0.8353 0.6251 0.7906
No log 3.5789 272 0.6482 0.7564 0.6482 0.8051
No log 3.6053 274 0.7157 0.7248 0.7157 0.8460
No log 3.6316 276 0.7222 0.7059 0.7222 0.8498
No log 3.6579 278 0.6714 0.7805 0.6714 0.8194
No log 3.6842 280 0.7464 0.7865 0.7464 0.8639
No log 3.7105 282 0.8891 0.7119 0.8891 0.9429
No log 3.7368 284 0.9347 0.6818 0.9347 0.9668
No log 3.7632 286 0.8418 0.7283 0.8418 0.9175
No log 3.7895 288 0.6637 0.8235 0.6637 0.8147
No log 3.8158 290 0.6030 0.8121 0.6030 0.7765
No log 3.8421 292 0.5799 0.8 0.5799 0.7615
No log 3.8684 294 0.5422 0.8193 0.5422 0.7363
No log 3.8947 296 0.5942 0.8268 0.5942 0.7708
No log 3.9211 298 0.6779 0.8043 0.6779 0.8234
No log 3.9474 300 0.6550 0.7869 0.6550 0.8093
No log 3.9737 302 0.6288 0.8140 0.6288 0.7930
No log 4.0 304 0.6732 0.7857 0.6732 0.8205
No log 4.0263 306 0.7014 0.7811 0.7014 0.8375
No log 4.0526 308 0.7202 0.7719 0.7202 0.8487
No log 4.0789 310 0.7662 0.7263 0.7662 0.8753
No log 4.1053 312 0.7186 0.7797 0.7186 0.8477
No log 4.1316 314 0.6563 0.8256 0.6563 0.8101
No log 4.1579 316 0.7066 0.8118 0.7066 0.8406
No log 4.1842 318 0.7249 0.7619 0.7249 0.8514
No log 4.2105 320 0.7131 0.7619 0.7131 0.8444
No log 4.2368 322 0.6966 0.7976 0.6966 0.8346
No log 4.2632 324 0.6832 0.8166 0.6832 0.8266
No log 4.2895 326 0.7064 0.7976 0.7064 0.8405
No log 4.3158 328 0.7407 0.7784 0.7407 0.8606
No log 4.3421 330 0.8175 0.7514 0.8175 0.9042
No log 4.3684 332 0.8600 0.7176 0.8600 0.9274
No log 4.3947 334 0.7888 0.7771 0.7888 0.8881
No log 4.4211 336 0.6659 0.8276 0.6659 0.8160
No log 4.4474 338 0.6348 0.8343 0.6348 0.7967
No log 4.4737 340 0.6563 0.7935 0.6563 0.8101
No log 4.5 342 0.7099 0.7897 0.7099 0.8426
No log 4.5263 344 0.6384 0.8132 0.6384 0.7990
No log 4.5526 346 0.7040 0.7978 0.7040 0.8390
No log 4.5789 348 0.7117 0.8046 0.7117 0.8436
No log 4.6053 350 0.6538 0.7879 0.6538 0.8086
No log 4.6316 352 0.7033 0.7625 0.7033 0.8386
No log 4.6579 354 0.8742 0.6905 0.8742 0.9350
No log 4.6842 356 1.0551 0.6776 1.0551 1.0272
No log 4.7105 358 1.0342 0.6667 1.0342 1.0170
No log 4.7368 360 1.0933 0.6667 1.0933 1.0456
No log 4.7632 362 1.2170 0.6633 1.2170 1.1032
No log 4.7895 364 0.9602 0.7083 0.9602 0.9799
No log 4.8158 366 0.6920 0.7953 0.6920 0.8319
No log 4.8421 368 0.7080 0.7857 0.7080 0.8414
No log 4.8684 370 0.7495 0.7407 0.7495 0.8657
No log 4.8947 372 0.7651 0.7179 0.7651 0.8747
No log 4.9211 374 0.8991 0.7152 0.8991 0.9482
No log 4.9474 376 0.9924 0.6813 0.9924 0.9962
No log 4.9737 378 1.0107 0.6703 1.0107 1.0053
No log 5.0 380 0.7935 0.7416 0.7935 0.8908
No log 5.0263 382 0.5794 0.8304 0.5794 0.7612
No log 5.0526 384 0.5746 0.8166 0.5746 0.7580
No log 5.0789 386 0.6447 0.8287 0.6447 0.8029
No log 5.1053 388 0.8184 0.7784 0.8184 0.9047
No log 5.1316 390 0.9428 0.7111 0.9428 0.9710
No log 5.1579 392 0.8837 0.7168 0.8837 0.9400
No log 5.1842 394 0.7764 0.7831 0.7764 0.8811
No log 5.2105 396 0.6562 0.7901 0.6562 0.8100
No log 5.2368 398 0.6278 0.75 0.6278 0.7924
No log 5.2632 400 0.7019 0.7436 0.7019 0.8378
No log 5.2895 402 0.8424 0.6792 0.8424 0.9178
No log 5.3158 404 0.8105 0.75 0.8105 0.9003
No log 5.3421 406 0.6646 0.7711 0.6646 0.8152
No log 5.3684 408 0.5389 0.7843 0.5389 0.7341
No log 5.3947 410 0.5269 0.8052 0.5269 0.7259
No log 5.4211 412 0.5339 0.8098 0.5339 0.7307
No log 5.4474 414 0.6472 0.8249 0.6472 0.8045
No log 5.4737 416 0.7652 0.7614 0.7652 0.8747
No log 5.5 418 0.7308 0.7797 0.7308 0.8549
No log 5.5263 420 0.6945 0.7727 0.6945 0.8334
No log 5.5526 422 0.6944 0.7861 0.6944 0.8333
No log 5.5789 424 0.6038 0.8 0.6038 0.7771
No log 5.6053 426 0.5720 0.8 0.5720 0.7563
No log 5.6316 428 0.5873 0.8 0.5873 0.7663
No log 5.6579 430 0.6112 0.8 0.6112 0.7818
No log 5.6842 432 0.6627 0.8276 0.6627 0.8141
No log 5.7105 434 0.6469 0.8276 0.6469 0.8043
No log 5.7368 436 0.6365 0.8276 0.6365 0.7978
No log 5.7632 438 0.6216 0.8391 0.6216 0.7884
No log 5.7895 440 0.6472 0.8439 0.6472 0.8045
No log 5.8158 442 0.6581 0.8439 0.6581 0.8112
No log 5.8421 444 0.7107 0.8391 0.7107 0.8430
No log 5.8684 446 0.7277 0.8391 0.7277 0.8530
No log 5.8947 448 0.7513 0.7912 0.7513 0.8668
No log 5.9211 450 0.7056 0.8 0.7056 0.8400
No log 5.9474 452 0.6275 0.8111 0.6275 0.7921
No log 5.9737 454 0.5408 0.8457 0.5408 0.7354
No log 6.0 456 0.5264 0.8457 0.5264 0.7255
No log 6.0263 458 0.5684 0.8315 0.5684 0.7539
No log 6.0526 460 0.6473 0.8 0.6473 0.8046
No log 6.0789 462 0.6968 0.8 0.6968 0.8347
No log 6.1053 464 0.6603 0.8276 0.6603 0.8126
No log 6.1316 466 0.6297 0.8391 0.6297 0.7936
No log 6.1579 468 0.6462 0.8121 0.6462 0.8039
No log 6.1842 470 0.6623 0.8171 0.6623 0.8138
No log 6.2105 472 0.7091 0.8050 0.7091 0.8421
No log 6.2368 474 0.7252 0.7848 0.7252 0.8516
No log 6.2632 476 0.6848 0.825 0.6848 0.8275
No log 6.2895 478 0.6310 0.8098 0.6310 0.7944
No log 6.3158 480 0.6350 0.8353 0.6350 0.7969
No log 6.3421 482 0.6453 0.8249 0.6453 0.8033
No log 6.3684 484 0.5992 0.8391 0.5992 0.7741
No log 6.3947 486 0.5343 0.8353 0.5343 0.7310
No log 6.4211 488 0.5342 0.8284 0.5342 0.7309
No log 6.4474 490 0.6059 0.8304 0.6059 0.7784
No log 6.4737 492 0.6734 0.7578 0.6734 0.8206
No log 6.5 494 0.6621 0.7974 0.6621 0.8137
No log 6.5263 496 0.6749 0.7799 0.6749 0.8215
No log 6.5526 498 0.6811 0.7976 0.6811 0.8253
0.4019 6.5789 500 0.6985 0.8046 0.6985 0.8358
0.4019 6.6053 502 0.6386 0.8182 0.6386 0.7991
0.4019 6.6316 504 0.5864 0.8391 0.5864 0.7658
0.4019 6.6579 506 0.5607 0.8214 0.5607 0.7488
0.4019 6.6842 508 0.5494 0.8158 0.5494 0.7412
0.4019 6.7105 510 0.5668 0.8027 0.5668 0.7528
0.4019 6.7368 512 0.5799 0.7681 0.5799 0.7615
0.4019 6.7632 514 0.5960 0.8 0.5960 0.7720
0.4019 6.7895 516 0.6854 0.7843 0.6854 0.8279
0.4019 6.8158 518 0.7486 0.7170 0.7486 0.8652

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k16_task1_organization

Finetuned
(4019)
this model