ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run2_AugV5_k4_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.1823
  • Qwk: 0.4656
  • Mse: 1.1823
  • Rmse: 1.0873

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0952 2 5.3364 -0.0214 5.3364 2.3101
No log 0.1905 4 3.5219 0.0579 3.5219 1.8767
No log 0.2857 6 1.9986 0.1201 1.9986 1.4137
No log 0.3810 8 1.4246 0.1209 1.4246 1.1936
No log 0.4762 10 1.2701 0.2451 1.2701 1.1270
No log 0.5714 12 1.4976 0.0391 1.4976 1.2238
No log 0.6667 14 1.2559 0.1813 1.2559 1.1207
No log 0.7619 16 1.2396 0.1372 1.2396 1.1134
No log 0.8571 18 1.2561 0.0738 1.2561 1.1207
No log 0.9524 20 1.2640 0.0738 1.2640 1.1243
No log 1.0476 22 1.1905 0.1771 1.1905 1.0911
No log 1.1429 24 1.1650 0.3037 1.1650 1.0793
No log 1.2381 26 1.2591 0.2390 1.2591 1.1221
No log 1.3333 28 1.2711 0.2054 1.2711 1.1274
No log 1.4286 30 1.1765 0.2459 1.1765 1.0847
No log 1.5238 32 1.2085 0.2330 1.2085 1.0993
No log 1.6190 34 1.1758 0.3205 1.1758 1.0843
No log 1.7143 36 1.0384 0.3983 1.0384 1.0190
No log 1.8095 38 1.0151 0.3688 1.0151 1.0075
No log 1.9048 40 1.0837 0.4065 1.0837 1.0410
No log 2.0 42 1.0426 0.3860 1.0426 1.0211
No log 2.0952 44 0.9901 0.3907 0.9901 0.9950
No log 2.1905 46 0.9337 0.3740 0.9337 0.9663
No log 2.2857 48 0.9175 0.4415 0.9175 0.9579
No log 2.3810 50 0.9200 0.4737 0.9200 0.9592
No log 2.4762 52 0.9552 0.4534 0.9552 0.9774
No log 2.5714 54 1.0596 0.4436 1.0596 1.0294
No log 2.6667 56 1.0283 0.4746 1.0283 1.0140
No log 2.7619 58 0.9944 0.3814 0.9944 0.9972
No log 2.8571 60 1.0571 0.3461 1.0571 1.0281
No log 2.9524 62 1.1081 0.4002 1.1081 1.0527
No log 3.0476 64 0.9339 0.4152 0.9339 0.9664
No log 3.1429 66 0.8737 0.5092 0.8737 0.9347
No log 3.2381 68 0.9342 0.4820 0.9342 0.9665
No log 3.3333 70 0.9251 0.5172 0.9251 0.9618
No log 3.4286 72 1.0293 0.5004 1.0293 1.0146
No log 3.5238 74 1.0302 0.5245 1.0302 1.0150
No log 3.6190 76 1.1120 0.5223 1.1120 1.0545
No log 3.7143 78 0.8959 0.5217 0.8959 0.9465
No log 3.8095 80 0.8090 0.5679 0.8090 0.8995
No log 3.9048 82 0.9232 0.5229 0.9232 0.9608
No log 4.0 84 1.1809 0.4891 1.1809 1.0867
No log 4.0952 86 1.1026 0.5253 1.1026 1.0501
No log 4.1905 88 0.9755 0.5826 0.9755 0.9877
No log 4.2857 90 0.9530 0.5606 0.9530 0.9762
No log 4.3810 92 1.0487 0.5718 1.0487 1.0241
No log 4.4762 94 1.0336 0.5658 1.0336 1.0166
No log 4.5714 96 1.2047 0.4896 1.2047 1.0976
No log 4.6667 98 1.2718 0.4757 1.2718 1.1277
No log 4.7619 100 1.4007 0.4416 1.4007 1.1835
No log 4.8571 102 1.2785 0.4472 1.2785 1.1307
No log 4.9524 104 1.1282 0.4720 1.1282 1.0622
No log 5.0476 106 1.4023 0.4272 1.4023 1.1842
No log 5.1429 108 1.7807 0.3174 1.7807 1.3344
No log 5.2381 110 1.6270 0.3535 1.6270 1.2756
No log 5.3333 112 1.2176 0.4205 1.2176 1.1034
No log 5.4286 114 0.8741 0.4988 0.8741 0.9349
No log 5.5238 116 0.8594 0.5259 0.8594 0.9270
No log 5.6190 118 1.1196 0.4531 1.1196 1.0581
No log 5.7143 120 1.3339 0.4199 1.3339 1.1549
No log 5.8095 122 1.1914 0.4376 1.1914 1.0915
No log 5.9048 124 0.9820 0.5440 0.9820 0.9909
No log 6.0 126 1.0019 0.5553 1.0019 1.0009
No log 6.0952 128 1.0837 0.5089 1.0837 1.0410
No log 6.1905 130 0.9988 0.5566 0.9988 0.9994
No log 6.2857 132 0.9064 0.5661 0.9064 0.9521
No log 6.3810 134 0.8129 0.5246 0.8129 0.9016
No log 6.4762 136 0.8330 0.5167 0.8330 0.9127
No log 6.5714 138 0.8434 0.5302 0.8434 0.9184
No log 6.6667 140 0.9948 0.5527 0.9948 0.9974
No log 6.7619 142 0.9843 0.5704 0.9843 0.9921
No log 6.8571 144 0.8258 0.6167 0.8258 0.9087
No log 6.9524 146 0.8353 0.6180 0.8353 0.9140
No log 7.0476 148 1.0529 0.5129 1.0529 1.0261
No log 7.1429 150 1.4280 0.4307 1.4280 1.1950
No log 7.2381 152 1.4690 0.4305 1.4690 1.2120
No log 7.3333 154 1.2122 0.4912 1.2122 1.1010
No log 7.4286 156 0.8962 0.6084 0.8962 0.9467
No log 7.5238 158 0.8605 0.6189 0.8605 0.9276
No log 7.6190 160 0.8840 0.6263 0.8840 0.9402
No log 7.7143 162 1.0074 0.5707 1.0074 1.0037
No log 7.8095 164 0.9064 0.5829 0.9064 0.9520
No log 7.9048 166 0.9773 0.5706 0.9773 0.9886
No log 8.0 168 0.9506 0.6098 0.9506 0.9750
No log 8.0952 170 0.7808 0.5783 0.7808 0.8836
No log 8.1905 172 0.7524 0.6146 0.7524 0.8674
No log 8.2857 174 0.9378 0.6082 0.9378 0.9684
No log 8.3810 176 1.4494 0.4345 1.4494 1.2039
No log 8.4762 178 1.5371 0.4224 1.5371 1.2398
No log 8.5714 180 1.3369 0.4487 1.3369 1.1563
No log 8.6667 182 1.1109 0.5033 1.1109 1.0540
No log 8.7619 184 1.1345 0.4789 1.1345 1.0651
No log 8.8571 186 1.1705 0.4687 1.1705 1.0819
No log 8.9524 188 0.9735 0.5755 0.9735 0.9867
No log 9.0476 190 0.9753 0.5667 0.9753 0.9876
No log 9.1429 192 1.0174 0.5278 1.0174 1.0086
No log 9.2381 194 1.0136 0.5531 1.0136 1.0068
No log 9.3333 196 1.0599 0.5072 1.0599 1.0295
No log 9.4286 198 0.9671 0.5536 0.9671 0.9834
No log 9.5238 200 0.8217 0.5755 0.8217 0.9065
No log 9.6190 202 0.8872 0.5890 0.8872 0.9419
No log 9.7143 204 1.1174 0.5009 1.1174 1.0571
No log 9.8095 206 1.2624 0.4243 1.2624 1.1236
No log 9.9048 208 1.2109 0.4373 1.2109 1.1004
No log 10.0 210 1.0771 0.5469 1.0771 1.0378
No log 10.0952 212 0.9423 0.5806 0.9423 0.9707
No log 10.1905 214 0.9909 0.5769 0.9909 0.9955
No log 10.2857 216 1.0653 0.5434 1.0653 1.0322
No log 10.3810 218 1.0150 0.5857 1.0150 1.0075
No log 10.4762 220 0.8800 0.5992 0.8800 0.9381
No log 10.5714 222 1.0157 0.5794 1.0157 1.0078
No log 10.6667 224 1.3574 0.4276 1.3574 1.1651
No log 10.7619 226 1.3344 0.4418 1.3344 1.1551
No log 10.8571 228 0.9870 0.5795 0.9870 0.9935
No log 10.9524 230 0.8357 0.6127 0.8357 0.9142
No log 11.0476 232 0.9044 0.5990 0.9044 0.9510
No log 11.1429 234 1.1757 0.4617 1.1757 1.0843
No log 11.2381 236 1.2901 0.4605 1.2901 1.1358
No log 11.3333 238 1.0387 0.5522 1.0387 1.0192
No log 11.4286 240 0.8061 0.5762 0.8061 0.8978
No log 11.5238 242 0.7982 0.5763 0.7982 0.8934
No log 11.6190 244 0.8245 0.5318 0.8245 0.9080
No log 11.7143 246 1.1150 0.5188 1.1150 1.0559
No log 11.8095 248 1.3951 0.4704 1.3951 1.1812
No log 11.9048 250 1.2880 0.4722 1.2880 1.1349
No log 12.0 252 1.1192 0.4820 1.1192 1.0579
No log 12.0952 254 1.0060 0.5020 1.0060 1.0030
No log 12.1905 256 0.9133 0.4812 0.9133 0.9557
No log 12.2857 258 0.8731 0.4940 0.8731 0.9344
No log 12.3810 260 0.9956 0.5094 0.9956 0.9978
No log 12.4762 262 1.1087 0.4692 1.1087 1.0529
No log 12.5714 264 1.0657 0.5082 1.0657 1.0323
No log 12.6667 266 1.0332 0.5222 1.0332 1.0165
No log 12.7619 268 1.0098 0.5473 1.0098 1.0049
No log 12.8571 270 1.1047 0.5209 1.1047 1.0511
No log 12.9524 272 1.1695 0.5002 1.1695 1.0814
No log 13.0476 274 1.0695 0.5479 1.0695 1.0341
No log 13.1429 276 0.9071 0.5918 0.9071 0.9524
No log 13.2381 278 0.9670 0.6003 0.9670 0.9834
No log 13.3333 280 1.2492 0.4601 1.2492 1.1177
No log 13.4286 282 1.5628 0.3855 1.5628 1.2501
No log 13.5238 284 1.5056 0.3679 1.5056 1.2270
No log 13.6190 286 1.2243 0.4770 1.2243 1.1065
No log 13.7143 288 0.9164 0.6224 0.9164 0.9573
No log 13.8095 290 0.7593 0.6675 0.7593 0.8714
No log 13.9048 292 0.7584 0.6706 0.7584 0.8708
No log 14.0 294 0.8721 0.6551 0.8721 0.9339
No log 14.0952 296 1.0623 0.5440 1.0623 1.0307
No log 14.1905 298 1.2261 0.4641 1.2261 1.1073
No log 14.2857 300 1.0455 0.5183 1.0455 1.0225
No log 14.3810 302 0.7938 0.5783 0.7938 0.8909
No log 14.4762 304 0.7684 0.5753 0.7684 0.8766
No log 14.5714 306 0.8145 0.5668 0.8145 0.9025
No log 14.6667 308 0.9322 0.5159 0.9322 0.9655
No log 14.7619 310 0.9911 0.5269 0.9911 0.9955
No log 14.8571 312 0.9735 0.5140 0.9735 0.9867
No log 14.9524 314 0.9437 0.5333 0.9437 0.9714
No log 15.0476 316 0.9644 0.5288 0.9644 0.9820
No log 15.1429 318 1.0199 0.5457 1.0199 1.0099
No log 15.2381 320 1.1768 0.4517 1.1768 1.0848
No log 15.3333 322 1.1123 0.4807 1.1123 1.0547
No log 15.4286 324 0.9081 0.5215 0.9081 0.9529
No log 15.5238 326 0.7743 0.6075 0.7743 0.8800
No log 15.6190 328 0.7595 0.6184 0.7595 0.8715
No log 15.7143 330 0.8425 0.6113 0.8425 0.9179
No log 15.8095 332 1.0375 0.5301 1.0375 1.0186
No log 15.9048 334 1.2188 0.4692 1.2188 1.1040
No log 16.0 336 1.2259 0.4688 1.2259 1.1072
No log 16.0952 338 1.0525 0.5044 1.0525 1.0259
No log 16.1905 340 0.9310 0.5861 0.9310 0.9649
No log 16.2857 342 0.8857 0.6061 0.8857 0.9411
No log 16.3810 344 0.9439 0.5614 0.9439 0.9715
No log 16.4762 346 1.1547 0.4816 1.1547 1.0746
No log 16.5714 348 1.1388 0.4777 1.1388 1.0671
No log 16.6667 350 1.0547 0.4987 1.0547 1.0270
No log 16.7619 352 0.9281 0.5901 0.9281 0.9634
No log 16.8571 354 0.8981 0.6204 0.8981 0.9477
No log 16.9524 356 1.0693 0.5083 1.0693 1.0341
No log 17.0476 358 1.3273 0.4449 1.3273 1.1521
No log 17.1429 360 1.4287 0.4360 1.4287 1.1953
No log 17.2381 362 1.2684 0.4750 1.2684 1.1262
No log 17.3333 364 1.0100 0.5781 1.0100 1.0050
No log 17.4286 366 0.8738 0.5851 0.8738 0.9348
No log 17.5238 368 0.9437 0.5872 0.9437 0.9714
No log 17.6190 370 1.0487 0.5289 1.0487 1.0241
No log 17.7143 372 1.1026 0.5094 1.1026 1.0501
No log 17.8095 374 1.0753 0.5267 1.0753 1.0370
No log 17.9048 376 1.0048 0.5660 1.0048 1.0024
No log 18.0 378 1.0838 0.5118 1.0838 1.0411
No log 18.0952 380 1.0545 0.5297 1.0545 1.0269
No log 18.1905 382 0.9920 0.5336 0.9920 0.9960
No log 18.2857 384 0.9395 0.5679 0.9395 0.9693
No log 18.3810 386 0.8428 0.5546 0.8428 0.9180
No log 18.4762 388 0.8248 0.5514 0.8248 0.9082
No log 18.5714 390 0.8856 0.5661 0.8856 0.9411
No log 18.6667 392 0.8902 0.5623 0.8902 0.9435
No log 18.7619 394 0.8652 0.6324 0.8652 0.9301
No log 18.8571 396 0.7984 0.6447 0.7984 0.8935
No log 18.9524 398 0.9277 0.6105 0.9277 0.9632
No log 19.0476 400 1.0815 0.5299 1.0815 1.0399
No log 19.1429 402 0.9832 0.5697 0.9832 0.9916
No log 19.2381 404 0.8164 0.6233 0.8164 0.9035
No log 19.3333 406 0.7727 0.6125 0.7727 0.8790
No log 19.4286 408 0.8854 0.5977 0.8854 0.9410
No log 19.5238 410 1.2728 0.4531 1.2728 1.1282
No log 19.6190 412 1.5282 0.4234 1.5282 1.2362
No log 19.7143 414 1.4828 0.4315 1.4828 1.2177
No log 19.8095 416 1.2255 0.4641 1.2255 1.1070
No log 19.9048 418 0.9492 0.5655 0.9492 0.9743
No log 20.0 420 0.8821 0.5688 0.8821 0.9392
No log 20.0952 422 0.9913 0.5631 0.9913 0.9956
No log 20.1905 424 1.2537 0.5064 1.2537 1.1197
No log 20.2857 426 1.4504 0.3922 1.4504 1.2043
No log 20.3810 428 1.4436 0.4111 1.4436 1.2015
No log 20.4762 430 1.1524 0.4861 1.1524 1.0735
No log 20.5714 432 0.9780 0.5369 0.9780 0.9889
No log 20.6667 434 0.9281 0.5538 0.9281 0.9634
No log 20.7619 436 1.0594 0.5637 1.0594 1.0293
No log 20.8571 438 1.2666 0.5261 1.2666 1.1254
No log 20.9524 440 1.3375 0.5167 1.3375 1.1565
No log 21.0476 442 1.2222 0.5176 1.2222 1.1055
No log 21.1429 444 0.9327 0.5812 0.9327 0.9658
No log 21.2381 446 0.8563 0.5801 0.8563 0.9254
No log 21.3333 448 0.8793 0.5702 0.8793 0.9377
No log 21.4286 450 0.8965 0.5476 0.8965 0.9468
No log 21.5238 452 0.9528 0.5573 0.9528 0.9761
No log 21.6190 454 1.0517 0.5368 1.0517 1.0255
No log 21.7143 456 1.0242 0.5251 1.0242 1.0120
No log 21.8095 458 1.0598 0.5484 1.0598 1.0294
No log 21.9048 460 1.1089 0.5314 1.1089 1.0530
No log 22.0 462 1.2503 0.5240 1.2503 1.1182
No log 22.0952 464 1.1438 0.5353 1.1438 1.0695
No log 22.1905 466 1.0141 0.5536 1.0141 1.0070
No log 22.2857 468 1.1379 0.5368 1.1379 1.0667
No log 22.3810 470 1.3319 0.4639 1.3319 1.1541
No log 22.4762 472 1.3788 0.4264 1.3788 1.1742
No log 22.5714 474 1.4397 0.3999 1.4397 1.1999
No log 22.6667 476 1.2411 0.4601 1.2411 1.1141
No log 22.7619 478 0.9694 0.5732 0.9694 0.9846
No log 22.8571 480 0.8941 0.5987 0.8941 0.9456
No log 22.9524 482 0.8672 0.5861 0.8672 0.9313
No log 23.0476 484 1.0105 0.5585 1.0105 1.0052
No log 23.1429 486 1.0999 0.5267 1.0999 1.0488
No log 23.2381 488 1.1288 0.5196 1.1288 1.0625
No log 23.3333 490 1.0737 0.5375 1.0737 1.0362
No log 23.4286 492 0.9394 0.6056 0.9394 0.9692
No log 23.5238 494 0.8352 0.6054 0.8352 0.9139
No log 23.6190 496 0.8737 0.6192 0.8737 0.9347
No log 23.7143 498 0.9787 0.5657 0.9787 0.9893
0.3768 23.8095 500 1.0500 0.5456 1.0500 1.0247
0.3768 23.9048 502 1.1351 0.5342 1.1351 1.0654
0.3768 24.0 504 1.0314 0.5270 1.0314 1.0156
0.3768 24.0952 506 1.0225 0.5371 1.0225 1.0112
0.3768 24.1905 508 0.9611 0.5667 0.9611 0.9804
0.3768 24.2857 510 1.0315 0.5545 1.0315 1.0156
0.3768 24.3810 512 1.1006 0.5150 1.1006 1.0491
0.3768 24.4762 514 1.3494 0.4583 1.3494 1.1616
0.3768 24.5714 516 1.4896 0.4314 1.4896 1.2205
0.3768 24.6667 518 1.4054 0.4487 1.4054 1.1855
0.3768 24.7619 520 1.1823 0.4656 1.1823 1.0873

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run2_AugV5_k4_task1_organization

Finetuned
(4023)
this model