ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k2_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0157
  • Qwk: 0.5354
  • Mse: 1.0157
  • Rmse: 1.0078

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.2 2 7.0457 0.0167 7.0457 2.6544
No log 0.4 4 4.1515 0.0769 4.1515 2.0375
No log 0.6 6 2.6392 0.1410 2.6392 1.6246
No log 0.8 8 2.1222 0.1926 2.1222 1.4568
No log 1.0 10 2.0386 0.1955 2.0386 1.4278
No log 1.2 12 1.8759 0.2439 1.8759 1.3697
No log 1.4 14 1.7688 0.1404 1.7688 1.3300
No log 1.6 16 2.0393 0.1148 2.0393 1.4281
No log 1.8 18 2.3592 0.1240 2.3592 1.5360
No log 2.0 20 2.2286 0.1550 2.2286 1.4928
No log 2.2 22 1.9298 0.2764 1.9298 1.3892
No log 2.4 24 2.1090 0.1527 2.1090 1.4522
No log 2.6 26 2.3665 0.0571 2.3665 1.5384
No log 2.8 28 2.2771 0.1471 2.2771 1.5090
No log 3.0 30 1.8746 0.2764 1.8746 1.3692
No log 3.2 32 1.6620 0.2881 1.6620 1.2892
No log 3.4 34 1.6728 0.2689 1.6728 1.2934
No log 3.6 36 1.6458 0.2833 1.6458 1.2829
No log 3.8 38 1.4108 0.3009 1.4108 1.1878
No log 4.0 40 1.4592 0.2087 1.4592 1.2080
No log 4.2 42 1.6107 0.3607 1.6107 1.2691
No log 4.4 44 1.6353 0.3968 1.6353 1.2788
No log 4.6 46 1.4670 0.4308 1.4670 1.2112
No log 4.8 48 1.5685 0.3810 1.5685 1.2524
No log 5.0 50 2.4311 0.1316 2.4311 1.5592
No log 5.2 52 3.1374 0.0814 3.1374 1.7713
No log 5.4 54 2.5362 0.0886 2.5362 1.5926
No log 5.6 56 1.4245 0.4228 1.4245 1.1935
No log 5.8 58 1.2025 0.3866 1.2025 1.0966
No log 6.0 60 1.1964 0.4370 1.1964 1.0938
No log 6.2 62 1.1123 0.4516 1.1123 1.0547
No log 6.4 64 1.1213 0.4882 1.1213 1.0589
No log 6.6 66 1.2231 0.5180 1.2231 1.1059
No log 6.8 68 1.0333 0.5758 1.0333 1.0165
No log 7.0 70 0.9350 0.6565 0.9350 0.9670
No log 7.2 72 0.9413 0.6565 0.9413 0.9702
No log 7.4 74 0.9215 0.6260 0.9215 0.9600
No log 7.6 76 0.9285 0.6364 0.9285 0.9636
No log 7.8 78 1.0094 0.6074 1.0094 1.0047
No log 8.0 80 0.9788 0.5821 0.9788 0.9893
No log 8.2 82 0.9829 0.6222 0.9829 0.9914
No log 8.4 84 1.1230 0.5373 1.1230 1.0597
No log 8.6 86 1.2178 0.4853 1.2178 1.1035
No log 8.8 88 1.1869 0.5 1.1869 1.0894
No log 9.0 90 1.1533 0.5333 1.1533 1.0739
No log 9.2 92 1.1795 0.5692 1.1795 1.0860
No log 9.4 94 1.1886 0.5692 1.1886 1.0902
No log 9.6 96 1.1647 0.5312 1.1647 1.0792
No log 9.8 98 1.1607 0.5185 1.1607 1.0774
No log 10.0 100 1.3254 0.4328 1.3254 1.1513
No log 10.2 102 1.3140 0.4478 1.3140 1.1463
No log 10.4 104 1.1460 0.6 1.1460 1.0705
No log 10.6 106 1.1320 0.6131 1.1320 1.0640
No log 10.8 108 1.1243 0.5970 1.1243 1.0603
No log 11.0 110 1.1268 0.5846 1.1268 1.0615
No log 11.2 112 1.0421 0.6047 1.0421 1.0208
No log 11.4 114 0.9420 0.6154 0.9420 0.9706
No log 11.6 116 0.9271 0.6667 0.9271 0.9629
No log 11.8 118 0.8972 0.6667 0.8972 0.9472
No log 12.0 120 0.8381 0.6714 0.8381 0.9155
No log 12.2 122 0.8282 0.6713 0.8282 0.9101
No log 12.4 124 0.8549 0.6618 0.8549 0.9246
No log 12.6 126 0.9325 0.6061 0.9325 0.9657
No log 12.8 128 0.9569 0.6308 0.9569 0.9782
No log 13.0 130 0.9749 0.6515 0.9749 0.9874
No log 13.2 132 1.0065 0.6519 1.0065 1.0033
No log 13.4 134 0.9869 0.6154 0.9869 0.9934
No log 13.6 136 0.9545 0.6389 0.9545 0.9770
No log 13.8 138 0.9443 0.6197 0.9443 0.9718
No log 14.0 140 0.9334 0.6619 0.9334 0.9661
No log 14.2 142 0.9360 0.6714 0.9360 0.9675
No log 14.4 144 0.9386 0.6713 0.9386 0.9688
No log 14.6 146 0.9044 0.6809 0.9044 0.9510
No log 14.8 148 0.8767 0.7083 0.8767 0.9363
No log 15.0 150 0.8936 0.6533 0.8936 0.9453
No log 15.2 152 0.9334 0.6225 0.9334 0.9661
No log 15.4 154 0.9514 0.6187 0.9514 0.9754
No log 15.6 156 1.0114 0.6565 1.0114 1.0057
No log 15.8 158 1.0405 0.6212 1.0405 1.0201
No log 16.0 160 0.9861 0.6165 0.9861 0.9930
No log 16.2 162 0.9470 0.6370 0.9470 0.9732
No log 16.4 164 0.9821 0.6418 0.9821 0.9910
No log 16.6 166 1.0323 0.6212 1.0323 1.0160
No log 16.8 168 1.0308 0.6165 1.0308 1.0153
No log 17.0 170 1.0358 0.5191 1.0358 1.0178
No log 17.2 172 1.0207 0.5874 1.0207 1.0103
No log 17.4 174 1.0110 0.5986 1.0110 1.0055
No log 17.6 176 1.0253 0.6187 1.0253 1.0126
No log 17.8 178 1.0156 0.5821 1.0156 1.0078
No log 18.0 180 1.0101 0.6061 1.0101 1.0051
No log 18.2 182 0.9989 0.6418 0.9989 0.9995
No log 18.4 184 0.9606 0.6370 0.9606 0.9801
No log 18.6 186 0.9435 0.6119 0.9435 0.9713
No log 18.8 188 0.9697 0.6515 0.9697 0.9847
No log 19.0 190 1.0184 0.6 1.0184 1.0092
No log 19.2 192 1.0287 0.5625 1.0287 1.0142
No log 19.4 194 1.0163 0.5469 1.0163 1.0081
No log 19.6 196 1.0085 0.6047 1.0085 1.0043
No log 19.8 198 1.0484 0.5781 1.0484 1.0239
No log 20.0 200 1.0467 0.6154 1.0467 1.0231
No log 20.2 202 1.0017 0.5692 1.0017 1.0009
No log 20.4 204 0.9706 0.5714 0.9706 0.9852
No log 20.6 206 0.9721 0.5758 0.9721 0.9859
No log 20.8 208 0.9688 0.5672 0.9688 0.9843
No log 21.0 210 0.9475 0.6222 0.9475 0.9734
No log 21.2 212 0.9324 0.6202 0.9324 0.9656
No log 21.4 214 0.9717 0.5649 0.9717 0.9857
No log 21.6 216 1.0569 0.5231 1.0569 1.0281
No log 21.8 218 1.0613 0.5564 1.0613 1.0302
No log 22.0 220 1.0556 0.6047 1.0556 1.0274
No log 22.2 222 1.0833 0.5649 1.0833 1.0408
No log 22.4 224 1.0822 0.5344 1.0822 1.0403
No log 22.6 226 1.0400 0.5191 1.0400 1.0198
No log 22.8 228 1.0008 0.5802 1.0008 1.0004
No log 23.0 230 1.0044 0.5846 1.0044 1.0022
No log 23.2 232 0.9924 0.6107 0.9924 0.9962
No log 23.4 234 0.9693 0.6107 0.9693 0.9845
No log 23.6 236 0.9314 0.6418 0.9314 0.9651
No log 23.8 238 0.9203 0.6165 0.9203 0.9593
No log 24.0 240 0.9237 0.6418 0.9237 0.9611
No log 24.2 242 0.9387 0.6364 0.9387 0.9689
No log 24.4 244 0.9881 0.6308 0.9881 0.9940
No log 24.6 246 0.9874 0.6260 0.9874 0.9937
No log 24.8 248 0.9666 0.6260 0.9666 0.9831
No log 25.0 250 0.9590 0.5802 0.9590 0.9793
No log 25.2 252 0.9628 0.6212 0.9628 0.9812
No log 25.4 254 1.0068 0.5538 1.0068 1.0034
No log 25.6 256 1.0028 0.5354 1.0028 1.0014
No log 25.8 258 0.9513 0.5736 0.9513 0.9754
No log 26.0 260 0.9146 0.6519 0.9146 0.9564
No log 26.2 262 0.9184 0.6074 0.9184 0.9583
No log 26.4 264 0.9184 0.6074 0.9184 0.9583
No log 26.6 266 0.9123 0.6074 0.9123 0.9552
No log 26.8 268 0.9077 0.6466 0.9077 0.9527
No log 27.0 270 0.9321 0.6364 0.9321 0.9655
No log 27.2 272 0.9606 0.6515 0.9606 0.9801
No log 27.4 274 0.9913 0.6119 0.9913 0.9956
No log 27.6 276 1.0058 0.6061 1.0058 1.0029
No log 27.8 278 0.9832 0.6370 0.9832 0.9916
No log 28.0 280 0.9349 0.6423 0.9349 0.9669
No log 28.2 282 0.9270 0.6809 0.9270 0.9628
No log 28.4 284 0.9335 0.6716 0.9335 0.9662
No log 28.6 286 0.9484 0.6667 0.9484 0.9739
No log 28.8 288 0.9552 0.6412 0.9552 0.9774
No log 29.0 290 0.9737 0.5672 0.9737 0.9868
No log 29.2 292 1.0289 0.5588 1.0289 1.0143
No log 29.4 294 0.9953 0.5481 0.9953 0.9976
No log 29.6 296 0.9067 0.6519 0.9067 0.9522
No log 29.8 298 0.8799 0.6466 0.8799 0.9380
No log 30.0 300 0.9090 0.6466 0.9090 0.9534
No log 30.2 302 0.9096 0.6522 0.9096 0.9537
No log 30.4 304 0.9067 0.6471 0.9067 0.9522
No log 30.6 306 0.9063 0.6056 0.9063 0.9520
No log 30.8 308 0.8398 0.6759 0.8398 0.9164
No log 31.0 310 0.8049 0.7133 0.8049 0.8972
No log 31.2 312 0.8101 0.7123 0.8101 0.9001
No log 31.4 314 0.8337 0.7 0.8337 0.9131
No log 31.6 316 0.8579 0.6866 0.8579 0.9262
No log 31.8 318 0.8772 0.6567 0.8772 0.9366
No log 32.0 320 0.9042 0.6364 0.9042 0.9509
No log 32.2 322 0.9303 0.6515 0.9303 0.9645
No log 32.4 324 0.9809 0.6015 0.9809 0.9904
No log 32.6 326 1.0160 0.6015 1.0160 1.0080
No log 32.8 328 1.0438 0.6061 1.0438 1.0216
No log 33.0 330 1.0340 0.6212 1.0340 1.0169
No log 33.2 332 1.0109 0.6119 1.0109 1.0054
No log 33.4 334 1.0154 0.5481 1.0154 1.0077
No log 33.6 336 1.0146 0.5441 1.0146 1.0073
No log 33.8 338 0.9963 0.5674 0.9963 0.9981
No log 34.0 340 0.9566 0.6207 0.9566 0.9781
No log 34.2 342 0.9156 0.6301 0.9156 0.9569
No log 34.4 344 0.8940 0.6122 0.8940 0.9455
No log 34.6 346 0.8946 0.6122 0.8946 0.9458
No log 34.8 348 0.9268 0.6525 0.9268 0.9627
No log 35.0 350 0.9610 0.6423 0.9610 0.9803
No log 35.2 352 0.9984 0.6471 0.9984 0.9992
No log 35.4 354 1.0336 0.5802 1.0336 1.0166
No log 35.6 356 1.0575 0.5909 1.0575 1.0283
No log 35.8 358 1.0587 0.6165 1.0587 1.0289
No log 36.0 360 1.0490 0.6165 1.0490 1.0242
No log 36.2 362 1.0207 0.6165 1.0207 1.0103
No log 36.4 364 1.0059 0.6165 1.0059 1.0030
No log 36.6 366 0.9756 0.6119 0.9756 0.9877
No log 36.8 368 0.9397 0.6316 0.9397 0.9694
No log 37.0 370 0.9229 0.5821 0.9229 0.9607
No log 37.2 372 0.9281 0.5985 0.9281 0.9634
No log 37.4 374 0.9487 0.5882 0.9487 0.9740
No log 37.6 376 0.9737 0.5909 0.9737 0.9868
No log 37.8 378 0.9665 0.5781 0.9665 0.9831
No log 38.0 380 0.9618 0.6015 0.9618 0.9807
No log 38.2 382 0.9713 0.6165 0.9713 0.9855
No log 38.4 384 0.9601 0.6165 0.9601 0.9799
No log 38.6 386 0.9297 0.6222 0.9297 0.9642
No log 38.8 388 0.9142 0.6222 0.9142 0.9562
No log 39.0 390 0.8985 0.6377 0.8985 0.9479
No log 39.2 392 0.9009 0.6277 0.9009 0.9492
No log 39.4 394 0.9257 0.5926 0.9257 0.9621
No log 39.6 396 0.9300 0.5926 0.9300 0.9643
No log 39.8 398 0.9329 0.5926 0.9329 0.9659
No log 40.0 400 0.9527 0.6269 0.9527 0.9761
No log 40.2 402 0.9658 0.6316 0.9658 0.9827
No log 40.4 404 0.9654 0.6316 0.9654 0.9825
No log 40.6 406 0.9606 0.6316 0.9606 0.9801
No log 40.8 408 0.9586 0.6316 0.9586 0.9791
No log 41.0 410 0.9578 0.6316 0.9578 0.9787
No log 41.2 412 0.9524 0.6316 0.9524 0.9759
No log 41.4 414 0.9578 0.6061 0.9578 0.9787
No log 41.6 416 0.9592 0.5970 0.9592 0.9794
No log 41.8 418 0.9599 0.5970 0.9599 0.9797
No log 42.0 420 0.9602 0.6269 0.9602 0.9799
No log 42.2 422 0.9793 0.6316 0.9793 0.9896
No log 42.4 424 0.9830 0.6316 0.9830 0.9914
No log 42.6 426 0.9647 0.6269 0.9647 0.9822
No log 42.8 428 0.9446 0.6324 0.9446 0.9719
No log 43.0 430 0.9404 0.5778 0.9404 0.9698
No log 43.2 432 0.9581 0.6074 0.9581 0.9788
No log 43.4 434 0.9617 0.5970 0.9617 0.9807
No log 43.6 436 0.9545 0.6176 0.9545 0.9770
No log 43.8 438 0.9375 0.6119 0.9375 0.9682
No log 44.0 440 0.9293 0.5821 0.9293 0.9640
No log 44.2 442 0.9211 0.5926 0.9211 0.9597
No log 44.4 444 0.9068 0.5970 0.9068 0.9523
No log 44.6 446 0.8987 0.6074 0.8987 0.9480
No log 44.8 448 0.8929 0.6277 0.8929 0.9450
No log 45.0 450 0.8978 0.6119 0.8978 0.9475
No log 45.2 452 0.9210 0.6107 0.9210 0.9597
No log 45.4 454 0.9689 0.6 0.9689 0.9843
No log 45.6 456 1.0147 0.6047 1.0147 1.0073
No log 45.8 458 1.0298 0.6308 1.0298 1.0148
No log 46.0 460 1.0312 0.6512 1.0312 1.0155
No log 46.2 462 1.0235 0.6462 1.0235 1.0117
No log 46.4 464 1.0097 0.6202 1.0097 1.0049
No log 46.6 466 0.9994 0.5938 0.9994 0.9997
No log 46.8 468 0.9955 0.5669 0.9955 0.9977
No log 47.0 470 0.9913 0.5581 0.9913 0.9957
No log 47.2 472 1.0064 0.6107 1.0064 1.0032
No log 47.4 474 1.0405 0.6107 1.0405 1.0200
No log 47.6 476 1.0807 0.5736 1.0807 1.0396
No log 47.8 478 1.0912 0.5736 1.0912 1.0446
No log 48.0 480 1.1092 0.5736 1.1092 1.0532
No log 48.2 482 1.0990 0.5625 1.0990 1.0483
No log 48.4 484 1.0541 0.5736 1.0541 1.0267
No log 48.6 486 1.0088 0.5669 1.0088 1.0044
No log 48.8 488 0.9744 0.5649 0.9744 0.9871
No log 49.0 490 0.9562 0.5344 0.9562 0.9779
No log 49.2 492 0.9480 0.5455 0.9480 0.9737
No log 49.4 494 0.9529 0.5758 0.9529 0.9761
No log 49.6 496 0.9627 0.6061 0.9627 0.9812
No log 49.8 498 0.9762 0.6061 0.9762 0.9880
0.2819 50.0 500 0.9938 0.6364 0.9938 0.9969
0.2819 50.2 502 1.0184 0.6364 1.0184 1.0091
0.2819 50.4 504 1.0226 0.6 1.0226 1.0113
0.2819 50.6 506 1.0179 0.6107 1.0179 1.0089
0.2819 50.8 508 1.0092 0.5581 1.0092 1.0046
0.2819 51.0 510 1.0157 0.5354 1.0157 1.0078

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k2_task1_organization

Finetuned
(4023)
this model