ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k12_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8004
  • Qwk: 0.3740
  • Mse: 0.8004
  • Rmse: 0.8947

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.05 2 4.3996 -0.0387 4.3996 2.0975
No log 0.1 4 2.5039 0.1069 2.5039 1.5824
No log 0.15 6 1.9026 -0.0082 1.9026 1.3794
No log 0.2 8 2.2262 0.0468 2.2262 1.4921
No log 0.25 10 1.4947 0.0294 1.4947 1.2226
No log 0.3 12 1.1570 0.1261 1.1570 1.0757
No log 0.35 14 1.1811 0.1525 1.1811 1.0868
No log 0.4 16 1.2951 0.0 1.2951 1.1380
No log 0.45 18 1.4303 0.0279 1.4303 1.1959
No log 0.5 20 1.4203 0.0767 1.4203 1.1918
No log 0.55 22 1.1506 0.1790 1.1506 1.0727
No log 0.6 24 1.0489 0.1779 1.0489 1.0242
No log 0.65 26 1.2126 -0.0344 1.2126 1.1012
No log 0.7 28 1.3209 -0.0824 1.3209 1.1493
No log 0.75 30 1.3966 -0.0411 1.3966 1.1818
No log 0.8 32 1.5516 -0.0411 1.5516 1.2456
No log 0.85 34 1.5064 0.0 1.5064 1.2274
No log 0.9 36 1.3683 0.0 1.3683 1.1698
No log 0.95 38 1.1934 0.1790 1.1934 1.0924
No log 1.0 40 1.0998 0.1821 1.0998 1.0487
No log 1.05 42 1.0880 0.1493 1.0880 1.0431
No log 1.1 44 1.0884 0.1313 1.0884 1.0433
No log 1.15 46 1.1484 0.2035 1.1484 1.0716
No log 1.2 48 1.2845 0.3052 1.2845 1.1334
No log 1.25 50 1.3421 0.3243 1.3421 1.1585
No log 1.3 52 1.3056 0.3734 1.3056 1.1426
No log 1.35 54 1.2724 0.3486 1.2724 1.1280
No log 1.4 56 1.1410 0.2665 1.1410 1.0682
No log 1.45 58 1.1118 0.2455 1.1118 1.0544
No log 1.5 60 1.0709 0.2268 1.0709 1.0349
No log 1.55 62 1.0471 0.2812 1.0471 1.0233
No log 1.6 64 1.0542 0.2615 1.0542 1.0268
No log 1.65 66 1.0995 0.1304 1.0995 1.0486
No log 1.7 68 1.1592 0.1825 1.1592 1.0767
No log 1.75 70 1.0988 0.2217 1.0988 1.0482
No log 1.8 72 1.0861 0.2366 1.0861 1.0421
No log 1.85 74 1.1094 0.2735 1.1094 1.0533
No log 1.9 76 1.0473 0.2716 1.0473 1.0234
No log 1.95 78 1.1016 0.2207 1.1016 1.0496
No log 2.0 80 1.0817 0.3649 1.0817 1.0400
No log 2.05 82 0.9870 0.3948 0.9870 0.9935
No log 2.1 84 0.9535 0.3522 0.9535 0.9765
No log 2.15 86 0.9354 0.4613 0.9354 0.9672
No log 2.2 88 0.9196 0.5040 0.9196 0.9589
No log 2.25 90 0.9365 0.5534 0.9365 0.9677
No log 2.3 92 0.9700 0.4625 0.9700 0.9849
No log 2.35 94 0.8919 0.4783 0.8919 0.9444
No log 2.4 96 0.8617 0.5401 0.8617 0.9283
No log 2.45 98 0.8580 0.5291 0.8580 0.9263
No log 2.5 100 1.1009 0.4287 1.1009 1.0492
No log 2.55 102 1.1637 0.3629 1.1637 1.0788
No log 2.6 104 0.9717 0.4161 0.9717 0.9857
No log 2.65 106 0.9248 0.4843 0.9248 0.9617
No log 2.7 108 0.8966 0.5094 0.8966 0.9469
No log 2.75 110 0.8894 0.5015 0.8894 0.9431
No log 2.8 112 0.9255 0.4949 0.9255 0.9620
No log 2.85 114 1.0043 0.4728 1.0043 1.0021
No log 2.9 116 1.1315 0.3862 1.1315 1.0637
No log 2.95 118 1.1955 0.3253 1.1955 1.0934
No log 3.0 120 1.1425 0.4369 1.1425 1.0689
No log 3.05 122 0.9830 0.5826 0.9830 0.9915
No log 3.1 124 0.9449 0.4936 0.9449 0.9721
No log 3.15 126 0.9766 0.4862 0.9766 0.9882
No log 3.2 128 1.1082 0.4217 1.1082 1.0527
No log 3.25 130 1.0347 0.4517 1.0347 1.0172
No log 3.3 132 0.9277 0.4835 0.9277 0.9632
No log 3.35 134 0.9052 0.4354 0.9052 0.9514
No log 3.4 136 0.8927 0.4772 0.8927 0.9448
No log 3.45 138 0.9405 0.3351 0.9405 0.9698
No log 3.5 140 0.9059 0.3992 0.9059 0.9518
No log 3.55 142 0.8576 0.5077 0.8576 0.9261
No log 3.6 144 0.8070 0.6083 0.8070 0.8983
No log 3.65 146 0.8063 0.6225 0.8063 0.8979
No log 3.7 148 0.7939 0.6343 0.7939 0.8910
No log 3.75 150 0.8683 0.5953 0.8683 0.9318
No log 3.8 152 1.1542 0.3696 1.1542 1.0743
No log 3.85 154 1.2629 0.3248 1.2629 1.1238
No log 3.9 156 1.0738 0.3094 1.0738 1.0362
No log 3.95 158 0.8322 0.5819 0.8322 0.9123
No log 4.0 160 0.9500 0.4163 0.9500 0.9747
No log 4.05 162 1.1048 0.4436 1.1048 1.0511
No log 4.1 164 0.9506 0.3942 0.9506 0.9750
No log 4.15 166 0.8129 0.5349 0.8129 0.9016
No log 4.2 168 0.9573 0.4278 0.9573 0.9784
No log 4.25 170 1.0369 0.375 1.0369 1.0183
No log 4.3 172 0.9337 0.3103 0.9337 0.9663
No log 4.35 174 0.8926 0.3547 0.8926 0.9448
No log 4.4 176 0.8903 0.4106 0.8903 0.9435
No log 4.45 178 0.8715 0.4954 0.8715 0.9335
No log 4.5 180 0.9006 0.4841 0.9006 0.9490
No log 4.55 182 0.8084 0.5905 0.8084 0.8991
No log 4.6 184 0.7970 0.6066 0.7970 0.8928
No log 4.65 186 0.8136 0.5763 0.8136 0.9020
No log 4.7 188 0.7880 0.5425 0.7880 0.8877
No log 4.75 190 0.7994 0.5425 0.7994 0.8941
No log 4.8 192 0.8260 0.5208 0.8260 0.9089
No log 4.85 194 1.0105 0.4162 1.0105 1.0052
No log 4.9 196 0.9988 0.4510 0.9988 0.9994
No log 4.95 198 0.8457 0.5559 0.8457 0.9196
No log 5.0 200 0.8121 0.6168 0.8121 0.9012
No log 5.05 202 0.8152 0.6068 0.8152 0.9029
No log 5.1 204 0.9088 0.5023 0.9088 0.9533
No log 5.15 206 1.0782 0.3893 1.0782 1.0384
No log 5.2 208 1.1182 0.3590 1.1182 1.0575
No log 5.25 210 1.0044 0.3161 1.0044 1.0022
No log 5.3 212 0.8810 0.3802 0.8810 0.9386
No log 5.35 214 0.8332 0.5054 0.8332 0.9128
No log 5.4 216 0.8439 0.4984 0.8439 0.9187
No log 5.45 218 0.8516 0.5300 0.8516 0.9228
No log 5.5 220 0.8908 0.4990 0.8908 0.9438
No log 5.55 222 1.0469 0.3922 1.0469 1.0232
No log 5.6 224 1.0718 0.3935 1.0718 1.0353
No log 5.65 226 0.9632 0.4548 0.9632 0.9814
No log 5.7 228 0.8963 0.5291 0.8963 0.9467
No log 5.75 230 0.8738 0.5302 0.8738 0.9348
No log 5.8 232 0.8830 0.5182 0.8830 0.9397
No log 5.85 234 0.9354 0.4722 0.9354 0.9671
No log 5.9 236 0.9944 0.4459 0.9944 0.9972
No log 5.95 238 0.9374 0.4695 0.9374 0.9682
No log 6.0 240 0.9306 0.4334 0.9306 0.9647
No log 6.05 242 0.9389 0.4341 0.9389 0.9690
No log 6.1 244 0.9725 0.4565 0.9725 0.9862
No log 6.15 246 0.9747 0.4565 0.9747 0.9873
No log 6.2 248 0.9055 0.4810 0.9055 0.9516
No log 6.25 250 0.9095 0.4810 0.9095 0.9537
No log 6.3 252 0.9101 0.4924 0.9101 0.9540
No log 6.35 254 0.8756 0.4710 0.8756 0.9358
No log 6.4 256 0.9137 0.5144 0.9137 0.9559
No log 6.45 258 1.0273 0.4570 1.0273 1.0136
No log 6.5 260 1.0436 0.4467 1.0436 1.0216
No log 6.55 262 0.8602 0.5159 0.8602 0.9275
No log 6.6 264 0.7974 0.5950 0.7974 0.8930
No log 6.65 266 0.7846 0.5773 0.7846 0.8858
No log 6.7 268 0.7935 0.5495 0.7935 0.8908
No log 6.75 270 0.9303 0.4790 0.9303 0.9645
No log 6.8 272 0.9293 0.4790 0.9293 0.9640
No log 6.85 274 0.7923 0.5374 0.7923 0.8901
No log 6.9 276 0.7704 0.5501 0.7704 0.8778
No log 6.95 278 0.7664 0.5726 0.7664 0.8754
No log 7.0 280 0.7518 0.5861 0.7518 0.8671
No log 7.05 282 0.7620 0.5315 0.7620 0.8729
No log 7.1 284 0.8208 0.4833 0.8208 0.9060
No log 7.15 286 0.8296 0.4940 0.8296 0.9108
No log 7.2 288 0.8140 0.5279 0.8140 0.9022
No log 7.25 290 0.7781 0.5523 0.7781 0.8821
No log 7.3 292 0.8000 0.5291 0.8000 0.8944
No log 7.35 294 0.8401 0.5607 0.8401 0.9166
No log 7.4 296 0.8444 0.5291 0.8444 0.9189
No log 7.45 298 0.8281 0.5291 0.8281 0.9100
No log 7.5 300 0.8568 0.5668 0.8568 0.9256
No log 7.55 302 0.9738 0.4705 0.9738 0.9868
No log 7.6 304 1.0736 0.4407 1.0736 1.0361
No log 7.65 306 0.9279 0.5315 0.9279 0.9633
No log 7.7 308 0.7833 0.6129 0.7833 0.8850
No log 7.75 310 0.8189 0.4876 0.8189 0.9049
No log 7.8 312 0.8002 0.4761 0.8002 0.8946
No log 7.85 314 0.7902 0.4279 0.7902 0.8889
No log 7.9 316 0.8549 0.4570 0.8549 0.9246
No log 7.95 318 0.8472 0.4455 0.8472 0.9204
No log 8.0 320 0.8016 0.4395 0.8016 0.8953
No log 8.05 322 0.7749 0.5169 0.7749 0.8803
No log 8.1 324 0.7915 0.5073 0.7915 0.8897
No log 8.15 326 0.7937 0.5073 0.7937 0.8909
No log 8.2 328 0.7668 0.6167 0.7668 0.8757
No log 8.25 330 0.8215 0.4840 0.8215 0.9064
No log 8.3 332 0.8557 0.5033 0.8557 0.9251
No log 8.35 334 0.8447 0.5033 0.8447 0.9191
No log 8.4 336 0.7766 0.5763 0.7766 0.8812
No log 8.45 338 0.7628 0.6167 0.7628 0.8734
No log 8.5 340 0.7641 0.5884 0.7641 0.8741
No log 8.55 342 0.7523 0.6138 0.7523 0.8673
No log 8.6 344 0.8025 0.5527 0.8025 0.8958
No log 8.65 346 0.8564 0.5406 0.8564 0.9254
No log 8.7 348 0.8000 0.5161 0.8000 0.8944
No log 8.75 350 0.7716 0.5681 0.7716 0.8784
No log 8.8 352 0.7438 0.6226 0.7438 0.8624
No log 8.85 354 0.7506 0.5681 0.7506 0.8664
No log 8.9 356 0.7691 0.5268 0.7691 0.8770
No log 8.95 358 0.8227 0.5808 0.8227 0.9070
No log 9.0 360 0.8440 0.5617 0.8440 0.9187
No log 9.05 362 0.9250 0.5680 0.9250 0.9618
No log 9.1 364 0.8754 0.5779 0.8754 0.9357
No log 9.15 366 0.7807 0.5569 0.7807 0.8836
No log 9.2 368 0.7745 0.5610 0.7745 0.8801
No log 9.25 370 0.7750 0.5500 0.7750 0.8804
No log 9.3 372 0.7828 0.4136 0.7828 0.8847
No log 9.35 374 0.8203 0.3990 0.8203 0.9057
No log 9.4 376 0.8733 0.4708 0.8733 0.9345
No log 9.45 378 0.8522 0.4386 0.8522 0.9232
No log 9.5 380 0.8288 0.4853 0.8288 0.9104
No log 9.55 382 0.8344 0.5458 0.8344 0.9135
No log 9.6 384 0.8794 0.5036 0.8794 0.9378
No log 9.65 386 0.8358 0.5437 0.8358 0.9142
No log 9.7 388 0.8069 0.5637 0.8069 0.8983
No log 9.75 390 0.7741 0.5991 0.7741 0.8798
No log 9.8 392 0.7598 0.5913 0.7598 0.8717
No log 9.85 394 0.7804 0.5540 0.7804 0.8834
No log 9.9 396 0.7984 0.5072 0.7984 0.8935
No log 9.95 398 0.8192 0.5131 0.8192 0.9051
No log 10.0 400 0.7696 0.5378 0.7696 0.8773
No log 10.05 402 0.7403 0.5534 0.7403 0.8604
No log 10.1 404 0.7255 0.5808 0.7255 0.8517
No log 10.15 406 0.7305 0.4932 0.7305 0.8547
No log 10.2 408 0.7402 0.5455 0.7402 0.8604
No log 10.25 410 0.7890 0.5059 0.7890 0.8883
No log 10.3 412 0.7828 0.5067 0.7828 0.8848
No log 10.35 414 0.7753 0.5528 0.7753 0.8805
No log 10.4 416 0.7332 0.5139 0.7332 0.8563
No log 10.45 418 0.7238 0.5152 0.7238 0.8508
No log 10.5 420 0.7286 0.4893 0.7286 0.8536
No log 10.55 422 0.7235 0.5681 0.7235 0.8506
No log 10.6 424 0.7145 0.6179 0.7145 0.8453
No log 10.65 426 0.7236 0.6133 0.7236 0.8506
No log 10.7 428 0.7686 0.5736 0.7686 0.8767
No log 10.75 430 0.7778 0.5436 0.7778 0.8819
No log 10.8 432 0.7165 0.5422 0.7165 0.8464
No log 10.85 434 0.7150 0.5786 0.7150 0.8456
No log 10.9 436 0.7207 0.5577 0.7207 0.8490
No log 10.95 438 0.7253 0.5121 0.7253 0.8516
No log 11.0 440 0.7431 0.5654 0.7431 0.8620
No log 11.05 442 0.7325 0.5786 0.7325 0.8558
No log 11.1 444 0.7283 0.5905 0.7283 0.8534
No log 11.15 446 0.7559 0.5788 0.7559 0.8694
No log 11.2 448 0.7216 0.5959 0.7216 0.8495
No log 11.25 450 0.6977 0.5370 0.6977 0.8353
No log 11.3 452 0.7005 0.5485 0.7005 0.8370
No log 11.35 454 0.7345 0.5642 0.7345 0.8570
No log 11.4 456 0.7221 0.5975 0.7221 0.8498
No log 11.45 458 0.6921 0.5146 0.6921 0.8319
No log 11.5 460 0.7071 0.4594 0.7071 0.8409
No log 11.55 462 0.6961 0.5594 0.6961 0.8343
No log 11.6 464 0.7398 0.5993 0.7398 0.8601
No log 11.65 466 0.7966 0.5425 0.7966 0.8925
No log 11.7 468 0.7599 0.5462 0.7599 0.8717
No log 11.75 470 0.7354 0.5993 0.7354 0.8576
No log 11.8 472 0.7236 0.6187 0.7236 0.8506
No log 11.85 474 0.7205 0.5495 0.7205 0.8488
No log 11.9 476 0.7191 0.6206 0.7191 0.8480
No log 11.95 478 0.7336 0.6075 0.7336 0.8565
No log 12.0 480 0.7931 0.5650 0.7931 0.8905
No log 12.05 482 0.8842 0.5210 0.8842 0.9403
No log 12.1 484 0.9600 0.4572 0.9600 0.9798
No log 12.15 486 0.9016 0.4459 0.9016 0.9495
No log 12.2 488 0.8222 0.4455 0.8222 0.9068
No log 12.25 490 0.7511 0.4642 0.7511 0.8667
No log 12.3 492 0.7252 0.4537 0.7252 0.8516
No log 12.35 494 0.7172 0.4527 0.7172 0.8469
No log 12.4 496 0.7638 0.5160 0.7638 0.8740
No log 12.45 498 0.8045 0.4924 0.8045 0.8970
0.3142 12.5 500 0.7644 0.5243 0.7644 0.8743
0.3142 12.55 502 0.7444 0.5766 0.7444 0.8628
0.3142 12.6 504 0.7190 0.5534 0.7190 0.8479
0.3142 12.65 506 0.7126 0.5534 0.7126 0.8441
0.3142 12.7 508 0.7039 0.5643 0.7039 0.8390
0.3142 12.75 510 0.7070 0.5691 0.7070 0.8408
0.3142 12.8 512 0.7084 0.5869 0.7084 0.8417
0.3142 12.85 514 0.7059 0.5911 0.7059 0.8402
0.3142 12.9 516 0.7115 0.6241 0.7115 0.8435
0.3142 12.95 518 0.7227 0.5879 0.7227 0.8501
0.3142 13.0 520 0.7413 0.5569 0.7413 0.8610
0.3142 13.05 522 0.7774 0.5543 0.7774 0.8817
0.3142 13.1 524 0.7827 0.5268 0.7827 0.8847
0.3142 13.15 526 0.7678 0.4916 0.7678 0.8763
0.3142 13.2 528 0.7778 0.4455 0.7778 0.8819
0.3142 13.25 530 0.7783 0.4907 0.7783 0.8822
0.3142 13.3 532 0.7602 0.5238 0.7602 0.8719
0.3142 13.35 534 0.7546 0.6232 0.7546 0.8687
0.3142 13.4 536 0.7585 0.5958 0.7585 0.8709
0.3142 13.45 538 0.7498 0.6196 0.7498 0.8659
0.3142 13.5 540 0.7502 0.5161 0.7502 0.8662
0.3142 13.55 542 0.7543 0.5340 0.7543 0.8685
0.3142 13.6 544 0.7871 0.4875 0.7871 0.8872
0.3142 13.65 546 0.7930 0.4014 0.7930 0.8905
0.3142 13.7 548 0.8002 0.3717 0.8002 0.8945
0.3142 13.75 550 0.8013 0.3760 0.8013 0.8951
0.3142 13.8 552 0.8004 0.3740 0.8004 0.8947

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k12_task5_organization

Finetuned
(4019)
this model