ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k8_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0432
  • Qwk: 0.5755
  • Mse: 1.0432
  • Rmse: 1.0213

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0526 2 7.4764 -0.0211 7.4764 2.7343
No log 0.1053 4 4.3570 0.0417 4.3570 2.0873
No log 0.1579 6 3.1163 0.1222 3.1163 1.7653
No log 0.2105 8 2.2323 0.2013 2.2323 1.4941
No log 0.2632 10 2.0576 0.0992 2.0576 1.4344
No log 0.3158 12 2.7089 -0.0448 2.7089 1.6459
No log 0.3684 14 1.8907 0.1416 1.8907 1.3750
No log 0.4211 16 1.5998 0.1538 1.5998 1.2648
No log 0.4737 18 1.6607 0.1887 1.6607 1.2887
No log 0.5263 20 1.8446 0.2787 1.8446 1.3582
No log 0.5789 22 2.0390 0.1940 2.0390 1.4279
No log 0.6316 24 2.0476 0.25 2.0476 1.4310
No log 0.6842 26 2.1884 0.2302 2.1884 1.4793
No log 0.7368 28 2.4291 0.1103 2.4291 1.5585
No log 0.7895 30 2.3907 0.1259 2.3907 1.5462
No log 0.8421 32 2.2014 0.1259 2.2014 1.4837
No log 0.8947 34 2.2126 0.1259 2.2126 1.4875
No log 0.9474 36 2.1302 0.2429 2.1302 1.4595
No log 1.0 38 2.2622 0.1678 2.2622 1.5040
No log 1.0526 40 2.1343 0.1972 2.1343 1.4609
No log 1.1053 42 1.9601 0.2286 1.9601 1.4001
No log 1.1579 44 1.8686 0.2754 1.8686 1.3670
No log 1.2105 46 1.7659 0.3066 1.7659 1.3289
No log 1.2632 48 1.7944 0.3066 1.7944 1.3395
No log 1.3158 50 2.0679 0.2014 2.0679 1.4380
No log 1.3684 52 2.1497 0.2083 2.1497 1.4662
No log 1.4211 54 1.9674 0.3014 1.9674 1.4026
No log 1.4737 56 1.7112 0.4148 1.7112 1.3081
No log 1.5263 58 1.6026 0.3721 1.6026 1.2659
No log 1.5789 60 1.5866 0.3065 1.5866 1.2596
No log 1.6316 62 1.4444 0.4298 1.4444 1.2018
No log 1.6842 64 1.2536 0.4138 1.2536 1.1197
No log 1.7368 66 1.3868 0.3966 1.3868 1.1776
No log 1.7895 68 1.4822 0.35 1.4822 1.2175
No log 1.8421 70 1.3012 0.3770 1.3012 1.1407
No log 1.8947 72 1.3597 0.3492 1.3597 1.1661
No log 1.9474 74 1.3936 0.4427 1.3936 1.1805
No log 2.0 76 1.3263 0.4651 1.3263 1.1516
No log 2.0526 78 1.3097 0.4567 1.3097 1.1444
No log 2.1053 80 1.2873 0.4848 1.2873 1.1346
No log 2.1579 82 1.1881 0.5156 1.1881 1.0900
No log 2.2105 84 1.2886 0.4154 1.2886 1.1352
No log 2.2632 86 1.4831 0.4412 1.4831 1.2178
No log 2.3158 88 1.4214 0.4211 1.4214 1.1922
No log 2.3684 90 1.3996 0.4308 1.3996 1.1831
No log 2.4211 92 1.2706 0.4444 1.2706 1.1272
No log 2.4737 94 1.1897 0.4921 1.1897 1.0907
No log 2.5263 96 1.1518 0.4806 1.1518 1.0732
No log 2.5789 98 1.0599 0.5882 1.0599 1.0295
No log 2.6316 100 1.0399 0.5882 1.0399 1.0197
No log 2.6842 102 1.0416 0.5714 1.0416 1.0206
No log 2.7368 104 1.1532 0.5652 1.1532 1.0739
No log 2.7895 106 1.2850 0.5390 1.2850 1.1336
No log 2.8421 108 1.4356 0.5419 1.4356 1.1982
No log 2.8947 110 1.1619 0.5674 1.1619 1.0779
No log 2.9474 112 1.0530 0.5926 1.0530 1.0261
No log 3.0 114 1.0444 0.5714 1.0444 1.0219
No log 3.0526 116 0.9222 0.5781 0.9222 0.9603
No log 3.1053 118 0.9313 0.5938 0.9313 0.9650
No log 3.1579 120 0.9888 0.6047 0.9888 0.9944
No log 3.2105 122 0.9870 0.6202 0.9870 0.9935
No log 3.2632 124 0.9903 0.6154 0.9903 0.9951
No log 3.3158 126 0.9932 0.5954 0.9932 0.9966
No log 3.3684 128 1.0564 0.5816 1.0564 1.0278
No log 3.4211 130 1.0969 0.5816 1.0969 1.0474
No log 3.4737 132 1.2594 0.5185 1.2594 1.1223
No log 3.5263 134 1.4205 0.4219 1.4205 1.1918
No log 3.5789 136 1.2047 0.5512 1.2047 1.0976
No log 3.6316 138 1.0365 0.6202 1.0365 1.0181
No log 3.6842 140 0.9902 0.6107 0.9902 0.9951
No log 3.7368 142 1.1226 0.5606 1.1226 1.0595
No log 3.7895 144 1.3832 0.4154 1.3832 1.1761
No log 3.8421 146 1.5051 0.4091 1.5051 1.2268
No log 3.8947 148 1.3237 0.4806 1.3237 1.1505
No log 3.9474 150 1.1453 0.4640 1.1453 1.0702
No log 4.0 152 1.0937 0.5714 1.0937 1.0458
No log 4.0526 154 1.0553 0.5736 1.0553 1.0273
No log 4.1053 156 0.9833 0.6 0.9833 0.9916
No log 4.1579 158 0.9431 0.5846 0.9431 0.9712
No log 4.2105 160 0.9488 0.5758 0.9488 0.9741
No log 4.2632 162 1.0016 0.5455 1.0016 1.0008
No log 4.3158 164 1.0155 0.5455 1.0155 1.0077
No log 4.3684 166 0.9666 0.5802 0.9666 0.9832
No log 4.4211 168 0.9615 0.6202 0.9615 0.9805
No log 4.4737 170 1.0098 0.5891 1.0098 1.0049
No log 4.5263 172 1.0859 0.5385 1.0859 1.0421
No log 4.5789 174 1.1565 0.5038 1.1565 1.0754
No log 4.6316 176 1.0833 0.5426 1.0833 1.0408
No log 4.6842 178 1.1105 0.5303 1.1105 1.0538
No log 4.7368 180 1.1031 0.5755 1.1031 1.0503
No log 4.7895 182 0.8821 0.5942 0.8821 0.9392
No log 4.8421 184 0.8084 0.6901 0.8084 0.8991
No log 4.8947 186 0.9427 0.6713 0.9427 0.9709
No log 4.9474 188 1.0081 0.5674 1.0081 1.0040
No log 5.0 190 0.9169 0.6957 0.9169 0.9575
No log 5.0526 192 0.9244 0.6917 0.9244 0.9615
No log 5.1053 194 0.9715 0.5736 0.9715 0.9857
No log 5.1579 196 0.9718 0.5758 0.9718 0.9858
No log 5.2105 198 0.9627 0.5865 0.9627 0.9812
No log 5.2632 200 0.9608 0.5970 0.9608 0.9802
No log 5.3158 202 0.9298 0.5970 0.9298 0.9643
No log 5.3684 204 0.9639 0.6099 0.9639 0.9818
No log 5.4211 206 1.0515 0.5899 1.0515 1.0254
No log 5.4737 208 0.9815 0.6074 0.9815 0.9907
No log 5.5263 210 0.9223 0.6767 0.9223 0.9604
No log 5.5789 212 0.9834 0.6202 0.9834 0.9916
No log 5.6316 214 0.9930 0.5669 0.9930 0.9965
No log 5.6842 216 0.9888 0.5736 0.9888 0.9944
No log 5.7368 218 1.0219 0.5496 1.0219 1.0109
No log 5.7895 220 1.2339 0.5655 1.2339 1.1108
No log 5.8421 222 1.2120 0.5694 1.2120 1.1009
No log 5.8947 224 1.0166 0.5564 1.0166 1.0082
No log 5.9474 226 0.8719 0.6418 0.8719 0.9338
No log 6.0 228 0.8581 0.6716 0.8581 0.9264
No log 6.0526 230 0.8873 0.6364 0.8873 0.9420
No log 6.1053 232 0.9502 0.5692 0.9502 0.9748
No log 6.1579 234 0.9642 0.5781 0.9642 0.9819
No log 6.2105 236 0.9499 0.6094 0.9499 0.9746
No log 6.2632 238 0.9324 0.6357 0.9324 0.9656
No log 6.3158 240 0.9285 0.6565 0.9285 0.9636
No log 6.3684 242 0.9138 0.6462 0.9138 0.9560
No log 6.4211 244 0.9257 0.5954 0.9257 0.9622
No log 6.4737 246 1.0256 0.5385 1.0256 1.0127
No log 6.5263 248 1.0125 0.5263 1.0125 1.0063
No log 6.5789 250 0.8934 0.6119 0.8934 0.9452
No log 6.6316 252 0.8296 0.7143 0.8296 0.9108
No log 6.6842 254 0.8457 0.6857 0.8457 0.9196
No log 6.7368 256 0.8443 0.6483 0.8443 0.9189
No log 6.7895 258 0.9017 0.6447 0.9017 0.9496
No log 6.8421 260 1.0572 0.6065 1.0572 1.0282
No log 6.8947 262 0.9772 0.6538 0.9772 0.9885
No log 6.9474 264 0.8256 0.6423 0.8256 0.9086
No log 7.0 266 0.8399 0.6906 0.8399 0.9165
No log 7.0526 268 0.8965 0.6667 0.8965 0.9468
No log 7.1053 270 0.9318 0.6565 0.9318 0.9653
No log 7.1579 272 0.9926 0.6 0.9926 0.9963
No log 7.2105 274 0.9320 0.6260 0.9320 0.9654
No log 7.2632 276 0.8866 0.6308 0.8866 0.9416
No log 7.3158 278 0.8660 0.6617 0.8660 0.9306
No log 7.3684 280 0.8923 0.5581 0.8923 0.9446
No log 7.4211 282 0.9539 0.5538 0.9539 0.9767
No log 7.4737 284 1.0298 0.5538 1.0298 1.0148
No log 7.5263 286 1.0130 0.5821 1.0130 1.0065
No log 7.5789 288 0.9196 0.5538 0.9196 0.9590
No log 7.6316 290 0.8757 0.5692 0.8757 0.9358
No log 7.6842 292 0.8515 0.6565 0.8515 0.9228
No log 7.7368 294 0.8553 0.6615 0.8553 0.9248
No log 7.7895 296 0.8622 0.6061 0.8622 0.9285
No log 7.8421 298 0.8940 0.5692 0.8940 0.9455
No log 7.8947 300 0.8836 0.5692 0.8836 0.9400
No log 7.9474 302 0.8540 0.6212 0.8540 0.9241
No log 8.0 304 0.8569 0.6466 0.8569 0.9257
No log 8.0526 306 0.8646 0.6107 0.8646 0.9298
No log 8.1053 308 0.9155 0.5649 0.9155 0.9568
No log 8.1579 310 0.9695 0.5672 0.9695 0.9846
No log 8.2105 312 0.9279 0.5821 0.9279 0.9633
No log 8.2632 314 0.8446 0.6569 0.8446 0.9190
No log 8.3158 316 0.8447 0.6519 0.8447 0.9191
No log 8.3684 318 0.8900 0.6569 0.8900 0.9434
No log 8.4211 320 0.9166 0.6567 0.9166 0.9574
No log 8.4737 322 0.9363 0.6462 0.9363 0.9676
No log 8.5263 324 0.9943 0.5512 0.9943 0.9972
No log 8.5789 326 1.0157 0.5238 1.0157 1.0078
No log 8.6316 328 0.9664 0.5397 0.9664 0.9831
No log 8.6842 330 0.9129 0.6061 0.9129 0.9554
No log 8.7368 332 0.8846 0.6364 0.8846 0.9405
No log 8.7895 334 0.8812 0.6767 0.8812 0.9387
No log 8.8421 336 0.8731 0.6767 0.8731 0.9344
No log 8.8947 338 0.8532 0.6418 0.8532 0.9237
No log 8.9474 340 0.8511 0.6471 0.8511 0.9225
No log 9.0 342 0.8471 0.6714 0.8471 0.9204
No log 9.0526 344 0.8559 0.6667 0.8559 0.9251
No log 9.1053 346 0.8812 0.6667 0.8812 0.9387
No log 9.1579 348 0.8927 0.6667 0.8927 0.9448
No log 9.2105 350 0.8991 0.6912 0.8991 0.9482
No log 9.2632 352 0.8864 0.6667 0.8864 0.9415
No log 9.3158 354 0.8934 0.6316 0.8934 0.9452
No log 9.3684 356 0.9299 0.5970 0.9299 0.9643
No log 9.4211 358 0.9334 0.6087 0.9334 0.9661
No log 9.4737 360 0.8730 0.6316 0.8730 0.9343
No log 9.5263 362 0.8621 0.6618 0.8621 0.9285
No log 9.5789 364 0.8818 0.6471 0.8818 0.9391
No log 9.6316 366 0.9305 0.5758 0.9305 0.9646
No log 9.6842 368 1.0185 0.5938 1.0185 1.0092
No log 9.7368 370 1.0519 0.5891 1.0519 1.0256
No log 9.7895 372 0.9977 0.5891 0.9977 0.9989
No log 9.8421 374 0.9013 0.6061 0.9013 0.9494
No log 9.8947 376 0.8883 0.6316 0.8883 0.9425
No log 9.9474 378 0.8896 0.6370 0.8896 0.9432
No log 10.0 380 0.9053 0.6861 0.9053 0.9514
No log 10.0526 382 0.9130 0.6716 0.9130 0.9555
No log 10.1053 384 0.9414 0.6061 0.9414 0.9703
No log 10.1579 386 1.1173 0.5481 1.1173 1.0570
No log 10.2105 388 1.2484 0.5143 1.2484 1.1173
No log 10.2632 390 1.2306 0.5143 1.2306 1.1093
No log 10.3158 392 1.0956 0.5116 1.0956 1.0467
No log 10.3684 394 0.9902 0.5846 0.9902 0.9951
No log 10.4211 396 0.9364 0.6418 0.9364 0.9677
No log 10.4737 398 0.9061 0.6519 0.9061 0.9519
No log 10.5263 400 0.8927 0.6418 0.8927 0.9449
No log 10.5789 402 1.0137 0.5532 1.0137 1.0068
No log 10.6316 404 1.2487 0.5844 1.2487 1.1175
No log 10.6842 406 1.3015 0.5621 1.3015 1.1408
No log 10.7368 408 1.2019 0.5658 1.2019 1.0963
No log 10.7895 410 1.0223 0.5362 1.0223 1.0111
No log 10.8421 412 0.9193 0.6061 0.9193 0.9588
No log 10.8947 414 0.9466 0.6618 0.9466 0.9729
No log 10.9474 416 0.9633 0.6370 0.9633 0.9815
No log 11.0 418 0.9761 0.5714 0.9761 0.9880
No log 11.0526 420 1.0293 0.5197 1.0293 1.0146
No log 11.1053 422 1.0277 0.5312 1.0277 1.0138
No log 11.1579 424 0.9575 0.5692 0.9575 0.9785
No log 11.2105 426 0.9226 0.5865 0.9226 0.9605
No log 11.2632 428 0.9255 0.6165 0.9255 0.9620
No log 11.3158 430 0.9739 0.6154 0.9739 0.9868
No log 11.3684 432 1.0515 0.5469 1.0515 1.0254
No log 11.4211 434 1.0410 0.5469 1.0410 1.0203
No log 11.4737 436 1.0220 0.5469 1.0220 1.0110
No log 11.5263 438 0.9884 0.5469 0.9884 0.9942
No log 11.5789 440 0.9593 0.5512 0.9593 0.9794
No log 11.6316 442 0.9261 0.5581 0.9261 0.9623
No log 11.6842 444 0.8926 0.5692 0.8926 0.9448
No log 11.7368 446 0.8625 0.6364 0.8625 0.9287
No log 11.7895 448 0.8703 0.6763 0.8703 0.9329
No log 11.8421 450 0.8890 0.6471 0.8890 0.9429
No log 11.8947 452 0.8686 0.6715 0.8686 0.9320
No log 11.9474 454 0.8447 0.6471 0.8447 0.9191
No log 12.0 456 0.8400 0.6815 0.8400 0.9165
No log 12.0526 458 0.8452 0.7007 0.8452 0.9194
No log 12.1053 460 0.8641 0.6957 0.8641 0.9296
No log 12.1579 462 0.8513 0.7007 0.8513 0.9227
No log 12.2105 464 0.8310 0.6963 0.8310 0.9116
No log 12.2632 466 0.8531 0.6667 0.8531 0.9236
No log 12.3158 468 0.8774 0.6471 0.8774 0.9367
No log 12.3684 470 0.9402 0.5775 0.9402 0.9696
No log 12.4211 472 0.9754 0.5594 0.9754 0.9876
No log 12.4737 474 0.9809 0.5594 0.9809 0.9904
No log 12.5263 476 0.9470 0.5594 0.9470 0.9731
No log 12.5789 478 0.8977 0.6316 0.8977 0.9475
No log 12.6316 480 0.8870 0.6617 0.8870 0.9418
No log 12.6842 482 0.8969 0.6269 0.8969 0.9471
No log 12.7368 484 0.9068 0.6618 0.9068 0.9522
No log 12.7895 486 0.9222 0.6212 0.9222 0.9603
No log 12.8421 488 0.9411 0.5692 0.9411 0.9701
No log 12.8947 490 0.9941 0.5581 0.9941 0.9970
No log 12.9474 492 1.0233 0.5755 1.0233 1.0116
No log 13.0 494 1.0266 0.6014 1.0266 1.0132
No log 13.0526 496 0.9467 0.6014 0.9467 0.9730
No log 13.1053 498 0.8733 0.6475 0.8733 0.9345
0.3835 13.1579 500 0.8726 0.6667 0.8726 0.9342
0.3835 13.2105 502 0.9145 0.6667 0.9145 0.9563
0.3835 13.2632 504 0.9266 0.6471 0.9266 0.9626
0.3835 13.3158 506 0.8935 0.6370 0.8935 0.9452
0.3835 13.3684 508 0.9461 0.5926 0.9461 0.9727
0.3835 13.4211 510 1.0432 0.5755 1.0432 1.0213

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k8_task1_organization

Finetuned
(4023)
this model