ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k10_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8472
  • Qwk: 0.4948
  • Mse: 0.8472
  • Rmse: 0.9204

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0588 2 3.9919 -0.0174 3.9919 1.9980
No log 0.1176 4 2.2112 0.0816 2.2112 1.4870
No log 0.1765 6 1.8036 0.0307 1.8036 1.3430
No log 0.2353 8 1.6503 0.0185 1.6503 1.2847
No log 0.2941 10 1.8896 0.0727 1.8896 1.3746
No log 0.3529 12 1.6729 0.0294 1.6729 1.2934
No log 0.4118 14 1.5004 0.0 1.5004 1.2249
No log 0.4706 16 1.4264 0.0 1.4264 1.1943
No log 0.5294 18 1.3752 0.0 1.3752 1.1727
No log 0.5882 20 1.1862 0.1324 1.1862 1.0891
No log 0.6471 22 1.1246 0.2049 1.1246 1.0604
No log 0.7059 24 1.1967 0.1591 1.1967 1.0939
No log 0.7647 26 1.3957 0.0883 1.3957 1.1814
No log 0.8235 28 1.3742 0.0769 1.3742 1.1722
No log 0.8824 30 1.1998 0.0996 1.1998 1.0954
No log 0.9412 32 1.0725 0.1779 1.0725 1.0356
No log 1.0 34 1.0388 0.2441 1.0388 1.0192
No log 1.0588 36 1.0535 0.1537 1.0535 1.0264
No log 1.1176 38 1.4351 0.1197 1.4351 1.1980
No log 1.1765 40 1.6012 0.0724 1.6012 1.2654
No log 1.2353 42 1.2947 0.1460 1.2947 1.1378
No log 1.2941 44 0.9933 0.2068 0.9933 0.9966
No log 1.3529 46 0.9751 0.2015 0.9751 0.9875
No log 1.4118 48 1.0549 0.2094 1.0549 1.0271
No log 1.4706 50 1.1618 0.2496 1.1618 1.0778
No log 1.5294 52 1.3814 0.2120 1.3814 1.1753
No log 1.5882 54 1.6037 0.1130 1.6037 1.2664
No log 1.6471 56 1.5409 0.0844 1.5409 1.2413
No log 1.7059 58 1.4342 0.1419 1.4342 1.1976
No log 1.7647 60 1.2406 0.1784 1.2406 1.1138
No log 1.8235 62 1.1093 0.1799 1.1093 1.0532
No log 1.8824 64 1.1214 0.2711 1.1214 1.0590
No log 1.9412 66 1.1581 0.2478 1.1581 1.0762
No log 2.0 68 1.2410 0.1967 1.2410 1.1140
No log 2.0588 70 1.2513 0.2120 1.2513 1.1186
No log 2.1176 72 1.2178 0.2998 1.2178 1.1036
No log 2.1765 74 1.0490 0.2872 1.0490 1.0242
No log 2.2353 76 0.9469 0.2812 0.9469 0.9731
No log 2.2941 78 0.9565 0.2128 0.9565 0.9780
No log 2.3529 80 0.9512 0.3615 0.9512 0.9753
No log 2.4118 82 0.9487 0.2984 0.9487 0.9740
No log 2.4706 84 1.0758 0.3833 1.0758 1.0372
No log 2.5294 86 0.9495 0.3646 0.9495 0.9744
No log 2.5882 88 0.8845 0.3666 0.8845 0.9405
No log 2.6471 90 0.8699 0.3540 0.8699 0.9327
No log 2.7059 92 0.8630 0.3985 0.8630 0.9290
No log 2.7647 94 0.8548 0.3250 0.8548 0.9246
No log 2.8235 96 0.8576 0.2693 0.8576 0.9261
No log 2.8824 98 0.8841 0.2742 0.8841 0.9403
No log 2.9412 100 0.9003 0.2742 0.9003 0.9489
No log 3.0 102 0.8797 0.2857 0.8797 0.9379
No log 3.0588 104 0.9160 0.3960 0.9160 0.9571
No log 3.1176 106 0.8808 0.4224 0.8808 0.9385
No log 3.1765 108 0.8363 0.4963 0.8363 0.9145
No log 3.2353 110 0.8313 0.3914 0.8313 0.9118
No log 3.2941 112 0.9651 0.4334 0.9651 0.9824
No log 3.3529 114 1.0125 0.4552 1.0125 1.0063
No log 3.4118 116 0.9286 0.3879 0.9286 0.9636
No log 3.4706 118 0.9312 0.1961 0.9312 0.9650
No log 3.5294 120 0.9741 0.2424 0.9741 0.9870
No log 3.5882 122 0.9667 0.2424 0.9667 0.9832
No log 3.6471 124 0.9065 0.2220 0.9065 0.9521
No log 3.7059 126 0.9120 0.4010 0.9120 0.9550
No log 3.7647 128 0.9695 0.4586 0.9695 0.9846
No log 3.8235 130 0.9041 0.4507 0.9041 0.9508
No log 3.8824 132 0.8678 0.3457 0.8678 0.9316
No log 3.9412 134 0.8757 0.3800 0.8757 0.9358
No log 4.0 136 0.9269 0.4507 0.9269 0.9628
No log 4.0588 138 0.9647 0.4382 0.9647 0.9822
No log 4.1176 140 0.9721 0.4388 0.9721 0.9860
No log 4.1765 142 0.9423 0.5116 0.9423 0.9707
No log 4.2353 144 0.9688 0.5120 0.9688 0.9843
No log 4.2941 146 1.0048 0.4518 1.0048 1.0024
No log 4.3529 148 1.0965 0.3972 1.0965 1.0471
No log 4.4118 150 1.0561 0.4224 1.0561 1.0277
No log 4.4706 152 0.9800 0.3874 0.9800 0.9900
No log 4.5294 154 1.0282 0.3709 1.0282 1.0140
No log 4.5882 156 1.0350 0.3537 1.0350 1.0173
No log 4.6471 158 0.9498 0.3402 0.9498 0.9746
No log 4.7059 160 0.9166 0.3671 0.9166 0.9574
No log 4.7647 162 0.8835 0.3149 0.8835 0.9399
No log 4.8235 164 0.8591 0.3291 0.8591 0.9269
No log 4.8824 166 0.9139 0.3993 0.9139 0.9560
No log 4.9412 168 0.8494 0.4402 0.8494 0.9216
No log 5.0 170 0.7820 0.3967 0.7820 0.8843
No log 5.0588 172 0.8273 0.4231 0.8273 0.9095
No log 5.1176 174 0.9353 0.5102 0.9353 0.9671
No log 5.1765 176 0.8788 0.5447 0.8788 0.9374
No log 5.2353 178 0.7561 0.4690 0.7561 0.8696
No log 5.2941 180 0.7555 0.5430 0.7555 0.8692
No log 5.3529 182 0.7338 0.5205 0.7338 0.8566
No log 5.4118 184 0.7166 0.4048 0.7166 0.8465
No log 5.4706 186 0.8165 0.4570 0.8165 0.9036
No log 5.5294 188 0.8575 0.4326 0.8575 0.9260
No log 5.5882 190 0.8122 0.4662 0.8122 0.9012
No log 5.6471 192 0.8360 0.5184 0.8360 0.9143
No log 5.7059 194 0.8839 0.4838 0.8839 0.9402
No log 5.7647 196 0.9556 0.4526 0.9556 0.9775
No log 5.8235 198 0.9805 0.4023 0.9805 0.9902
No log 5.8824 200 0.9083 0.4419 0.9083 0.9530
No log 5.9412 202 0.8671 0.5029 0.8671 0.9312
No log 6.0 204 0.8585 0.5227 0.8585 0.9266
No log 6.0588 206 0.9240 0.4928 0.9240 0.9612
No log 6.1176 208 1.0047 0.4781 1.0047 1.0023
No log 6.1765 210 0.9231 0.4474 0.9231 0.9608
No log 6.2353 212 0.8462 0.3932 0.8462 0.9199
No log 6.2941 214 0.8307 0.3932 0.8307 0.9114
No log 6.3529 216 0.8205 0.4304 0.8205 0.9058
No log 6.4118 218 0.8556 0.3958 0.8556 0.9250
No log 6.4706 220 0.9059 0.3434 0.9059 0.9518
No log 6.5294 222 0.8926 0.3790 0.8926 0.9448
No log 6.5882 224 0.8061 0.3977 0.8061 0.8978
No log 6.6471 226 0.8230 0.3998 0.8230 0.9072
No log 6.7059 228 0.8108 0.3979 0.8108 0.9005
No log 6.7647 230 0.8057 0.3372 0.8057 0.8976
No log 6.8235 232 0.8254 0.4180 0.8254 0.9085
No log 6.8824 234 0.7888 0.4460 0.7888 0.8882
No log 6.9412 236 0.7301 0.5712 0.7301 0.8544
No log 7.0 238 0.7521 0.5325 0.7521 0.8672
No log 7.0588 240 0.7595 0.5098 0.7595 0.8715
No log 7.1176 242 0.7556 0.5274 0.7556 0.8693
No log 7.1765 244 0.8962 0.4130 0.8962 0.9467
No log 7.2353 246 0.9530 0.4232 0.9530 0.9762
No log 7.2941 248 0.8548 0.5304 0.8548 0.9246
No log 7.3529 250 0.8249 0.4916 0.8249 0.9083
No log 7.4118 252 0.8187 0.4916 0.8187 0.9048
No log 7.4706 254 0.8105 0.4444 0.8105 0.9003
No log 7.5294 256 0.7990 0.4816 0.7990 0.8939
No log 7.5882 258 0.8156 0.4354 0.8156 0.9031
No log 7.6471 260 0.8592 0.3780 0.8592 0.9269
No log 7.7059 262 0.8124 0.4471 0.8124 0.9013
No log 7.7647 264 0.8077 0.5317 0.8077 0.8987
No log 7.8235 266 0.8311 0.5304 0.8311 0.9117
No log 7.8824 268 0.7942 0.4676 0.7942 0.8912
No log 7.9412 270 0.8772 0.4175 0.8772 0.9366
No log 8.0 272 0.8950 0.4565 0.8950 0.9460
No log 8.0588 274 0.8186 0.4471 0.8186 0.9048
No log 8.1176 276 0.8468 0.4962 0.8468 0.9202
No log 8.1765 278 0.8581 0.4962 0.8581 0.9264
No log 8.2353 280 0.8964 0.4723 0.8964 0.9468
No log 8.2941 282 0.8799 0.4620 0.8799 0.9380
No log 8.3529 284 0.8497 0.4459 0.8497 0.9218
No log 8.4118 286 0.8682 0.4749 0.8682 0.9317
No log 8.4706 288 0.8808 0.4503 0.8808 0.9385
No log 8.5294 290 0.8715 0.4619 0.8715 0.9335
No log 8.5882 292 0.8334 0.3874 0.8334 0.9129
No log 8.6471 294 0.8648 0.4080 0.8648 0.9300
No log 8.7059 296 0.8267 0.4393 0.8267 0.9092
No log 8.7647 298 0.7969 0.5463 0.7969 0.8927
No log 8.8235 300 0.8091 0.4995 0.8091 0.8995
No log 8.8824 302 0.8009 0.4109 0.8009 0.8949
No log 8.9412 304 0.8715 0.4080 0.8715 0.9336
No log 9.0 306 0.9460 0.4100 0.9460 0.9726
No log 9.0588 308 1.0101 0.3687 1.0101 1.0050
No log 9.1176 310 1.0391 0.3402 1.0391 1.0194
No log 9.1765 312 1.0001 0.3025 1.0001 1.0000
No log 9.2353 314 0.9259 0.4017 0.9259 0.9623
No log 9.2941 316 0.8893 0.4248 0.8893 0.9430
No log 9.3529 318 0.8864 0.4964 0.8864 0.9415
No log 9.4118 320 0.8329 0.4575 0.8329 0.9126
No log 9.4706 322 0.8125 0.4690 0.8125 0.9014
No log 9.5294 324 0.8124 0.5057 0.8124 0.9014
No log 9.5882 326 0.8219 0.5090 0.8219 0.9066
No log 9.6471 328 0.8987 0.4935 0.8987 0.9480
No log 9.7059 330 0.8612 0.4850 0.8612 0.9280
No log 9.7647 332 0.8082 0.4313 0.8082 0.8990
No log 9.8235 334 0.8582 0.4232 0.8582 0.9264
No log 9.8824 336 0.8615 0.4707 0.8615 0.9282
No log 9.9412 338 0.8116 0.4321 0.8116 0.9009
No log 10.0 340 0.8783 0.4575 0.8783 0.9372
No log 10.0588 342 0.9477 0.4348 0.9477 0.9735
No log 10.1176 344 0.8808 0.4812 0.8808 0.9385
No log 10.1765 346 0.8078 0.5102 0.8078 0.8988
No log 10.2353 348 0.8156 0.4816 0.8156 0.9031
No log 10.2941 350 0.8022 0.4813 0.8022 0.8957
No log 10.3529 352 0.8090 0.5528 0.8090 0.8994
No log 10.4118 354 0.8128 0.5528 0.8128 0.9016
No log 10.4706 356 0.7991 0.5009 0.7991 0.8939
No log 10.5294 358 0.8511 0.4425 0.8511 0.9226
No log 10.5882 360 0.8992 0.4366 0.8992 0.9482
No log 10.6471 362 0.8643 0.5424 0.8643 0.9297
No log 10.7059 364 0.8864 0.5690 0.8864 0.9415
No log 10.7647 366 0.8636 0.5315 0.8636 0.9293
No log 10.8235 368 0.8714 0.3361 0.8714 0.9335
No log 10.8824 370 0.8652 0.3361 0.8652 0.9302
No log 10.9412 372 0.8500 0.4029 0.8500 0.9220
No log 11.0 374 0.8534 0.4129 0.8534 0.9238
No log 11.0588 376 0.8448 0.4248 0.8448 0.9191
No log 11.1176 378 0.8672 0.3282 0.8672 0.9312
No log 11.1765 380 0.8923 0.3332 0.8923 0.9446
No log 11.2353 382 0.8705 0.3765 0.8705 0.9330
No log 11.2941 384 0.8671 0.4337 0.8671 0.9312
No log 11.3529 386 0.8758 0.4754 0.8758 0.9358
No log 11.4118 388 0.8657 0.5002 0.8657 0.9304
No log 11.4706 390 0.8668 0.3765 0.8668 0.9310
No log 11.5294 392 0.8503 0.3503 0.8503 0.9221
No log 11.5882 394 0.8503 0.3363 0.8503 0.9221
No log 11.6471 396 0.8373 0.3363 0.8373 0.9150
No log 11.7059 398 0.8378 0.3363 0.8378 0.9153
No log 11.7647 400 0.8330 0.3633 0.8330 0.9127
No log 11.8235 402 0.8372 0.4898 0.8372 0.9150
No log 11.8824 404 0.8531 0.5315 0.8531 0.9236
No log 11.9412 406 0.8343 0.4690 0.8343 0.9134
No log 12.0 408 0.8280 0.4570 0.8280 0.9100
No log 12.0588 410 0.8277 0.4550 0.8277 0.9098
No log 12.1176 412 0.8535 0.4613 0.8535 0.9238
No log 12.1765 414 0.9119 0.3565 0.9119 0.9549
No log 12.2353 416 0.8600 0.5070 0.8600 0.9273
No log 12.2941 418 0.8049 0.4455 0.8049 0.8972
No log 12.3529 420 0.8337 0.4736 0.8337 0.9131
No log 12.4118 422 0.8294 0.4361 0.8294 0.9107
No log 12.4706 424 0.8589 0.5064 0.8590 0.9268
No log 12.5294 426 0.9216 0.5255 0.9216 0.9600
No log 12.5882 428 0.9242 0.5255 0.9242 0.9614
No log 12.6471 430 0.8725 0.5291 0.8725 0.9341
No log 12.7059 432 0.8436 0.4996 0.8436 0.9185
No log 12.7647 434 0.8324 0.4553 0.8324 0.9123
No log 12.8235 436 0.8355 0.4313 0.8355 0.9141
No log 12.8824 438 0.8439 0.4313 0.8439 0.9187
No log 12.9412 440 0.8651 0.5316 0.8651 0.9301
No log 13.0 442 0.8512 0.4996 0.8512 0.9226
No log 13.0588 444 0.8409 0.4455 0.8409 0.9170
No log 13.1176 446 0.8428 0.4816 0.8428 0.9180
No log 13.1765 448 0.8399 0.4455 0.8399 0.9165
No log 13.2353 450 0.8475 0.5120 0.8475 0.9206
No log 13.2941 452 0.9158 0.5178 0.9158 0.9570
No log 13.3529 454 0.9062 0.5178 0.9062 0.9519
No log 13.4118 456 0.8556 0.5124 0.8556 0.9250
No log 13.4706 458 0.8554 0.4824 0.8554 0.9249
No log 13.5294 460 0.8447 0.4810 0.8447 0.9191
No log 13.5882 462 0.8556 0.4647 0.8556 0.9250
No log 13.6471 464 0.8602 0.4968 0.8602 0.9274
No log 13.7059 466 0.8241 0.4782 0.8241 0.9078
No log 13.7647 468 0.8291 0.4722 0.8291 0.9105
No log 13.8235 470 0.8109 0.4466 0.8109 0.9005
No log 13.8824 472 0.8268 0.4742 0.8268 0.9093
No log 13.9412 474 0.8619 0.4832 0.8619 0.9284
No log 14.0 476 0.9321 0.5208 0.9321 0.9654
No log 14.0588 478 0.9148 0.5119 0.9148 0.9564
No log 14.1176 480 0.8308 0.5328 0.8308 0.9115
No log 14.1765 482 0.8256 0.4824 0.8256 0.9086
No log 14.2353 484 0.8460 0.4957 0.8460 0.9198
No log 14.2941 486 0.8341 0.4704 0.8341 0.9133
No log 14.3529 488 0.8770 0.5517 0.8770 0.9365
No log 14.4118 490 0.9177 0.5885 0.9177 0.9580
No log 14.4706 492 0.8984 0.5204 0.8984 0.9478
No log 14.5294 494 0.8854 0.5040 0.8854 0.9410
No log 14.5882 496 0.8703 0.4277 0.8703 0.9329
No log 14.6471 498 0.8449 0.3967 0.8449 0.9192
0.3081 14.7059 500 0.8366 0.5103 0.8366 0.9147
0.3081 14.7647 502 0.8600 0.4968 0.8600 0.9274
0.3081 14.8235 504 0.8285 0.5199 0.8285 0.9102
0.3081 14.8824 506 0.8079 0.5212 0.8079 0.8988
0.3081 14.9412 508 0.7867 0.5226 0.7867 0.8870
0.3081 15.0 510 0.8007 0.4486 0.8007 0.8948
0.3081 15.0588 512 0.7918 0.4720 0.7918 0.8898
0.3081 15.1176 514 0.7690 0.5340 0.7690 0.8769
0.3081 15.1765 516 0.8000 0.5392 0.8000 0.8944
0.3081 15.2353 518 0.8522 0.5988 0.8522 0.9232
0.3081 15.2941 520 0.8005 0.5392 0.8005 0.8947
0.3081 15.3529 522 0.7578 0.5155 0.7578 0.8705
0.3081 15.4118 524 0.7833 0.4393 0.7833 0.8850
0.3081 15.4706 526 0.7755 0.4377 0.7755 0.8806
0.3081 15.5294 528 0.7499 0.5136 0.7499 0.8659
0.3081 15.5882 530 0.7745 0.5203 0.7745 0.8801
0.3081 15.6471 532 0.7941 0.5489 0.7941 0.8911
0.3081 15.7059 534 0.7952 0.5571 0.7952 0.8918
0.3081 15.7647 536 0.7619 0.5431 0.7619 0.8729
0.3081 15.8235 538 0.7425 0.4691 0.7425 0.8617
0.3081 15.8824 540 0.7526 0.4705 0.7526 0.8675
0.3081 15.9412 542 0.7639 0.5023 0.7639 0.8740
0.3081 16.0 544 0.8225 0.5081 0.8225 0.9069
0.3081 16.0588 546 0.8472 0.4948 0.8472 0.9204

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k10_task5_organization

Finetuned
(4019)
this model