ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k17_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.2236
  • Qwk: -0.0704
  • Mse: 1.2236
  • Rmse: 1.1062

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0435 2 3.5873 0.0048 3.5873 1.8940
No log 0.0870 4 1.9188 0.0704 1.9188 1.3852
No log 0.1304 6 1.8244 -0.0284 1.8244 1.3507
No log 0.1739 8 1.4418 0.0 1.4418 1.2008
No log 0.2174 10 1.0361 -0.0423 1.0361 1.0179
No log 0.2609 12 0.8671 -0.0894 0.8671 0.9312
No log 0.3043 14 0.7707 0.0807 0.7707 0.8779
No log 0.3478 16 1.0831 0.0282 1.0831 1.0407
No log 0.3913 18 1.7604 -0.0063 1.7604 1.3268
No log 0.4348 20 1.0971 -0.0987 1.0971 1.0474
No log 0.4783 22 0.7533 0.0460 0.7533 0.8679
No log 0.5217 24 0.8135 -0.0711 0.8135 0.9019
No log 0.5652 26 1.1733 -0.1019 1.1733 1.0832
No log 0.6087 28 1.3305 -0.0234 1.3305 1.1535
No log 0.6522 30 1.0437 -0.0704 1.0437 1.0216
No log 0.6957 32 0.7443 -0.0035 0.7443 0.8627
No log 0.7391 34 0.7115 0.0 0.7115 0.8435
No log 0.7826 36 0.7213 0.0 0.7213 0.8493
No log 0.8261 38 0.7959 -0.1227 0.7959 0.8921
No log 0.8696 40 1.0828 -0.0133 1.0828 1.0406
No log 0.9130 42 1.4774 -0.0247 1.4774 1.2155
No log 0.9565 44 1.5555 0.0 1.5555 1.2472
No log 1.0 46 1.2685 -0.0490 1.2685 1.1263
No log 1.0435 48 1.0666 -0.0457 1.0666 1.0327
No log 1.0870 50 0.8191 -0.0331 0.8191 0.9051
No log 1.1304 52 0.8023 -0.0766 0.8023 0.8957
No log 1.1739 54 0.9323 -0.0079 0.9323 0.9655
No log 1.2174 56 1.1684 -0.0359 1.1684 1.0809
No log 1.2609 58 1.7983 -0.0920 1.7983 1.3410
No log 1.3043 60 1.7539 -0.0920 1.7539 1.3244
No log 1.3478 62 0.9175 0.0017 0.9175 0.9579
No log 1.3913 64 0.7764 -0.0695 0.7764 0.8812
No log 1.4348 66 0.9244 0.0017 0.9244 0.9615
No log 1.4783 68 1.5803 -0.0705 1.5803 1.2571
No log 1.5217 70 1.4264 -0.0456 1.4264 1.1943
No log 1.5652 72 0.9418 0.1064 0.9418 0.9705
No log 1.6087 74 0.7224 -0.0069 0.7224 0.8500
No log 1.6522 76 0.7199 -0.0571 0.7199 0.8485
No log 1.6957 78 0.7249 -0.0035 0.7249 0.8514
No log 1.7391 80 0.8414 0.0512 0.8414 0.9173
No log 1.7826 82 0.8989 0.1243 0.8989 0.9481
No log 1.8261 84 0.9059 -0.0122 0.9059 0.9518
No log 1.8696 86 0.9224 -0.1274 0.9224 0.9604
No log 1.9130 88 0.9303 -0.0923 0.9303 0.9645
No log 1.9565 90 1.0265 -0.1284 1.0265 1.0131
No log 2.0 92 1.1630 -0.1890 1.1630 1.0784
No log 2.0435 94 1.4723 -0.1575 1.4723 1.2134
No log 2.0870 96 1.2552 -0.1912 1.2552 1.1204
No log 2.1304 98 0.8987 0.0191 0.8987 0.9480
No log 2.1739 100 0.8145 -0.0228 0.8145 0.9025
No log 2.2174 102 0.7442 -0.0125 0.7442 0.8627
No log 2.2609 104 0.7494 -0.0033 0.7494 0.8657
No log 2.3043 106 1.0532 -0.0133 1.0532 1.0263
No log 2.3478 108 1.2162 0.0104 1.2162 1.1028
No log 2.3913 110 0.8520 0.0639 0.8520 0.9230
No log 2.4348 112 0.7956 0.0214 0.7956 0.8920
No log 2.4783 114 0.7560 0.0247 0.7560 0.8695
No log 2.5217 116 0.7723 0.0732 0.7723 0.8788
No log 2.5652 118 0.7558 0.0303 0.7558 0.8694
No log 2.6087 120 0.7510 0.0303 0.7510 0.8666
No log 2.6522 122 0.8865 0.0438 0.8865 0.9415
No log 2.6957 124 0.8204 0.1095 0.8204 0.9058
No log 2.7391 126 0.7682 -0.0612 0.7682 0.8765
No log 2.7826 128 0.8061 0.0999 0.8061 0.8978
No log 2.8261 130 0.8352 0.0999 0.8352 0.9139
No log 2.8696 132 0.9415 -0.0425 0.9415 0.9703
No log 2.9130 134 1.1808 -0.0961 1.1808 1.0866
No log 2.9565 136 0.9286 0.0118 0.9286 0.9636
No log 3.0 138 0.8551 0.0412 0.8551 0.9247
No log 3.0435 140 0.8515 -0.0113 0.8515 0.9227
No log 3.0870 142 0.9873 -0.1265 0.9873 0.9937
No log 3.1304 144 0.9507 -0.1265 0.9507 0.9751
No log 3.1739 146 0.8979 -0.0033 0.8979 0.9476
No log 3.2174 148 0.9239 0.0346 0.9239 0.9612
No log 3.2609 150 1.0265 -0.0163 1.0265 1.0132
No log 3.3043 152 1.0163 -0.0175 1.0163 1.0081
No log 3.3478 154 1.0753 0.0086 1.0753 1.0369
No log 3.3913 156 0.8291 -0.0351 0.8291 0.9106
No log 3.4348 158 0.7341 0.0334 0.7341 0.8568
No log 3.4783 160 0.7347 -0.0541 0.7347 0.8572
No log 3.5217 162 0.6994 0.0374 0.6994 0.8363
No log 3.5652 164 0.8091 0.0953 0.8091 0.8995
No log 3.6087 166 0.8118 0.1879 0.8118 0.9010
No log 3.6522 168 0.7890 0.1758 0.7890 0.8883
No log 3.6957 170 0.8519 -0.0008 0.8519 0.9230
No log 3.7391 172 0.8442 -0.0373 0.8442 0.9188
No log 3.7826 174 0.8405 0.0095 0.8405 0.9168
No log 3.8261 176 0.7567 -0.0030 0.7567 0.8699
No log 3.8696 178 0.7737 0.0061 0.7737 0.8796
No log 3.9130 180 0.8059 0.0628 0.8059 0.8977
No log 3.9565 182 0.8467 0.0549 0.8467 0.9202
No log 4.0 184 0.9436 -0.0122 0.9436 0.9714
No log 4.0435 186 0.8004 0.0714 0.8004 0.8947
No log 4.0870 188 0.7665 -0.0662 0.7665 0.8755
No log 4.1304 190 0.7801 -0.1230 0.7801 0.8833
No log 4.1739 192 0.9527 -0.0163 0.9527 0.9761
No log 4.2174 194 0.9607 -0.0532 0.9607 0.9802
No log 4.2609 196 0.8264 -0.0711 0.8264 0.9091
No log 4.3043 198 0.8408 -0.0958 0.8408 0.9170
No log 4.3478 200 0.8507 -0.2126 0.8507 0.9224
No log 4.3913 202 1.0908 -0.0885 1.0908 1.0444
No log 4.4348 204 1.1118 -0.0558 1.1118 1.0544
No log 4.4783 206 0.7727 0.0807 0.7727 0.8790
No log 4.5217 208 0.7449 -0.0571 0.7449 0.8631
No log 4.5652 210 0.7440 -0.0551 0.7440 0.8625
No log 4.6087 212 0.7037 -0.0035 0.7037 0.8389
No log 4.6522 214 0.8129 0.0129 0.8129 0.9016
No log 4.6957 216 0.7934 0.0159 0.7934 0.8907
No log 4.7391 218 0.7361 -0.0069 0.7361 0.8580
No log 4.7826 220 0.7797 0.0395 0.7797 0.8830
No log 4.8261 222 0.8918 0.0095 0.8918 0.9443
No log 4.8696 224 0.8048 -0.0578 0.8048 0.8971
No log 4.9130 226 0.8554 -0.1033 0.8554 0.9249
No log 4.9565 228 0.8200 -0.1397 0.8200 0.9055
No log 5.0 230 0.8574 0.0191 0.8574 0.9260
No log 5.0435 232 1.0885 0.0182 1.0885 1.0433
No log 5.0870 234 0.9400 -0.0442 0.9400 0.9696
No log 5.1304 236 0.7937 0.0296 0.7937 0.8909
No log 5.1739 238 0.7433 -0.0069 0.7433 0.8621
No log 5.2174 240 0.7549 -0.0131 0.7549 0.8689
No log 5.2609 242 0.9307 0.0748 0.9307 0.9647
No log 5.3043 244 0.9539 0.0287 0.9539 0.9767
No log 5.3478 246 0.7977 0.1097 0.7977 0.8931
No log 5.3913 248 0.7936 -0.2056 0.7936 0.8908
No log 5.4348 250 0.7999 -0.0145 0.7999 0.8943
No log 5.4783 252 1.1229 0.0458 1.1229 1.0596
No log 5.5217 254 1.2349 -0.0972 1.2349 1.1112
No log 5.5652 256 0.9067 0.0909 0.9067 0.9522
No log 5.6087 258 0.8488 -0.0996 0.8488 0.9213
No log 5.6522 260 0.8810 -0.0317 0.8810 0.9386
No log 5.6957 262 0.8737 -0.0150 0.8737 0.9347
No log 5.7391 264 0.7822 -0.1067 0.7822 0.8844
No log 5.7826 266 0.6994 0.0 0.6994 0.8363
No log 5.8261 268 0.8856 0.0576 0.8856 0.9410
No log 5.8696 270 1.0770 -0.0658 1.0770 1.0378
No log 5.9130 272 0.9317 -0.0285 0.9317 0.9653
No log 5.9565 274 0.7274 0.0964 0.7274 0.8529
No log 6.0 276 0.6798 0.0 0.6798 0.8245
No log 6.0435 278 0.6925 -0.0035 0.6925 0.8322
No log 6.0870 280 0.7474 0.1097 0.7474 0.8645
No log 6.1304 282 0.8058 0.2141 0.8058 0.8977
No log 6.1739 284 0.7379 0.2258 0.7379 0.8590
No log 6.2174 286 0.7288 -0.0101 0.7288 0.8537
No log 6.2609 288 0.7459 0.1259 0.7459 0.8637
No log 6.3043 290 0.7659 0.2105 0.7659 0.8752
No log 6.3478 292 0.7348 0.0416 0.7348 0.8572
No log 6.3913 294 0.7342 0.0863 0.7342 0.8569
No log 6.4348 296 0.7617 0.1395 0.7617 0.8727
No log 6.4783 298 0.8754 0.0609 0.8754 0.9356
No log 6.5217 300 0.7790 0.1449 0.7790 0.8826
No log 6.5652 302 0.7158 -0.0035 0.7158 0.8461
No log 6.6087 304 0.7402 -0.0571 0.7402 0.8603
No log 6.6522 306 0.7438 -0.0035 0.7438 0.8624
No log 6.6957 308 0.9262 -0.0269 0.9262 0.9624
No log 6.7391 310 1.0451 -0.0638 1.0451 1.0223
No log 6.7826 312 1.0174 -0.0617 1.0174 1.0086
No log 6.8261 314 0.9885 0.0111 0.9885 0.9943
No log 6.8696 316 1.1191 -0.0638 1.1191 1.0579
No log 6.9130 318 1.3108 -0.1289 1.3108 1.1449
No log 6.9565 320 1.0799 -0.0301 1.0799 1.0392
No log 7.0 322 0.8322 0.0146 0.8322 0.9123
No log 7.0435 324 0.7923 -0.0449 0.7923 0.8901
No log 7.0870 326 0.7692 0.1927 0.7692 0.8771
No log 7.1304 328 0.7761 0.1047 0.7761 0.8810
No log 7.1739 330 0.7940 0.0071 0.7940 0.8910
No log 7.2174 332 0.8221 0.0867 0.8221 0.9067
No log 7.2609 334 0.7621 0.1691 0.7621 0.8730
No log 7.3043 336 0.7400 0.1318 0.7400 0.8602
No log 7.3478 338 0.8285 -0.0056 0.8285 0.9102
No log 7.3913 340 0.9057 -0.0101 0.9057 0.9517
No log 7.4348 342 0.9118 -0.0101 0.9118 0.9549
No log 7.4783 344 0.9925 0.0157 0.9925 0.9963
No log 7.5217 346 1.2091 -0.0977 1.2091 1.0996
No log 7.5652 348 1.2220 -0.0977 1.2220 1.1055
No log 7.6087 350 1.1226 -0.0648 1.1226 1.0595
No log 7.6522 352 0.8599 0.1243 0.8599 0.9273
No log 7.6957 354 0.7745 0.0296 0.7745 0.8800
No log 7.7391 356 0.7992 0.1506 0.7992 0.8940
No log 7.7826 358 0.8157 0.0549 0.8157 0.9031
No log 7.8261 360 0.8751 0.1196 0.8751 0.9355
No log 7.8696 362 0.9079 0.1150 0.9079 0.9529
No log 7.9130 364 0.9947 0.0984 0.9947 0.9973
No log 7.9565 366 0.9606 0.0786 0.9606 0.9801
No log 8.0 368 1.0340 0.0986 1.0340 1.0169
No log 8.0435 370 1.1082 -0.0187 1.1082 1.0527
No log 8.0870 372 0.9027 0.0826 0.9027 0.9501
No log 8.1304 374 0.7594 0.1659 0.7594 0.8715
No log 8.1739 376 0.7321 0.0414 0.7321 0.8557
No log 8.2174 378 0.7357 0.1148 0.7357 0.8577
No log 8.2609 380 0.8926 0.0233 0.8926 0.9448
No log 8.3043 382 0.9165 0.0515 0.9165 0.9573
No log 8.3478 384 0.8804 0.0134 0.8804 0.9383
No log 8.3913 386 0.8942 0.0111 0.8942 0.9456
No log 8.4348 388 0.9937 -0.0638 0.9937 0.9968
No log 8.4783 390 0.9994 -0.0638 0.9994 0.9997
No log 8.5217 392 0.8908 0.0224 0.8908 0.9438
No log 8.5652 394 0.8615 0.2092 0.8615 0.9282
No log 8.6087 396 0.8590 0.2092 0.8590 0.9268
No log 8.6522 398 1.0448 -0.0316 1.0448 1.0222
No log 8.6957 400 1.1434 -0.0345 1.1434 1.0693
No log 8.7391 402 0.9523 0.0157 0.9523 0.9759
No log 8.7826 404 0.8124 0.1395 0.8124 0.9013
No log 8.8261 406 0.8611 -0.0101 0.8611 0.9279
No log 8.8696 408 1.1393 -0.0331 1.1393 1.0674
No log 8.9130 410 1.3227 -0.0435 1.3227 1.1501
No log 8.9565 412 1.1419 -0.0359 1.1419 1.0686
No log 9.0 414 0.8799 0.0182 0.8799 0.9380
No log 9.0435 416 0.8485 -0.0122 0.8485 0.9211
No log 9.0870 418 0.9594 -0.0236 0.9594 0.9795
No log 9.1304 420 1.0222 -0.0301 1.0222 1.0111
No log 9.1739 422 1.1046 -0.0359 1.1046 1.0510
No log 9.2174 424 0.9110 0.0134 0.9110 0.9545
No log 9.2609 426 0.7891 0.0377 0.7891 0.8883
No log 9.3043 428 0.7654 0.1716 0.7654 0.8749
No log 9.3478 430 0.8176 0.0676 0.8176 0.9042
No log 9.3913 432 0.9030 0.0157 0.9030 0.9503
No log 9.4348 434 0.8471 0.1064 0.8471 0.9204
No log 9.4783 436 0.8726 0.1024 0.8726 0.9341
No log 9.5217 438 0.8631 0.0711 0.8631 0.9290
No log 9.5652 440 0.7750 0.0525 0.7750 0.8803
No log 9.6087 442 0.7373 0.0914 0.7373 0.8587
No log 9.6522 444 0.7759 0.1449 0.7759 0.8809
No log 9.6957 446 0.9406 -0.0253 0.9406 0.9699
No log 9.7391 448 1.0754 -0.0385 1.0754 1.0370
No log 9.7826 450 0.9554 -0.0949 0.9554 0.9774
No log 9.8261 452 0.8205 0.1453 0.8205 0.9058
No log 9.8696 454 0.7533 -0.0069 0.7533 0.8680
No log 9.9130 456 0.7568 0.0460 0.7568 0.8699
No log 9.9565 458 0.8033 0.1605 0.8033 0.8963
No log 10.0 460 1.0310 -0.0285 1.0310 1.0154
No log 10.0435 462 1.2372 -0.0411 1.2372 1.1123
No log 10.0870 464 1.1316 -0.0372 1.1316 1.0638
No log 10.1304 466 0.8667 0.0182 0.8667 0.9310
No log 10.1739 468 0.7509 0.0909 0.7509 0.8666
No log 10.2174 470 0.7512 0.1318 0.7512 0.8667
No log 10.2609 472 0.8661 0.0946 0.8661 0.9306
No log 10.3043 474 1.0738 -0.0686 1.0738 1.0363
No log 10.3478 476 1.1116 -0.0695 1.1116 1.0543
No log 10.3913 478 0.9415 -0.0583 0.9415 0.9703
No log 10.4348 480 0.7118 0.1021 0.7118 0.8437
No log 10.4783 482 0.7145 -0.1001 0.7145 0.8453
No log 10.5217 484 0.6934 -0.1001 0.6934 0.8327
No log 10.5652 486 0.7433 0.1775 0.7433 0.8621
No log 10.6087 488 1.0448 -0.0218 1.0448 1.0221
No log 10.6522 490 1.2897 -0.0457 1.2897 1.1356
No log 10.6957 492 1.1994 -0.0446 1.1994 1.0952
No log 10.7391 494 0.9223 0.0111 0.9223 0.9604
No log 10.7826 496 0.7059 0.2180 0.7059 0.8401
No log 10.8261 498 0.6763 0.2059 0.6763 0.8224
0.3436 10.8696 500 0.7224 0.1775 0.7224 0.8499
0.3436 10.9130 502 0.8345 0.0909 0.8345 0.9135
0.3436 10.9565 504 0.7751 0.1788 0.7751 0.8804
0.3436 11.0 506 0.6666 0.2424 0.6666 0.8164
0.3436 11.0435 508 0.6499 0.2513 0.6499 0.8062
0.3436 11.0870 510 0.6456 0.2513 0.6456 0.8035
0.3436 11.1304 512 0.6466 0.2513 0.6466 0.8041
0.3436 11.1739 514 0.6910 0.1775 0.6910 0.8313
0.3436 11.2174 516 0.7309 0.2424 0.7309 0.8549
0.3436 11.2609 518 0.7353 0.2077 0.7353 0.8575
0.3436 11.3043 520 0.8186 0.1788 0.8186 0.9048
0.3436 11.3478 522 0.8926 0.0157 0.8926 0.9448
0.3436 11.3913 524 0.8766 0.0157 0.8766 0.9363
0.3436 11.4348 526 0.7862 0.1897 0.7862 0.8867
0.3436 11.4783 528 0.7343 0.1899 0.7343 0.8569
0.3436 11.5217 530 0.7623 0.1660 0.7623 0.8731
0.3436 11.5652 532 0.9095 -0.0218 0.9095 0.9537
0.3436 11.6087 534 1.1604 -0.0992 1.1604 1.0772
0.3436 11.6522 536 1.3269 -0.0758 1.3269 1.1519
0.3436 11.6957 538 1.2236 -0.0704 1.2236 1.1062

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k17_task3_organization

Finetuned
(4023)
this model