ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k1_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0464
  • Qwk: 0.2239
  • Mse: 1.0464
  • Rmse: 1.0229

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.5 2 4.1928 0.0086 4.1928 2.0476
No log 1.0 4 2.7140 -0.0426 2.7140 1.6474
No log 1.5 6 2.4462 -0.0109 2.4462 1.5640
No log 2.0 8 1.6164 -0.0394 1.6164 1.2714
No log 2.5 10 1.3162 0.1186 1.3162 1.1473
No log 3.0 12 1.3234 0.0928 1.3234 1.1504
No log 3.5 14 1.3008 0.1775 1.3008 1.1405
No log 4.0 16 1.2176 0.2204 1.2176 1.1034
No log 4.5 18 1.1903 0.1961 1.1903 1.0910
No log 5.0 20 1.2690 0.1838 1.2690 1.1265
No log 5.5 22 1.2495 0.0811 1.2495 1.1178
No log 6.0 24 1.1421 0.1351 1.1421 1.0687
No log 6.5 26 1.2025 0.2304 1.2025 1.0966
No log 7.0 28 1.5191 0.1112 1.5191 1.2325
No log 7.5 30 1.2437 0.1497 1.2437 1.1152
No log 8.0 32 1.1490 0.2440 1.1490 1.0719
No log 8.5 34 1.2674 0.1750 1.2674 1.1258
No log 9.0 36 1.1231 0.1435 1.1231 1.0598
No log 9.5 38 1.1231 0.1698 1.1231 1.0598
No log 10.0 40 1.2178 0.0107 1.2178 1.1035
No log 10.5 42 1.1214 0.1698 1.1214 1.0590
No log 11.0 44 1.0860 0.1782 1.0860 1.0421
No log 11.5 46 1.1334 0.1632 1.1334 1.0646
No log 12.0 48 1.0908 0.1969 1.0908 1.0444
No log 12.5 50 1.0856 0.2473 1.0856 1.0419
No log 13.0 52 1.1533 0.1874 1.1533 1.0739
No log 13.5 54 1.2346 0.0075 1.2346 1.1111
No log 14.0 56 1.1525 0.1304 1.1525 1.0736
No log 14.5 58 1.1011 0.1532 1.1011 1.0493
No log 15.0 60 1.2550 0.1697 1.2550 1.1203
No log 15.5 62 1.4027 0.1850 1.4027 1.1844
No log 16.0 64 1.2978 0.2446 1.2978 1.1392
No log 16.5 66 1.1150 0.1830 1.1150 1.0559
No log 17.0 68 1.1261 0.1971 1.1261 1.0612
No log 17.5 70 1.1328 0.1183 1.1328 1.0643
No log 18.0 72 1.1095 0.1426 1.1095 1.0533
No log 18.5 74 1.1292 0.1482 1.1292 1.0627
No log 19.0 76 1.1190 0.1671 1.1190 1.0578
No log 19.5 78 1.1015 0.1160 1.1015 1.0495
No log 20.0 80 1.1057 0.1007 1.1057 1.0515
No log 20.5 82 1.1255 0.1680 1.1255 1.0609
No log 21.0 84 1.1691 0.2494 1.1691 1.0813
No log 21.5 86 1.1732 0.2155 1.1732 1.0831
No log 22.0 88 1.1911 0.2564 1.1911 1.0914
No log 22.5 90 1.1011 0.1625 1.1011 1.0493
No log 23.0 92 1.0976 0.0914 1.0976 1.0477
No log 23.5 94 1.0989 0.1408 1.0989 1.0483
No log 24.0 96 1.0965 0.1476 1.0965 1.0471
No log 24.5 98 1.1233 0.1625 1.1233 1.0598
No log 25.0 100 1.1308 0.1351 1.1308 1.0634
No log 25.5 102 1.1064 0.1292 1.1064 1.0519
No log 26.0 104 1.1248 0.0639 1.1248 1.0605
No log 26.5 106 1.1400 0.0639 1.1400 1.0677
No log 27.0 108 1.1030 0.0854 1.1030 1.0502
No log 27.5 110 1.0984 0.1292 1.0984 1.0480
No log 28.0 112 1.1873 0.2089 1.1873 1.0896
No log 28.5 114 1.2519 0.2260 1.2519 1.1189
No log 29.0 116 1.1940 0.1770 1.1940 1.0927
No log 29.5 118 1.1063 0.1504 1.1063 1.0518
No log 30.0 120 1.1147 0.0517 1.1147 1.0558
No log 30.5 122 1.0841 0.0517 1.0841 1.0412
No log 31.0 124 1.0635 0.1226 1.0635 1.0312
No log 31.5 126 1.1475 0.2027 1.1475 1.0712
No log 32.0 128 1.1873 0.1628 1.1873 1.0896
No log 32.5 130 1.1720 0.1628 1.1720 1.0826
No log 33.0 132 1.0867 0.1893 1.0867 1.0425
No log 33.5 134 1.0530 0.1504 1.0530 1.0262
No log 34.0 136 1.0628 0.1740 1.0628 1.0309
No log 34.5 138 1.0690 0.1284 1.0690 1.0339
No log 35.0 140 1.0626 0.1857 1.0626 1.0308
No log 35.5 142 1.0629 0.1927 1.0629 1.0310
No log 36.0 144 1.0459 0.1935 1.0459 1.0227
No log 36.5 146 1.0335 0.1408 1.0335 1.0166
No log 37.0 148 1.0293 0.1908 1.0293 1.0146
No log 37.5 150 1.1117 0.2448 1.1117 1.0544
No log 38.0 152 1.2236 0.2089 1.2236 1.1061
No log 38.5 154 1.2290 0.2089 1.2290 1.1086
No log 39.0 156 1.1434 0.2130 1.1434 1.0693
No log 39.5 158 1.0766 0.2440 1.0766 1.0376
No log 40.0 160 1.0314 0.2276 1.0314 1.0156
No log 40.5 162 1.0216 0.1408 1.0216 1.0108
No log 41.0 164 1.0147 0.1810 1.0147 1.0073
No log 41.5 166 1.0237 0.2456 1.0237 1.0118
No log 42.0 168 1.0831 0.1970 1.0831 1.0407
No log 42.5 170 1.1249 0.2886 1.1249 1.0606
No log 43.0 172 1.1332 0.2886 1.1332 1.0645
No log 43.5 174 1.0828 0.1970 1.0828 1.0406
No log 44.0 176 1.0339 0.2251 1.0339 1.0168
No log 44.5 178 1.0180 0.2187 1.0180 1.0089
No log 45.0 180 1.0136 0.1783 1.0136 1.0068
No log 45.5 182 1.0364 0.1502 1.0364 1.0180
No log 46.0 184 1.0974 0.1649 1.0974 1.0476
No log 46.5 186 1.1698 0.2979 1.1698 1.0816
No log 47.0 188 1.1988 0.2979 1.1988 1.0949
No log 47.5 190 1.1926 0.2979 1.1926 1.0920
No log 48.0 192 1.1188 0.2308 1.1188 1.0577
No log 48.5 194 1.0634 0.1351 1.0634 1.0312
No log 49.0 196 1.0320 0.1783 1.0320 1.0159
No log 49.5 198 1.0372 0.1589 1.0372 1.0184
No log 50.0 200 1.0413 0.1589 1.0413 1.0204
No log 50.5 202 1.0338 0.1589 1.0338 1.0167
No log 51.0 204 1.0300 0.1589 1.0300 1.0149
No log 51.5 206 1.0321 0.1561 1.0321 1.0159
No log 52.0 208 1.0460 0.1476 1.0460 1.0227
No log 52.5 210 1.0626 0.1137 1.0626 1.0308
No log 53.0 212 1.0529 0.1137 1.0529 1.0261
No log 53.5 214 1.0369 0.1629 1.0369 1.0183
No log 54.0 216 1.0452 0.1589 1.0452 1.0224
No log 54.5 218 1.0515 0.2166 1.0515 1.0254
No log 55.0 220 1.0438 0.2517 1.0438 1.0216
No log 55.5 222 1.0679 0.1502 1.0679 1.0334
No log 56.0 224 1.0991 0.1605 1.0991 1.0484
No log 56.5 226 1.0926 0.1935 1.0926 1.0453
No log 57.0 228 1.0986 0.1868 1.0986 1.0481
No log 57.5 230 1.0891 0.1528 1.0891 1.0436
No log 58.0 232 1.0561 0.1418 1.0561 1.0277
No log 58.5 234 1.0424 0.2035 1.0424 1.0210
No log 59.0 236 1.0437 0.2114 1.0437 1.0216
No log 59.5 238 1.0492 0.1713 1.0492 1.0243
No log 60.0 240 1.0494 0.2517 1.0494 1.0244
No log 60.5 242 1.0735 0.1573 1.0735 1.0361
No log 61.0 244 1.1580 0.2308 1.1580 1.0761
No log 61.5 246 1.2289 0.1911 1.2289 1.1086
No log 62.0 248 1.2363 0.1911 1.2363 1.1119
No log 62.5 250 1.1860 0.2448 1.1860 1.0890
No log 63.0 252 1.1514 0.1893 1.1514 1.0730
No log 63.5 254 1.1365 0.1435 1.1365 1.0661
No log 64.0 256 1.1249 0.1076 1.1249 1.0606
No log 64.5 258 1.1316 0.1199 1.1316 1.0638
No log 65.0 260 1.1288 0.1199 1.1288 1.0624
No log 65.5 262 1.1268 0.1555 1.1268 1.0615
No log 66.0 264 1.1234 0.1702 1.1234 1.0599
No log 66.5 266 1.1228 0.1702 1.1228 1.0596
No log 67.0 268 1.1158 0.1919 1.1158 1.0563
No log 67.5 270 1.1053 0.1919 1.1053 1.0513
No log 68.0 272 1.0890 0.1011 1.0890 1.0436
No log 68.5 274 1.0825 0.1292 1.0825 1.0404
No log 69.0 276 1.0807 0.1601 1.0807 1.0396
No log 69.5 278 1.0777 0.1601 1.0777 1.0381
No log 70.0 280 1.0750 0.1783 1.0750 1.0368
No log 70.5 282 1.0727 0.1783 1.0727 1.0357
No log 71.0 284 1.0769 0.1989 1.0769 1.0377
No log 71.5 286 1.0790 0.1589 1.0790 1.0388
No log 72.0 288 1.0761 0.1935 1.0761 1.0374
No log 72.5 290 1.0825 0.1601 1.0825 1.0404
No log 73.0 292 1.1036 0.1823 1.1036 1.0505
No log 73.5 294 1.1138 0.1676 1.1138 1.0554
No log 74.0 296 1.1167 0.1676 1.1167 1.0567
No log 74.5 298 1.1208 0.1676 1.1208 1.0587
No log 75.0 300 1.1273 0.1528 1.1273 1.0617
No log 75.5 302 1.1347 0.1868 1.1347 1.0652
No log 76.0 304 1.1297 0.1528 1.1297 1.0629
No log 76.5 306 1.1128 0.1322 1.1128 1.0549
No log 77.0 308 1.0874 0.1474 1.0874 1.0428
No log 77.5 310 1.0684 0.1292 1.0684 1.0336
No log 78.0 312 1.0574 0.1447 1.0574 1.0283
No log 78.5 314 1.0531 0.1601 1.0531 1.0262
No log 79.0 316 1.0507 0.2187 1.0507 1.0250
No log 79.5 318 1.0528 0.2187 1.0528 1.0260
No log 80.0 320 1.0554 0.1935 1.0554 1.0273
No log 80.5 322 1.0566 0.2088 1.0566 1.0279
No log 81.0 324 1.0581 0.1685 1.0581 1.0287
No log 81.5 326 1.0589 0.1685 1.0589 1.0290
No log 82.0 328 1.0616 0.1284 1.0616 1.0304
No log 82.5 330 1.0650 0.1864 1.0650 1.0320
No log 83.0 332 1.0673 0.1864 1.0673 1.0331
No log 83.5 334 1.0687 0.1067 1.0687 1.0338
No log 84.0 336 1.0664 0.1465 1.0664 1.0327
No log 84.5 338 1.0609 0.2265 1.0609 1.0300
No log 85.0 340 1.0561 0.2340 1.0561 1.0277
No log 85.5 342 1.0551 0.2187 1.0551 1.0272
No log 86.0 344 1.0557 0.1601 1.0557 1.0275
No log 86.5 346 1.0580 0.1292 1.0580 1.0286
No log 87.0 348 1.0585 0.1292 1.0585 1.0288
No log 87.5 350 1.0579 0.1292 1.0579 1.0285
No log 88.0 352 1.0600 0.1292 1.0600 1.0296
No log 88.5 354 1.0631 0.1292 1.0631 1.0311
No log 89.0 356 1.0680 0.1292 1.0680 1.0334
No log 89.5 358 1.0698 0.1292 1.0698 1.0343
No log 90.0 360 1.0677 0.1292 1.0677 1.0333
No log 90.5 362 1.0617 0.1292 1.0617 1.0304
No log 91.0 364 1.0551 0.1601 1.0551 1.0272
No log 91.5 366 1.0503 0.2340 1.0503 1.0248
No log 92.0 368 1.0479 0.2340 1.0479 1.0237
No log 92.5 370 1.0468 0.2088 1.0468 1.0231
No log 93.0 372 1.0457 0.2088 1.0457 1.0226
No log 93.5 374 1.0452 0.2088 1.0452 1.0224
No log 94.0 376 1.0449 0.2088 1.0449 1.0222
No log 94.5 378 1.0452 0.2088 1.0452 1.0223
No log 95.0 380 1.0454 0.2239 1.0454 1.0225
No log 95.5 382 1.0460 0.2239 1.0460 1.0227
No log 96.0 384 1.0461 0.2667 1.0461 1.0228
No log 96.5 386 1.0464 0.2667 1.0464 1.0230
No log 97.0 388 1.0466 0.2667 1.0466 1.0231
No log 97.5 390 1.0466 0.2667 1.0466 1.0231
No log 98.0 392 1.0465 0.2667 1.0465 1.0230
No log 98.5 394 1.0465 0.2667 1.0465 1.0230
No log 99.0 396 1.0465 0.2667 1.0465 1.0230
No log 99.5 398 1.0464 0.2667 1.0464 1.0229
No log 100.0 400 1.0464 0.2239 1.0464 1.0229

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k1_task5_organization

Finetuned
(4019)
this model