ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k1_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.2253
  • Qwk: 0.0548
  • Mse: 1.2253
  • Rmse: 1.1069

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.5 2 4.1309 0.0086 4.1309 2.0325
No log 1.0 4 2.5360 -0.0040 2.5360 1.5925
No log 1.5 6 2.2236 -0.0007 2.2236 1.4912
No log 2.0 8 1.5782 0.0341 1.5782 1.2563
No log 2.5 10 1.3638 0.0446 1.3638 1.1678
No log 3.0 12 1.7801 0.0520 1.7801 1.3342
No log 3.5 14 1.6635 -0.0038 1.6635 1.2898
No log 4.0 16 1.4265 0.1131 1.4265 1.1944
No log 4.5 18 1.4948 0.2101 1.4948 1.2226
No log 5.0 20 1.4596 0.0399 1.4596 1.2081
No log 5.5 22 1.9566 -0.0095 1.9566 1.3988
No log 6.0 24 1.9223 0.0194 1.9223 1.3865
No log 6.5 26 1.3609 0.0549 1.3609 1.1666
No log 7.0 28 1.3171 0.1221 1.3171 1.1477
No log 7.5 30 1.3632 0.0693 1.3632 1.1675
No log 8.0 32 1.4829 -0.0102 1.4829 1.2178
No log 8.5 34 1.3532 0.0427 1.3532 1.1633
No log 9.0 36 1.3265 0.0545 1.3265 1.1517
No log 9.5 38 1.3043 0.0730 1.3043 1.1420
No log 10.0 40 1.3527 0.0882 1.3527 1.1630
No log 10.5 42 1.3544 0.1304 1.3544 1.1638
No log 11.0 44 1.2793 0.0639 1.2793 1.1310
No log 11.5 46 1.3294 0.0839 1.3294 1.1530
No log 12.0 48 1.3037 0.0961 1.3037 1.1418
No log 12.5 50 1.2570 0.0792 1.2570 1.1212
No log 13.0 52 1.4620 0.0166 1.4620 1.2091
No log 13.5 54 1.6125 -0.0488 1.6125 1.2698
No log 14.0 56 1.4917 -0.0320 1.4917 1.2213
No log 14.5 58 1.2859 0.0640 1.2859 1.1340
No log 15.0 60 1.2640 0.1161 1.2640 1.1243
No log 15.5 62 1.3079 0.0839 1.3079 1.1436
No log 16.0 64 1.2670 0.0610 1.2670 1.1256
No log 16.5 66 1.3114 -0.0764 1.3114 1.1451
No log 17.0 68 1.2587 0.0400 1.2587 1.1219
No log 17.5 70 1.2253 0.1493 1.2253 1.1069
No log 18.0 72 1.2089 0.1125 1.2089 1.0995
No log 18.5 74 1.2338 0.0791 1.2338 1.1108
No log 19.0 76 1.2533 0.0312 1.2533 1.1195
No log 19.5 78 1.2237 0.1062 1.2237 1.1062
No log 20.0 80 1.2581 0.0970 1.2581 1.1216
No log 20.5 82 1.2219 0.1454 1.2219 1.1054
No log 21.0 84 1.2100 0.1186 1.2100 1.1000
No log 21.5 86 1.2269 0.1361 1.2269 1.1076
No log 22.0 88 1.3118 0.0909 1.3118 1.1453
No log 22.5 90 1.3522 0.0820 1.3522 1.1628
No log 23.0 92 1.2734 0.0405 1.2734 1.1284
No log 23.5 94 1.1871 0.1671 1.1871 1.0895
No log 24.0 96 1.1809 0.0822 1.1809 1.0867
No log 24.5 98 1.1865 0.0822 1.1865 1.0893
No log 25.0 100 1.2105 0.1183 1.2105 1.1002
No log 25.5 102 1.2150 0.1183 1.2150 1.1022
No log 26.0 104 1.2338 0.1240 1.2338 1.1108
No log 26.5 106 1.2188 0.0882 1.2188 1.1040
No log 27.0 108 1.1993 0.1067 1.1993 1.0951
No log 27.5 110 1.1939 0.0485 1.1939 1.0927
No log 28.0 112 1.2014 0.0022 1.2014 1.0961
No log 28.5 114 1.2157 0.0358 1.2157 1.1026
No log 29.0 116 1.1884 0.0331 1.1884 1.0902
No log 29.5 118 1.2283 0.1033 1.2283 1.1083
No log 30.0 120 1.2957 0.1591 1.2957 1.1383
No log 30.5 122 1.2696 0.1176 1.2696 1.1268
No log 31.0 124 1.1839 0.1794 1.1839 1.0881
No log 31.5 126 1.2495 0.0574 1.2495 1.1178
No log 32.0 128 1.3496 0.1136 1.3496 1.1617
No log 32.5 130 1.3202 0.0492 1.3202 1.1490
No log 33.0 132 1.2329 0.0078 1.2329 1.1104
No log 33.5 134 1.1797 0.1218 1.1797 1.0861
No log 34.0 136 1.2072 0.0882 1.2072 1.0987
No log 34.5 138 1.2336 0.1268 1.2336 1.1107
No log 35.0 140 1.2704 0.1176 1.2704 1.1271
No log 35.5 142 1.2612 0.1296 1.2612 1.1230
No log 36.0 144 1.2166 0.1276 1.2166 1.1030
No log 36.5 146 1.2026 0.0974 1.2026 1.0967
No log 37.0 148 1.2222 0.0177 1.2222 1.1055
No log 37.5 150 1.2649 0.0234 1.2649 1.1247
No log 38.0 152 1.2773 0.0234 1.2773 1.1302
No log 38.5 154 1.2531 -0.0011 1.2531 1.1194
No log 39.0 156 1.2259 0.0914 1.2259 1.1072
No log 39.5 158 1.2534 0.0791 1.2534 1.1196
No log 40.0 160 1.3111 0.0731 1.3111 1.1450
No log 40.5 162 1.3149 0.0493 1.3149 1.1467
No log 41.0 164 1.2689 0.0400 1.2689 1.1265
No log 41.5 166 1.2265 0.1096 1.2265 1.1075
No log 42.0 168 1.2174 0.0517 1.2174 1.1034
No log 42.5 170 1.2222 0.0331 1.2222 1.1056
No log 43.0 172 1.2431 -0.0011 1.2431 1.1149
No log 43.5 174 1.2528 -0.0011 1.2528 1.1193
No log 44.0 176 1.2471 0.0299 1.2471 1.1168
No log 44.5 178 1.2290 0.0485 1.2290 1.1086
No log 45.0 180 1.2287 0.0517 1.2287 1.1084
No log 45.5 182 1.2323 0.0822 1.2323 1.1101
No log 46.0 184 1.2277 0.0822 1.2277 1.1080
No log 46.5 186 1.2154 0.0944 1.2154 1.1025
No log 47.0 188 1.2141 0.0944 1.2141 1.1019
No log 47.5 190 1.2210 0.0033 1.2210 1.1050
No log 48.0 192 1.2199 0.0579 1.2199 1.1045
No log 48.5 194 1.2172 0.0944 1.2172 1.1033
No log 49.0 196 1.2122 0.0944 1.2122 1.1010
No log 49.5 198 1.2058 0.0517 1.2058 1.0981
No log 50.0 200 1.2029 0.0944 1.2029 1.0968
No log 50.5 202 1.1950 0.0517 1.1950 1.0931
No log 51.0 204 1.1903 0.0944 1.1903 1.0910
No log 51.5 206 1.1888 0.0517 1.1888 1.0903
No log 52.0 208 1.1976 0.0761 1.1976 1.0944
No log 52.5 210 1.2011 0.0608 1.2011 1.0960
No log 53.0 212 1.1956 0.0608 1.1956 1.0934
No log 53.5 214 1.1843 0.0761 1.1843 1.0882
No log 54.0 216 1.1753 0.0944 1.1753 1.0841
No log 54.5 218 1.1757 0.1521 1.1757 1.0843
No log 55.0 220 1.1751 0.1521 1.1751 1.0840
No log 55.5 222 1.1757 0.0944 1.1757 1.0843
No log 56.0 224 1.1868 0.0761 1.1868 1.0894
No log 56.5 226 1.2052 -0.0255 1.2052 1.0978
No log 57.0 228 1.2230 0.0358 1.2230 1.1059
No log 57.5 230 1.2275 0.0730 1.2275 1.1079
No log 58.0 232 1.2094 0.0761 1.2094 1.0997
No log 58.5 234 1.1952 0.0144 1.1952 1.0933
No log 59.0 236 1.1979 -0.0066 1.1979 1.0945
No log 59.5 238 1.2172 0.0242 1.2172 1.1033
No log 60.0 240 1.2271 0.0395 1.2271 1.1078
No log 60.5 242 1.2263 0.0395 1.2263 1.1074
No log 61.0 244 1.2197 0.0242 1.2197 1.1044
No log 61.5 246 1.2120 -0.0066 1.2120 1.1009
No log 62.0 248 1.2084 -0.0066 1.2084 1.0993
No log 62.5 250 1.2015 -0.0066 1.2015 1.0961
No log 63.0 252 1.1938 0.0331 1.1938 1.0926
No log 63.5 254 1.1843 0.0454 1.1843 1.0883
No log 64.0 256 1.1803 0.0331 1.1803 1.0864
No log 64.5 258 1.1775 0.0454 1.1775 1.0851
No log 65.0 260 1.1812 0.0454 1.1812 1.0868
No log 65.5 262 1.1942 0.0331 1.1942 1.0928
No log 66.0 264 1.2077 0.0331 1.2077 1.0989
No log 66.5 266 1.2182 0.0331 1.2182 1.1037
No log 67.0 268 1.2306 0.0299 1.2306 1.1093
No log 67.5 270 1.2408 0.0299 1.2408 1.1139
No log 68.0 272 1.2352 0.0299 1.2352 1.1114
No log 68.5 274 1.2177 0.0331 1.2177 1.1035
No log 69.0 276 1.2081 0.0363 1.2081 1.0991
No log 69.5 278 1.2039 0.0242 1.2039 1.0972
No log 70.0 280 1.2057 0.0670 1.2057 1.0980
No log 70.5 282 1.2054 0.0670 1.2054 1.0979
No log 71.0 284 1.2064 0.0822 1.2064 1.0984
No log 71.5 286 1.2039 0.0670 1.2039 1.0972
No log 72.0 288 1.1996 0.0363 1.1996 1.0953
No log 72.5 290 1.2011 0.0761 1.2011 1.0959
No log 73.0 292 1.2041 0.0761 1.2041 1.0973
No log 73.5 294 1.2055 0.0761 1.2055 1.0980
No log 74.0 296 1.2051 0.0761 1.2051 1.0978
No log 74.5 298 1.2044 0.0761 1.2044 1.0974
No log 75.0 300 1.2029 0.0761 1.2029 1.0968
No log 75.5 302 1.1981 0.0761 1.1981 1.0946
No log 76.0 304 1.1957 0.0761 1.1957 1.0935
No log 76.5 306 1.1967 0.0761 1.1967 1.0939
No log 77.0 308 1.1995 0.0944 1.1995 1.0952
No log 77.5 310 1.1992 0.0700 1.1992 1.0951
No log 78.0 312 1.1989 0.0700 1.1989 1.0949
No log 78.5 314 1.1996 0.0700 1.1996 1.0953
No log 79.0 316 1.2022 0.0306 1.2022 1.0964
No log 79.5 318 1.2107 0.0306 1.2107 1.1003
No log 80.0 320 1.2184 0.0731 1.2184 1.1038
No log 80.5 322 1.2252 0.0338 1.2252 1.1069
No log 81.0 324 1.2337 0.0489 1.2337 1.1107
No log 81.5 326 1.2426 0.0400 1.2426 1.1147
No log 82.0 328 1.2491 0.0400 1.2491 1.1176
No log 82.5 330 1.2597 0.0701 1.2597 1.1224
No log 83.0 332 1.2645 0.1119 1.2645 1.1245
No log 83.5 334 1.2674 0.1119 1.2674 1.1258
No log 84.0 336 1.2593 0.0701 1.2593 1.1222
No log 84.5 338 1.2456 0.0400 1.2456 1.1160
No log 85.0 340 1.2310 0.0338 1.2310 1.1095
No log 85.5 342 1.2181 0.0395 1.2181 1.1037
No log 86.0 344 1.2103 0.0761 1.2103 1.1001
No log 86.5 346 1.2080 0.0761 1.2080 1.0991
No log 87.0 348 1.2072 0.0761 1.2072 1.0987
No log 87.5 350 1.2078 0.0761 1.2078 1.0990
No log 88.0 352 1.2099 0.0761 1.2099 1.1000
No log 88.5 354 1.2134 0.0761 1.2134 1.1015
No log 89.0 356 1.2157 0.0761 1.2157 1.1026
No log 89.5 358 1.2169 0.0761 1.2169 1.1031
No log 90.0 360 1.2178 0.0761 1.2178 1.1035
No log 90.5 362 1.2189 0.0761 1.2189 1.1040
No log 91.0 364 1.2203 0.0242 1.2203 1.1047
No log 91.5 366 1.2217 0.0395 1.2217 1.1053
No log 92.0 368 1.2219 0.0395 1.2219 1.1054
No log 92.5 370 1.2224 0.0548 1.2224 1.1056
No log 93.0 372 1.2206 0.0395 1.2206 1.1048
No log 93.5 374 1.2188 0.0395 1.2188 1.1040
No log 94.0 376 1.2175 0.0639 1.2175 1.1034
No log 94.5 378 1.2179 0.0639 1.2179 1.1036
No log 95.0 380 1.2189 0.0242 1.2189 1.1041
No log 95.5 382 1.2201 0.0395 1.2201 1.1046
No log 96.0 384 1.2213 0.0395 1.2213 1.1051
No log 96.5 386 1.2227 0.0395 1.2227 1.1058
No log 97.0 388 1.2239 0.0395 1.2239 1.1063
No log 97.5 390 1.2245 0.0395 1.2245 1.1066
No log 98.0 392 1.2248 0.0395 1.2248 1.1067
No log 98.5 394 1.2253 0.0395 1.2253 1.1069
No log 99.0 396 1.2254 0.0548 1.2254 1.1070
No log 99.5 398 1.2254 0.0548 1.2254 1.1070
No log 100.0 400 1.2253 0.0548 1.2253 1.1069

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k1_task5_organization

Finetuned
(4019)
this model