ArabicNewSplits5_FineTuningAraBERT_run3_AugV5_k6_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7041
  • Qwk: 0.5140
  • Mse: 0.7041
  • Rmse: 0.8391

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.05 2 4.2345 -0.0066 4.2345 2.0578
No log 0.1 4 2.5189 0.0111 2.5189 1.5871
No log 0.15 6 1.6177 -0.0076 1.6177 1.2719
No log 0.2 8 1.3707 0.0353 1.3707 1.1708
No log 0.25 10 1.0898 0.0787 1.0898 1.0439
No log 0.3 12 0.9221 0.0386 0.9221 0.9603
No log 0.35 14 0.8610 0.0745 0.8610 0.9279
No log 0.4 16 0.8883 0.0470 0.8883 0.9425
No log 0.45 18 1.0904 0.0969 1.0904 1.0442
No log 0.5 20 1.3596 0.1194 1.3596 1.1660
No log 0.55 22 1.4447 0.1047 1.4447 1.2020
No log 0.6 24 1.5277 0.0644 1.5277 1.2360
No log 0.65 26 1.3045 0.0572 1.3045 1.1422
No log 0.7 28 0.9030 0.1094 0.9030 0.9502
No log 0.75 30 0.7592 0.1736 0.7592 0.8713
No log 0.8 32 0.6924 0.2742 0.6924 0.8321
No log 0.85 34 0.6722 0.3162 0.6722 0.8199
No log 0.9 36 0.6666 0.3420 0.6666 0.8165
No log 0.95 38 0.6837 0.2565 0.6837 0.8268
No log 1.0 40 0.7820 0.2947 0.7820 0.8843
No log 1.05 42 0.9433 0.2560 0.9433 0.9712
No log 1.1 44 0.9602 0.2774 0.9602 0.9799
No log 1.15 46 0.9694 0.2799 0.9694 0.9846
No log 1.2 48 1.0663 0.2469 1.0663 1.0326
No log 1.25 50 1.0964 0.2282 1.0964 1.0471
No log 1.3 52 0.9938 0.2660 0.9938 0.9969
No log 1.35 54 0.8477 0.1881 0.8477 0.9207
No log 1.4 56 0.7796 0.1488 0.7796 0.8830
No log 1.45 58 0.6956 0.2961 0.6956 0.8340
No log 1.5 60 0.7317 0.4141 0.7317 0.8554
No log 1.55 62 1.0914 0.3114 1.0914 1.0447
No log 1.6 64 1.1584 0.3099 1.1584 1.0763
No log 1.65 66 1.0617 0.2851 1.0617 1.0304
No log 1.7 68 0.9695 0.2888 0.9695 0.9846
No log 1.75 70 0.7529 0.3814 0.7529 0.8677
No log 1.8 72 0.6535 0.3595 0.6535 0.8084
No log 1.85 74 0.6282 0.4313 0.6282 0.7926
No log 1.9 76 0.6050 0.4561 0.6050 0.7778
No log 1.95 78 0.6171 0.4340 0.6171 0.7856
No log 2.0 80 0.6705 0.3796 0.6705 0.8188
No log 2.05 82 0.7847 0.3405 0.7847 0.8858
No log 2.1 84 0.8992 0.3295 0.8992 0.9483
No log 2.15 86 0.8629 0.3375 0.8629 0.9289
No log 2.2 88 0.8449 0.3495 0.8449 0.9192
No log 2.25 90 0.8443 0.3502 0.8443 0.9188
No log 2.3 92 0.8470 0.3665 0.8470 0.9203
No log 2.35 94 0.8616 0.3970 0.8616 0.9282
No log 2.4 96 0.7459 0.3883 0.7459 0.8637
No log 2.45 98 0.6125 0.4661 0.6125 0.7826
No log 2.5 100 0.6291 0.4532 0.6291 0.7931
No log 2.55 102 0.7107 0.3735 0.7107 0.8430
No log 2.6 104 0.7318 0.2592 0.7318 0.8554
No log 2.65 106 0.7158 0.3686 0.7158 0.8460
No log 2.7 108 0.7374 0.4176 0.7374 0.8587
No log 2.75 110 0.7211 0.4420 0.7211 0.8492
No log 2.8 112 0.6437 0.4404 0.6437 0.8023
No log 2.85 114 0.6052 0.4507 0.6052 0.7780
No log 2.9 116 0.6063 0.4019 0.6063 0.7786
No log 2.95 118 0.6032 0.4382 0.6032 0.7767
No log 3.0 120 0.6639 0.4039 0.6639 0.8148
No log 3.05 122 0.7689 0.4192 0.7689 0.8768
No log 3.1 124 0.8349 0.4214 0.8349 0.9137
No log 3.15 126 0.8471 0.3929 0.8471 0.9204
No log 3.2 128 0.7725 0.4378 0.7725 0.8789
No log 3.25 130 0.6931 0.4831 0.6931 0.8325
No log 3.3 132 0.6742 0.4972 0.6742 0.8211
No log 3.35 134 0.7779 0.4903 0.7779 0.8820
No log 3.4 136 0.8525 0.4511 0.8525 0.9233
No log 3.45 138 0.7620 0.5572 0.7620 0.8729
No log 3.5 140 0.7526 0.5214 0.7526 0.8675
No log 3.55 142 0.8435 0.5182 0.8435 0.9184
No log 3.6 144 0.8466 0.4781 0.8466 0.9201
No log 3.65 146 0.7391 0.5526 0.7391 0.8597
No log 3.7 148 0.6688 0.5250 0.6688 0.8178
No log 3.75 150 0.6461 0.5063 0.6461 0.8038
No log 3.8 152 0.6446 0.5265 0.6446 0.8029
No log 3.85 154 0.6671 0.5144 0.6671 0.8168
No log 3.9 156 0.7272 0.5118 0.7272 0.8528
No log 3.95 158 0.8237 0.4533 0.8237 0.9076
No log 4.0 160 0.8399 0.4536 0.8399 0.9164
No log 4.05 162 0.7687 0.4693 0.7687 0.8767
No log 4.1 164 0.7068 0.4597 0.7068 0.8407
No log 4.15 166 0.6993 0.4833 0.6993 0.8362
No log 4.2 168 0.7247 0.4615 0.7247 0.8513
No log 4.25 170 0.7348 0.4664 0.7348 0.8572
No log 4.3 172 0.7385 0.4517 0.7385 0.8594
No log 4.35 174 0.7651 0.4631 0.7651 0.8747
No log 4.4 176 0.8042 0.4552 0.8042 0.8968
No log 4.45 178 0.8349 0.4530 0.8349 0.9137
No log 4.5 180 0.8074 0.4471 0.8074 0.8985
No log 4.55 182 0.7482 0.3690 0.7482 0.8650
No log 4.6 184 0.7305 0.4068 0.7305 0.8547
No log 4.65 186 0.7180 0.4209 0.7180 0.8474
No log 4.7 188 0.7000 0.4145 0.7000 0.8367
No log 4.75 190 0.7194 0.4197 0.7194 0.8481
No log 4.8 192 0.7753 0.4683 0.7753 0.8805
No log 4.85 194 0.7578 0.4596 0.7578 0.8705
No log 4.9 196 0.7361 0.4731 0.7361 0.8579
No log 4.95 198 0.7114 0.4445 0.7114 0.8434
No log 5.0 200 0.7257 0.4451 0.7257 0.8519
No log 5.05 202 0.7285 0.4659 0.7285 0.8535
No log 5.1 204 0.7232 0.5105 0.7232 0.8504
No log 5.15 206 0.7113 0.5183 0.7113 0.8434
No log 5.2 208 0.6961 0.5133 0.6961 0.8343
No log 5.25 210 0.6928 0.5115 0.6928 0.8323
No log 5.3 212 0.7032 0.5486 0.7032 0.8386
No log 5.35 214 0.7344 0.5217 0.7344 0.8569
No log 5.4 216 0.7442 0.5455 0.7442 0.8627
No log 5.45 218 0.7452 0.5178 0.7452 0.8632
No log 5.5 220 0.7613 0.5119 0.7613 0.8725
No log 5.55 222 0.7739 0.5203 0.7739 0.8797
No log 5.6 224 0.7671 0.5178 0.7671 0.8758
No log 5.65 226 0.7580 0.5078 0.7580 0.8706
No log 5.7 228 0.7695 0.4898 0.7695 0.8772
No log 5.75 230 0.7725 0.4898 0.7725 0.8789
No log 5.8 232 0.7411 0.5286 0.7411 0.8609
No log 5.85 234 0.7303 0.5461 0.7303 0.8546
No log 5.9 236 0.7160 0.5461 0.7160 0.8462
No log 5.95 238 0.7138 0.5403 0.7138 0.8449
No log 6.0 240 0.7056 0.5446 0.7056 0.8400
No log 6.05 242 0.7033 0.5299 0.7033 0.8387
No log 6.1 244 0.7182 0.5463 0.7182 0.8475
No log 6.15 246 0.7382 0.5332 0.7382 0.8592
No log 6.2 248 0.7584 0.5364 0.7584 0.8708
No log 6.25 250 0.7781 0.5269 0.7781 0.8821
No log 6.3 252 0.7883 0.5188 0.7883 0.8878
No log 6.35 254 0.8093 0.5216 0.8093 0.8996
No log 6.4 256 0.8443 0.4855 0.8443 0.9189
No log 6.45 258 0.8209 0.5115 0.8209 0.9060
No log 6.5 260 0.7677 0.5232 0.7677 0.8762
No log 6.55 262 0.7504 0.5187 0.7504 0.8663
No log 6.6 264 0.7442 0.5292 0.7442 0.8626
No log 6.65 266 0.7556 0.5296 0.7556 0.8693
No log 6.7 268 0.7658 0.5284 0.7658 0.8751
No log 6.75 270 0.7740 0.5253 0.7740 0.8798
No log 6.8 272 0.7852 0.5227 0.7852 0.8861
No log 6.85 274 0.7855 0.5282 0.7855 0.8863
No log 6.9 276 0.7607 0.5293 0.7607 0.8722
No log 6.95 278 0.7320 0.5292 0.7320 0.8556
No log 7.0 280 0.7211 0.5303 0.7211 0.8492
No log 7.05 282 0.7247 0.5136 0.7247 0.8513
No log 7.1 284 0.7161 0.5316 0.7161 0.8462
No log 7.15 286 0.6996 0.5403 0.6996 0.8364
No log 7.2 288 0.6907 0.5095 0.6907 0.8311
No log 7.25 290 0.6813 0.5477 0.6813 0.8254
No log 7.3 292 0.6794 0.5397 0.6794 0.8243
No log 7.35 294 0.6849 0.5320 0.6849 0.8276
No log 7.4 296 0.6995 0.5148 0.6995 0.8364
No log 7.45 298 0.7298 0.5254 0.7298 0.8543
No log 7.5 300 0.7664 0.5068 0.7664 0.8754
No log 7.55 302 0.7857 0.5089 0.7857 0.8864
No log 7.6 304 0.7640 0.4947 0.7640 0.8741
No log 7.65 306 0.7244 0.5192 0.7244 0.8511
No log 7.7 308 0.7082 0.5184 0.7082 0.8416
No log 7.75 310 0.7146 0.5170 0.7146 0.8453
No log 7.8 312 0.7194 0.5306 0.7194 0.8482
No log 7.85 314 0.7189 0.5258 0.7189 0.8479
No log 7.9 316 0.7210 0.5184 0.7210 0.8491
No log 7.95 318 0.7242 0.5120 0.7242 0.8510
No log 8.0 320 0.7299 0.5087 0.7299 0.8543
No log 8.05 322 0.7251 0.5089 0.7251 0.8515
No log 8.1 324 0.7302 0.5235 0.7302 0.8545
No log 8.15 326 0.7249 0.5230 0.7249 0.8514
No log 8.2 328 0.7124 0.5127 0.7124 0.8441
No log 8.25 330 0.7097 0.5137 0.7097 0.8424
No log 8.3 332 0.7049 0.5049 0.7049 0.8396
No log 8.35 334 0.6979 0.5044 0.6979 0.8354
No log 8.4 336 0.6961 0.5119 0.6961 0.8343
No log 8.45 338 0.6967 0.5281 0.6967 0.8347
No log 8.5 340 0.7015 0.5434 0.7015 0.8376
No log 8.55 342 0.7071 0.5348 0.7071 0.8409
No log 8.6 344 0.7082 0.5348 0.7082 0.8416
No log 8.65 346 0.7065 0.5274 0.7065 0.8406
No log 8.7 348 0.7044 0.5052 0.7044 0.8393
No log 8.75 350 0.7113 0.5092 0.7113 0.8434
No log 8.8 352 0.7239 0.5073 0.7239 0.8508
No log 8.85 354 0.7339 0.5031 0.7339 0.8567
No log 8.9 356 0.7312 0.5031 0.7312 0.8551
No log 8.95 358 0.7181 0.5176 0.7181 0.8474
No log 9.0 360 0.7075 0.5092 0.7075 0.8412
No log 9.05 362 0.7017 0.5092 0.7017 0.8376
No log 9.1 364 0.6954 0.5167 0.6954 0.8339
No log 9.15 366 0.6911 0.5167 0.6911 0.8313
No log 9.2 368 0.6894 0.5194 0.6894 0.8303
No log 9.25 370 0.6906 0.5194 0.6906 0.8310
No log 9.3 372 0.6934 0.5109 0.6934 0.8327
No log 9.35 374 0.6958 0.5109 0.6958 0.8341
No log 9.4 376 0.6967 0.5109 0.6967 0.8347
No log 9.45 378 0.6975 0.5109 0.6975 0.8352
No log 9.5 380 0.6992 0.5167 0.6992 0.8362
No log 9.55 382 0.7033 0.5130 0.7033 0.8386
No log 9.6 384 0.7080 0.5123 0.7080 0.8414
No log 9.65 386 0.7106 0.5123 0.7106 0.8430
No log 9.7 388 0.7120 0.4983 0.7120 0.8438
No log 9.75 390 0.7114 0.4983 0.7114 0.8435
No log 9.8 392 0.7095 0.5123 0.7095 0.8423
No log 9.85 394 0.7070 0.5123 0.7070 0.8408
No log 9.9 396 0.7050 0.5130 0.7050 0.8397
No log 9.95 398 0.7044 0.5130 0.7044 0.8393
No log 10.0 400 0.7041 0.5140 0.7041 0.8391

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits5_FineTuningAraBERT_run3_AugV5_k6_task2_organization

Finetuned
(4023)
this model