ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k1_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4351
  • Qwk: 0.5970
  • Mse: 0.4351
  • Rmse: 0.6596

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.4 2 2.4735 -0.0788 2.4735 1.5727
No log 0.8 4 1.4778 0.1001 1.4778 1.2157
No log 1.2 6 0.7282 0.2430 0.7282 0.8533
No log 1.6 8 0.6639 0.2121 0.6639 0.8148
No log 2.0 10 0.7835 0.3051 0.7835 0.8851
No log 2.4 12 0.6741 0.4387 0.6741 0.8210
No log 2.8 14 0.5778 0.3907 0.5778 0.7602
No log 3.2 16 0.7846 0.4265 0.7846 0.8858
No log 3.6 18 1.0669 0.2894 1.0669 1.0329
No log 4.0 20 0.6874 0.5373 0.6874 0.8291
No log 4.4 22 0.4244 0.6683 0.4244 0.6515
No log 4.8 24 0.4244 0.6683 0.4244 0.6515
No log 5.2 26 0.5281 0.5862 0.5281 0.7267
No log 5.6 28 0.7427 0.4568 0.7427 0.8618
No log 6.0 30 0.5694 0.5739 0.5694 0.7546
No log 6.4 32 0.4939 0.6496 0.4939 0.7027
No log 6.8 34 0.5189 0.6152 0.5189 0.7203
No log 7.2 36 0.4846 0.7094 0.4846 0.6961
No log 7.6 38 0.5840 0.5887 0.5840 0.7642
No log 8.0 40 0.5220 0.6104 0.5220 0.7225
No log 8.4 42 0.4786 0.6009 0.4786 0.6918
No log 8.8 44 0.4485 0.6953 0.4485 0.6697
No log 9.2 46 0.7905 0.4338 0.7905 0.8891
No log 9.6 48 0.7824 0.4338 0.7824 0.8845
No log 10.0 50 0.4300 0.6395 0.4300 0.6558
No log 10.4 52 0.4288 0.6458 0.4288 0.6549
No log 10.8 54 0.4372 0.6118 0.4372 0.6612
No log 11.2 56 0.5298 0.6117 0.5298 0.7279
No log 11.6 58 0.4553 0.6235 0.4553 0.6748
No log 12.0 60 0.5204 0.5895 0.5204 0.7214
No log 12.4 62 0.6216 0.4874 0.6216 0.7884
No log 12.8 64 0.4519 0.6541 0.4519 0.6722
No log 13.2 66 0.4189 0.6481 0.4189 0.6472
No log 13.6 68 0.4676 0.6092 0.4676 0.6838
No log 14.0 70 0.4577 0.6271 0.4577 0.6766
No log 14.4 72 0.4104 0.6661 0.4104 0.6406
No log 14.8 74 0.5343 0.5510 0.5343 0.7309
No log 15.2 76 0.6467 0.3913 0.6467 0.8042
No log 15.6 78 0.5060 0.5907 0.5060 0.7114
No log 16.0 80 0.4300 0.6555 0.4300 0.6558
No log 16.4 82 0.5137 0.5015 0.5137 0.7167
No log 16.8 84 0.4532 0.5642 0.4532 0.6732
No log 17.2 86 0.4281 0.6914 0.4281 0.6543
No log 17.6 88 0.4531 0.6132 0.4531 0.6731
No log 18.0 90 0.4142 0.6860 0.4142 0.6436
No log 18.4 92 0.6580 0.5220 0.6580 0.8112
No log 18.8 94 0.8426 0.4474 0.8426 0.9179
No log 19.2 96 0.6451 0.5464 0.6451 0.8032
No log 19.6 98 0.4267 0.7022 0.4267 0.6532
No log 20.0 100 0.4450 0.6687 0.4450 0.6671
No log 20.4 102 0.4218 0.6828 0.4218 0.6495
No log 20.8 104 0.4704 0.6260 0.4704 0.6858
No log 21.2 106 0.5125 0.5664 0.5125 0.7159
No log 21.6 108 0.5410 0.5464 0.5410 0.7355
No log 22.0 110 0.4358 0.5751 0.4358 0.6602
No log 22.4 112 0.4274 0.6931 0.4274 0.6538
No log 22.8 114 0.4479 0.5852 0.4479 0.6692
No log 23.2 116 0.4367 0.6464 0.4367 0.6608
No log 23.6 118 0.4980 0.5765 0.4980 0.7057
No log 24.0 120 0.4686 0.6017 0.4686 0.6845
No log 24.4 122 0.4292 0.7041 0.4292 0.6552
No log 24.8 124 0.4304 0.6359 0.4304 0.6561
No log 25.2 126 0.4342 0.6359 0.4342 0.6589
No log 25.6 128 0.4154 0.7110 0.4154 0.6445
No log 26.0 130 0.5238 0.6072 0.5238 0.7237
No log 26.4 132 0.5487 0.5900 0.5487 0.7408
No log 26.8 134 0.4492 0.6886 0.4492 0.6702
No log 27.2 136 0.4106 0.72 0.4106 0.6408
No log 27.6 138 0.4091 0.7293 0.4091 0.6396
No log 28.0 140 0.4178 0.6772 0.4178 0.6464
No log 28.4 142 0.4528 0.6705 0.4528 0.6729
No log 28.8 144 0.4301 0.6690 0.4301 0.6558
No log 29.2 146 0.4373 0.6601 0.4373 0.6613
No log 29.6 148 0.4393 0.6601 0.4393 0.6628
No log 30.0 150 0.4256 0.6771 0.4256 0.6524
No log 30.4 152 0.4235 0.7489 0.4235 0.6508
No log 30.8 154 0.4290 0.6977 0.4290 0.6550
No log 31.2 156 0.4275 0.7489 0.4275 0.6538
No log 31.6 158 0.4774 0.5957 0.4774 0.6910
No log 32.0 160 0.6062 0.5220 0.6062 0.7786
No log 32.4 162 0.6117 0.5034 0.6117 0.7821
No log 32.8 164 0.5107 0.5650 0.5107 0.7146
No log 33.2 166 0.4278 0.6873 0.4278 0.6541
No log 33.6 168 0.4165 0.7032 0.4165 0.6454
No log 34.0 170 0.4179 0.7118 0.4179 0.6465
No log 34.4 172 0.4186 0.7042 0.4186 0.6470
No log 34.8 174 0.4701 0.5650 0.4701 0.6856
No log 35.2 176 0.4877 0.5581 0.4877 0.6984
No log 35.6 178 0.4890 0.5581 0.4890 0.6993
No log 36.0 180 0.4551 0.5836 0.4551 0.6746
No log 36.4 182 0.4243 0.7012 0.4243 0.6514
No log 36.8 184 0.4345 0.6627 0.4345 0.6592
No log 37.2 186 0.4319 0.6627 0.4319 0.6572
No log 37.6 188 0.4257 0.7032 0.4257 0.6525
No log 38.0 190 0.4371 0.6577 0.4371 0.6611
No log 38.4 192 0.4338 0.6566 0.4338 0.6586
No log 38.8 194 0.4368 0.6214 0.4368 0.6609
No log 39.2 196 0.4471 0.5995 0.4471 0.6686
No log 39.6 198 0.4368 0.6739 0.4368 0.6609
No log 40.0 200 0.4782 0.5621 0.4782 0.6915
No log 40.4 202 0.5267 0.5778 0.5267 0.7258
No log 40.8 204 0.4966 0.5778 0.4966 0.7047
No log 41.2 206 0.4384 0.6322 0.4384 0.6621
No log 41.6 208 0.4397 0.6787 0.4397 0.6631
No log 42.0 210 0.4341 0.6953 0.4341 0.6589
No log 42.4 212 0.4311 0.6503 0.4311 0.6566
No log 42.8 214 0.4446 0.5200 0.4446 0.6668
No log 43.2 216 0.4380 0.6143 0.4380 0.6618
No log 43.6 218 0.4170 0.6950 0.4170 0.6457
No log 44.0 220 0.4173 0.6771 0.4173 0.6460
No log 44.4 222 0.4201 0.6407 0.4201 0.6482
No log 44.8 224 0.4378 0.6627 0.4378 0.6617
No log 45.2 226 0.4457 0.6803 0.4457 0.6676
No log 45.6 228 0.4296 0.6143 0.4296 0.6554
No log 46.0 230 0.4143 0.6852 0.4143 0.6437
No log 46.4 232 0.4150 0.6747 0.4150 0.6442
No log 46.8 234 0.4200 0.6636 0.4200 0.6481
No log 47.2 236 0.4252 0.6517 0.4252 0.6521
No log 47.6 238 0.4302 0.5819 0.4302 0.6559
No log 48.0 240 0.4282 0.5836 0.4282 0.6544
No log 48.4 242 0.4324 0.6229 0.4324 0.6575
No log 48.8 244 0.4582 0.5587 0.4582 0.6769
No log 49.2 246 0.4596 0.5373 0.4596 0.6779
No log 49.6 248 0.4268 0.6678 0.4268 0.6533
No log 50.0 250 0.4144 0.6866 0.4144 0.6437
No log 50.4 252 0.4204 0.6866 0.4204 0.6484
No log 50.8 254 0.4262 0.6661 0.4262 0.6529
No log 51.2 256 0.4215 0.6448 0.4215 0.6493
No log 51.6 258 0.4240 0.6770 0.4240 0.6511
No log 52.0 260 0.4245 0.6566 0.4245 0.6516
No log 52.4 262 0.4279 0.6904 0.4279 0.6541
No log 52.8 264 0.4279 0.6904 0.4279 0.6541
No log 53.2 266 0.4274 0.6483 0.4274 0.6537
No log 53.6 268 0.4405 0.5335 0.4405 0.6637
No log 54.0 270 0.4397 0.5352 0.4397 0.6631
No log 54.4 272 0.4282 0.5611 0.4282 0.6544
No log 54.8 274 0.4228 0.6073 0.4228 0.6502
No log 55.2 276 0.4274 0.6648 0.4274 0.6538
No log 55.6 278 0.4279 0.6648 0.4279 0.6541
No log 56.0 280 0.4197 0.6759 0.4197 0.6479
No log 56.4 282 0.4230 0.5611 0.4230 0.6504
No log 56.8 284 0.4270 0.5611 0.4270 0.6535
No log 57.2 286 0.4264 0.5611 0.4264 0.6530
No log 57.6 288 0.4340 0.5826 0.4340 0.6588
No log 58.0 290 0.4385 0.5918 0.4385 0.6622
No log 58.4 292 0.4317 0.5611 0.4317 0.6571
No log 58.8 294 0.4383 0.5918 0.4383 0.6620
No log 59.2 296 0.4298 0.5702 0.4298 0.6556
No log 59.6 298 0.4299 0.5702 0.4299 0.6557
No log 60.0 300 0.4207 0.6667 0.4207 0.6486
No log 60.4 302 0.4152 0.6854 0.4152 0.6444
No log 60.8 304 0.4165 0.6854 0.4165 0.6454
No log 61.2 306 0.4190 0.6854 0.4190 0.6473
No log 61.6 308 0.4396 0.5495 0.4396 0.6630
No log 62.0 310 0.4665 0.5384 0.4665 0.6830
No log 62.4 312 0.4757 0.5528 0.4757 0.6897
No log 62.8 314 0.4645 0.5384 0.4645 0.6815
No log 63.2 316 0.4482 0.5438 0.4482 0.6695
No log 63.6 318 0.4329 0.5495 0.4329 0.6580
No log 64.0 320 0.4319 0.5571 0.4319 0.6572
No log 64.4 322 0.4351 0.5649 0.4351 0.6596
No log 64.8 324 0.4357 0.5861 0.4357 0.6601
No log 65.2 326 0.4353 0.6661 0.4353 0.6598
No log 65.6 328 0.4384 0.6725 0.4384 0.6621
No log 66.0 330 0.4383 0.6733 0.4383 0.6621
No log 66.4 332 0.4381 0.6101 0.4381 0.6619
No log 66.8 334 0.4379 0.6154 0.4379 0.6618
No log 67.2 336 0.4426 0.6154 0.4426 0.6653
No log 67.6 338 0.4422 0.6154 0.4422 0.6649
No log 68.0 340 0.4413 0.6154 0.4413 0.6643
No log 68.4 342 0.4394 0.6068 0.4394 0.6629
No log 68.8 344 0.4349 0.6101 0.4349 0.6594
No log 69.2 346 0.4348 0.6101 0.4348 0.6594
No log 69.6 348 0.4402 0.5826 0.4402 0.6635
No log 70.0 350 0.4634 0.5721 0.4634 0.6807
No log 70.4 352 0.4990 0.5408 0.4990 0.7064
No log 70.8 354 0.5073 0.5408 0.5073 0.7123
No log 71.2 356 0.4854 0.5408 0.4854 0.6967
No log 71.6 358 0.4558 0.5908 0.4558 0.6751
No log 72.0 360 0.4382 0.6156 0.4382 0.6620
No log 72.4 362 0.4320 0.5826 0.4320 0.6572
No log 72.8 364 0.4309 0.5611 0.4309 0.6564
No log 73.2 366 0.4297 0.6292 0.4297 0.6555
No log 73.6 368 0.4281 0.6292 0.4281 0.6543
No log 74.0 370 0.4290 0.6464 0.4290 0.6550
No log 74.4 372 0.4304 0.5890 0.4304 0.6561
No log 74.8 374 0.4331 0.5611 0.4331 0.6581
No log 75.2 376 0.4350 0.5611 0.4350 0.6595
No log 75.6 378 0.4366 0.5611 0.4366 0.6608
No log 76.0 380 0.4352 0.5611 0.4352 0.6597
No log 76.4 382 0.4345 0.5649 0.4345 0.6591
No log 76.8 384 0.4365 0.5611 0.4365 0.6606
No log 77.2 386 0.4380 0.5826 0.4380 0.6618
No log 77.6 388 0.4351 0.5611 0.4351 0.6596
No log 78.0 390 0.4325 0.5765 0.4325 0.6576
No log 78.4 392 0.4318 0.5970 0.4318 0.6571
No log 78.8 394 0.4317 0.6464 0.4317 0.6570
No log 79.2 396 0.4316 0.6464 0.4316 0.6570
No log 79.6 398 0.4312 0.5970 0.4312 0.6566
No log 80.0 400 0.4320 0.5812 0.4320 0.6573
No log 80.4 402 0.4334 0.5812 0.4334 0.6583
No log 80.8 404 0.4341 0.6020 0.4341 0.6588
No log 81.2 406 0.4341 0.6020 0.4341 0.6589
No log 81.6 408 0.4370 0.6223 0.4370 0.6611
No log 82.0 410 0.4435 0.6034 0.4435 0.6660
No log 82.4 412 0.4515 0.6235 0.4515 0.6719
No log 82.8 414 0.4613 0.5779 0.4613 0.6792
No log 83.2 416 0.4598 0.5779 0.4598 0.6781
No log 83.6 418 0.4554 0.5554 0.4554 0.6749
No log 84.0 420 0.4493 0.5554 0.4493 0.6703
No log 84.4 422 0.4459 0.6034 0.4459 0.6678
No log 84.8 424 0.4469 0.5554 0.4469 0.6685
No log 85.2 426 0.4447 0.5554 0.4447 0.6669
No log 85.6 428 0.4412 0.5611 0.4412 0.6642
No log 86.0 430 0.4396 0.5890 0.4396 0.6630
No log 86.4 432 0.4390 0.5890 0.4390 0.6625
No log 86.8 434 0.4389 0.5890 0.4389 0.6625
No log 87.2 436 0.4388 0.5890 0.4388 0.6624
No log 87.6 438 0.4392 0.5890 0.4392 0.6628
No log 88.0 440 0.4400 0.5687 0.4400 0.6633
No log 88.4 442 0.4402 0.5178 0.4402 0.6635
No log 88.8 444 0.4409 0.5266 0.4409 0.6640
No log 89.2 446 0.4417 0.5266 0.4417 0.6646
No log 89.6 448 0.4432 0.5266 0.4432 0.6657
No log 90.0 450 0.4433 0.5266 0.4433 0.6658
No log 90.4 452 0.4412 0.5266 0.4412 0.6642
No log 90.8 454 0.4394 0.5266 0.4394 0.6629
No log 91.2 456 0.4383 0.5266 0.4383 0.6620
No log 91.6 458 0.4372 0.5687 0.4372 0.6612
No log 92.0 460 0.4362 0.5687 0.4362 0.6605
No log 92.4 462 0.4360 0.5687 0.4360 0.6603
No log 92.8 464 0.4362 0.5687 0.4362 0.6605
No log 93.2 466 0.4364 0.5687 0.4364 0.6606
No log 93.6 468 0.4368 0.5890 0.4368 0.6609
No log 94.0 470 0.4371 0.5890 0.4371 0.6612
No log 94.4 472 0.4378 0.5890 0.4378 0.6617
No log 94.8 474 0.4381 0.5890 0.4381 0.6619
No log 95.2 476 0.4383 0.5890 0.4383 0.6621
No log 95.6 478 0.4380 0.5890 0.4380 0.6618
No log 96.0 480 0.4373 0.5970 0.4373 0.6613
No log 96.4 482 0.4365 0.5970 0.4365 0.6607
No log 96.8 484 0.4360 0.5970 0.4360 0.6603
No log 97.2 486 0.4357 0.5970 0.4357 0.6601
No log 97.6 488 0.4354 0.5970 0.4354 0.6599
No log 98.0 490 0.4353 0.5970 0.4353 0.6598
No log 98.4 492 0.4352 0.5970 0.4352 0.6597
No log 98.8 494 0.4351 0.5970 0.4351 0.6596
No log 99.2 496 0.4351 0.5970 0.4351 0.6596
No log 99.6 498 0.4351 0.5970 0.4351 0.6596
0.1804 100.0 500 0.4351 0.5970 0.4351 0.6596

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k1_task7_organization

Finetuned
(4019)
this model