udgrf609_20250704_025246
This model is a fine-tuned version of meta-llama/Llama-3.2-1B on an unknown dataset. It achieves the following results on the evaluation set:
- Loss: 1.3280
- Model Preparation Time: 0.0077
- Move Accuracy: 0.0219
- Token Accuracy: 0.4912
- Accuracy: 0.0219
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.001
- train_batch_size: 128
- eval_batch_size: 256
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: constant_with_warmup
- lr_scheduler_warmup_ratio: 0.001
- num_epochs: 100
Training results
| Training Loss | Epoch | Step | Validation Loss | Model Preparation Time | Move Accuracy | Token Accuracy | Accuracy |
|---|---|---|---|---|---|---|---|
| No log | 0 | 0 | 11.9474 | 0.0077 | 0.0 | 0.0000 | 0.0 |
| 1.7249 | 0.0098 | 100 | 1.7275 | 0.0077 | 0.0001 | 0.3398 | 0.0001 |
| 1.5741 | 0.0196 | 200 | 1.5610 | 0.0077 | 0.0070 | 0.4066 | 0.0070 |
| 1.4021 | 0.0295 | 300 | 1.4419 | 0.0077 | 0.0093 | 0.4513 | 0.0093 |
| 1.3719 | 0.0393 | 400 | 1.4199 | 0.0077 | 0.0070 | 0.4577 | 0.0070 |
| 1.3511 | 0.0491 | 500 | 1.3810 | 0.0077 | 0.0184 | 0.4710 | 0.0184 |
| 1.2881 | 0.0589 | 600 | 1.3280 | 0.0077 | 0.0219 | 0.4912 | 0.0219 |
| 1.308 | 0.0687 | 700 | 1.3274 | 0.0077 | 0.0208 | 0.4895 | 0.0208 |
| 1.3919 | 0.0785 | 800 | 1.3478 | 0.0077 | 0.0189 | 0.4783 | 0.0189 |
| 1.4178 | 0.0884 | 900 | 1.4284 | 0.0077 | 0.0121 | 0.4576 | 0.0121 |
| 1.4408 | 0.0982 | 1000 | 1.4418 | 0.0077 | 0.0093 | 0.4347 | 0.0093 |
| 1.5107 | 0.1080 | 1100 | 1.5621 | 0.0077 | 0.0023 | 0.3942 | 0.0023 |
| 1.5004 | 0.1178 | 1200 | 1.4907 | 0.0077 | 0.0004 | 0.4004 | 0.0004 |
| 1.5062 | 0.1276 | 1300 | 1.5485 | 0.0077 | 0.0023 | 0.3994 | 0.0023 |
| 1.5707 | 0.1374 | 1400 | 1.5250 | 0.0077 | 0.0021 | 0.3953 | 0.0021 |
| 1.6335 | 0.1473 | 1500 | 1.6275 | 0.0077 | 0.0012 | 0.3768 | 0.0012 |
| 2.4746 | 0.1571 | 1600 | 2.5088 | 0.0077 | 0.0008 | 0.2987 | 0.0008 |
| 2.4986 | 0.1669 | 1700 | 4.1257 | 0.0077 | 0.0 | 0.1539 | 0.0 |
| 2.2833 | 0.1767 | 1800 | 2.4301 | 0.0077 | 0.0 | 0.1124 | 0.0 |
| 2.5158 | 0.1865 | 1900 | 2.5386 | 0.0077 | 0.0 | 0.2412 | 0.0 |
| 2.114 | 0.1963 | 2000 | 2.0981 | 0.0077 | 0.0 | 0.2817 | 0.0 |
| 2.0344 | 0.2062 | 2100 | 2.0449 | 0.0077 | 0.0 | 0.2517 | 0.0 |
| 2.0666 | 0.2160 | 2200 | 2.0564 | 0.0077 | 0.0 | 0.2705 | 0.0 |
| 2.5462 | 0.2258 | 2300 | 2.6100 | 0.0077 | 0.0 | 0.2328 | 0.0 |
| 1.9724 | 0.2356 | 2400 | 2.0092 | 0.0077 | 0.0 | 0.2819 | 0.0 |
| 1.9781 | 0.2454 | 2500 | 1.9666 | 0.0077 | 0.0 | 0.2799 | 0.0 |
| 2.0801 | 0.2553 | 2600 | 2.0669 | 0.0077 | 0.0 | 0.2506 | 0.0 |
| 2.1864 | 0.2651 | 2700 | 2.1805 | 0.0077 | 0.0019 | 0.2184 | 0.0019 |
| 2.0327 | 0.2749 | 2800 | 2.0283 | 0.0077 | 0.0 | 0.2504 | 0.0 |
| 1.7155 | 0.2847 | 2900 | 1.7586 | 0.0077 | 0.0017 | 0.3205 | 0.0017 |
| 3.4484 | 0.2945 | 3000 | 3.2784 | 0.0077 | 0.0 | 0.1990 | 0.0 |
| 3.7537 | 0.3043 | 3100 | 3.3351 | 0.0077 | 0.0 | 0.1639 | 0.0 |
| 1.7822 | 0.3142 | 3200 | 1.7770 | 0.0077 | 0.0 | 0.2973 | 0.0 |
| 1.965 | 0.3240 | 3300 | 1.9679 | 0.0077 | 0.0 | 0.2606 | 0.0 |
| 1.9091 | 0.3338 | 3400 | 1.9359 | 0.0077 | 0.0 | 0.2708 | 0.0 |
| 2.0503 | 0.3436 | 3500 | 2.0767 | 0.0077 | 0.0 | 0.2203 | 0.0 |
| 2.0584 | 0.3534 | 3600 | 1.9688 | 0.0077 | 0.0 | 0.2537 | 0.0 |
| 1.7356 | 0.3632 | 3700 | 1.7311 | 0.0077 | 0.0003 | 0.3087 | 0.0003 |
| 1.7098 | 0.3731 | 3800 | 1.7020 | 0.0077 | 0.0 | 0.3182 | 0.0 |
| 1.6855 | 0.3829 | 3900 | 1.6984 | 0.0077 | 0.0 | 0.3244 | 0.0 |
| 1.6753 | 0.3927 | 4000 | 1.6868 | 0.0077 | 0.0008 | 0.3204 | 0.0008 |
| 1.7065 | 0.4025 | 4100 | 1.6848 | 0.0077 | 0.0 | 0.3147 | 0.0 |
| 1.6817 | 0.4123 | 4200 | 1.6763 | 0.0077 | 0.0002 | 0.3240 | 0.0002 |
| 1.6763 | 0.4221 | 4300 | 1.6726 | 0.0077 | 0.0 | 0.3204 | 0.0 |
| 1.6801 | 0.4320 | 4400 | 1.6730 | 0.0077 | 0.0008 | 0.3264 | 0.0008 |
| 1.6679 | 0.4418 | 4500 | 1.6748 | 0.0077 | 0.0 | 0.3218 | 0.0 |
| 1.6549 | 0.4516 | 4600 | 1.6692 | 0.0077 | 0.0 | 0.3258 | 0.0 |
| 1.6951 | 0.4614 | 4700 | 1.6681 | 0.0077 | 0.0 | 0.3193 | 0.0 |
| 1.6724 | 0.4712 | 4800 | 1.6708 | 0.0077 | 0.0 | 0.3253 | 0.0 |
| 1.658 | 0.4811 | 4900 | 1.6711 | 0.0077 | 0.0002 | 0.3161 | 0.0002 |
| 1.6677 | 0.4909 | 5000 | 1.6713 | 0.0077 | 0.0007 | 0.3272 | 0.0007 |
| 1.6797 | 0.5007 | 5100 | 1.6752 | 0.0077 | 0.0001 | 0.3190 | 0.0001 |
| 1.6786 | 0.5105 | 5200 | 1.6754 | 0.0077 | 0.0 | 0.3153 | 0.0 |
| 1.661 | 0.5203 | 5300 | 1.6681 | 0.0077 | 0.0039 | 0.3196 | 0.0039 |
| 1.6825 | 0.5301 | 5400 | 1.6894 | 0.0077 | 0.0014 | 0.3132 | 0.0014 |
| 1.658 | 0.5400 | 5500 | 1.6743 | 0.0077 | 0.0003 | 0.3078 | 0.0003 |
| 1.677 | 0.5498 | 5600 | 1.6720 | 0.0077 | 0.0001 | 0.3159 | 0.0001 |
| 1.6896 | 0.5596 | 5700 | 1.6728 | 0.0077 | 0.0 | 0.3091 | 0.0 |
| 1.6795 | 0.5694 | 5800 | 1.6704 | 0.0077 | 0.0012 | 0.3237 | 0.0012 |
| 1.6646 | 0.5792 | 5900 | 1.6657 | 0.0077 | 0.0001 | 0.3249 | 0.0001 |
| 1.6619 | 0.5890 | 6000 | 1.6738 | 0.0077 | 0.0 | 0.3193 | 0.0 |
| 1.6475 | 0.5989 | 6100 | 1.6734 | 0.0077 | 0.0 | 0.3198 | 0.0 |
| 1.7044 | 0.6087 | 6200 | 1.6690 | 0.0077 | 0.0 | 0.3178 | 0.0 |
| 1.6921 | 0.6185 | 6300 | 1.6898 | 0.0077 | 0.0 | 0.3245 | 0.0 |
| 1.6757 | 0.6283 | 6400 | 1.6668 | 0.0077 | 0.0001 | 0.3267 | 0.0001 |
| 1.8259 | 0.6381 | 6500 | 1.8418 | 0.0077 | 0.0001 | 0.3122 | 0.0001 |
| 1.9607 | 0.6479 | 6600 | 1.9611 | 0.0077 | 0.0 | 0.3098 | 0.0 |
| 1.6544 | 0.6578 | 6700 | 1.6741 | 0.0077 | 0.0006 | 0.3154 | 0.0006 |
| 1.6718 | 0.6676 | 6800 | 1.6760 | 0.0077 | 0.0001 | 0.3200 | 0.0001 |
| 2.1974 | 0.6774 | 6900 | 2.1437 | 0.0077 | 0.0002 | 0.3086 | 0.0002 |
| 1.9982 | 0.6872 | 7000 | 2.0012 | 0.0077 | 0.0001 | 0.2950 | 0.0001 |
| 1.7201 | 0.6970 | 7100 | 1.7016 | 0.0077 | 0.0 | 0.3156 | 0.0 |
| 1.7066 | 0.7069 | 7200 | 1.6970 | 0.0077 | 0.0 | 0.3155 | 0.0 |
| 1.6948 | 0.7167 | 7300 | 1.6821 | 0.0077 | 0.0 | 0.3244 | 0.0 |
| 1.6561 | 0.7265 | 7400 | 1.6726 | 0.0077 | 0.0 | 0.3150 | 0.0 |
| 1.6752 | 0.7363 | 7500 | 1.6649 | 0.0077 | 0.0010 | 0.3152 | 0.0010 |
| 1.6813 | 0.7461 | 7600 | 1.6649 | 0.0077 | 0.0002 | 0.3159 | 0.0002 |
| 1.7303 | 0.7559 | 7700 | 1.7081 | 0.0077 | 0.0 | 0.3045 | 0.0 |
| 1.6707 | 0.7658 | 7800 | 1.6670 | 0.0077 | 0.0001 | 0.3224 | 0.0001 |
| 2.067 | 0.7756 | 7900 | 2.0248 | 0.0077 | 0.0 | 0.3152 | 0.0 |
| 1.8621 | 0.7854 | 8000 | 1.7936 | 0.0077 | 0.0 | 0.3245 | 0.0 |
| 1.686 | 0.7952 | 8100 | 1.6674 | 0.0077 | 0.0 | 0.3280 | 0.0 |
| 1.6674 | 0.8050 | 8200 | 1.6914 | 0.0077 | 0.0005 | 0.3167 | 0.0005 |
| 1.6571 | 0.8148 | 8300 | 1.6685 | 0.0077 | 0.0 | 0.3106 | 0.0 |
| 1.657 | 0.8247 | 8400 | 1.6671 | 0.0077 | 0.0 | 0.3195 | 0.0 |
| 1.8014 | 0.8345 | 8500 | 1.8532 | 0.0077 | 0.0 | 0.3275 | 0.0 |
| 1.6653 | 0.8443 | 8600 | 1.6713 | 0.0077 | 0.0001 | 0.3248 | 0.0001 |
| 1.6546 | 0.8541 | 8700 | 1.6645 | 0.0077 | 0.0 | 0.3307 | 0.0 |
| 1.9097 | 0.8639 | 8800 | 1.9613 | 0.0077 | 0.0 | 0.3120 | 0.0 |
| 1.6568 | 0.8737 | 8900 | 1.6670 | 0.0077 | 0.0014 | 0.3208 | 0.0014 |
| 1.7859 | 0.8836 | 9000 | 1.8333 | 0.0077 | 0.0 | 0.3071 | 0.0 |
| 1.8024 | 0.8934 | 9100 | 1.7938 | 0.0077 | 0.0 | 0.3232 | 0.0 |
| 1.7572 | 0.9032 | 9200 | 1.7428 | 0.0077 | 0.0 | 0.3233 | 0.0 |
| 1.6981 | 0.9130 | 9300 | 1.7222 | 0.0077 | 0.0 | 0.3239 | 0.0 |
| 1.7396 | 0.9228 | 9400 | 1.7463 | 0.0077 | 0.0001 | 0.3098 | 0.0001 |
| 1.775 | 0.9327 | 9500 | 1.7762 | 0.0077 | 0.0007 | 0.3198 | 0.0007 |
| 1.7043 | 0.9425 | 9600 | 1.6886 | 0.0077 | 0.0002 | 0.3175 | 0.0002 |
| 1.7136 | 0.9523 | 9700 | 1.7339 | 0.0077 | 0.0 | 0.3190 | 0.0 |
| 1.751 | 0.9621 | 9800 | 1.7346 | 0.0077 | 0.0 | 0.3196 | 0.0 |
| 1.6863 | 0.9719 | 9900 | 1.7042 | 0.0077 | 0.0 | 0.3181 | 0.0 |
| 1.7162 | 0.9817 | 10000 | 1.7233 | 0.0077 | 0.0003 | 0.3166 | 0.0003 |
| 1.6858 | 0.9916 | 10100 | 1.6994 | 0.0077 | 0.0019 | 0.3227 | 0.0019 |
| 1.7365 | 1.0014 | 10200 | 1.7456 | 0.0077 | 0.0 | 0.3163 | 0.0 |
| 1.8212 | 1.0112 | 10300 | 1.8317 | 0.0077 | 0.0 | 0.2922 | 0.0 |
| 1.7881 | 1.0210 | 10400 | 1.7439 | 0.0077 | 0.0015 | 0.3125 | 0.0015 |
| 1.6968 | 1.0308 | 10500 | 1.7073 | 0.0077 | 0.0 | 0.3188 | 0.0 |
| 1.6854 | 1.0406 | 10600 | 1.6875 | 0.0077 | 0.0 | 0.3250 | 0.0 |
| 1.6646 | 1.0505 | 10700 | 1.6760 | 0.0077 | 0.0002 | 0.3209 | 0.0002 |
Framework versions
- PEFT 0.15.2
- Transformers 4.51.3
- Pytorch 2.6.0+cu124
- Datasets 3.5.0
- Tokenizers 0.21.1
- Downloads last month
- -
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support
Model tree for donoway/udgrf609_20250704_025246
Base model
meta-llama/Llama-3.2-1B