k738hh2f_20250704_111855
This model is a fine-tuned version of meta-llama/Llama-3.2-1B on an unknown dataset. It achieves the following results on the evaluation set:
- Loss: 1.3596
- Model Preparation Time: 0.0284
- Move Accuracy: 0.0148
- Token Accuracy: 0.4816
- Accuracy: 0.0148
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.001
- train_batch_size: 128
- eval_batch_size: 256
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: constant_with_warmup
- lr_scheduler_warmup_ratio: 0.001
- num_epochs: 100
Training results
| Training Loss | Epoch | Step | Validation Loss | Model Preparation Time | Move Accuracy | Token Accuracy | Accuracy |
|---|---|---|---|---|---|---|---|
| No log | 0 | 0 | 11.9474 | 0.0284 | 0.0 | 0.0000 | 0.0 |
| 1.6745 | 0.0098 | 100 | 1.7186 | 0.0284 | 0.0014 | 0.3616 | 0.0014 |
| 1.5711 | 0.0196 | 200 | 1.5492 | 0.0284 | 0.0041 | 0.4130 | 0.0041 |
| 1.38 | 0.0295 | 300 | 1.4138 | 0.0284 | 0.0101 | 0.4606 | 0.0101 |
| 1.3688 | 0.0393 | 400 | 1.3840 | 0.0284 | 0.0135 | 0.4673 | 0.0135 |
| 1.3486 | 0.0491 | 500 | 1.3596 | 0.0284 | 0.0148 | 0.4816 | 0.0148 |
| 1.3265 | 0.0589 | 600 | 1.3538 | 0.0284 | 0.0140 | 0.4781 | 0.0140 |
| 1.4607 | 0.0687 | 700 | 1.4251 | 0.0284 | 0.0123 | 0.4507 | 0.0123 |
| 1.5519 | 0.0785 | 800 | 1.5275 | 0.0284 | 0.0059 | 0.3862 | 0.0059 |
| 1.634 | 0.0884 | 900 | 1.6384 | 0.0284 | 0.0025 | 0.3704 | 0.0025 |
| 1.7167 | 0.0982 | 1000 | 1.7353 | 0.0284 | 0.0011 | 0.3243 | 0.0011 |
| 1.7161 | 0.1080 | 1100 | 1.7294 | 0.0284 | 0.0014 | 0.3460 | 0.0014 |
| 2.3118 | 0.1178 | 1200 | 1.9940 | 0.0284 | 0.0 | 0.3201 | 0.0 |
| 1.8426 | 0.1276 | 1300 | 1.8767 | 0.0284 | 0.0002 | 0.3057 | 0.0002 |
| 1.6701 | 0.1374 | 1400 | 1.6449 | 0.0284 | 0.0001 | 0.3295 | 0.0001 |
| 1.6402 | 0.1473 | 1500 | 1.6638 | 0.0284 | 0.0 | 0.3418 | 0.0 |
| 1.5796 | 0.1571 | 1600 | 1.6271 | 0.0284 | 0.0 | 0.3401 | 0.0 |
| 1.6263 | 0.1669 | 1700 | 1.6458 | 0.0284 | 0.0 | 0.3436 | 0.0 |
| 1.5907 | 0.1767 | 1800 | 1.6298 | 0.0284 | 0.0 | 0.3578 | 0.0 |
| 1.7561 | 0.1865 | 1900 | 1.7856 | 0.0284 | 0.0001 | 0.3409 | 0.0001 |
| 1.6389 | 0.1963 | 2000 | 1.6434 | 0.0284 | 0.0003 | 0.3160 | 0.0003 |
| 1.6 | 0.2062 | 2100 | 1.6313 | 0.0284 | 0.0 | 0.3492 | 0.0 |
| 1.6269 | 0.2160 | 2200 | 1.6260 | 0.0284 | 0.0001 | 0.3579 | 0.0001 |
| 1.7546 | 0.2258 | 2300 | 1.7161 | 0.0284 | 0.0010 | 0.3369 | 0.0010 |
| 2.539 | 0.2356 | 2400 | 3.0176 | 0.0284 | 0.0 | 0.1325 | 0.0 |
| 1.8561 | 0.2454 | 2500 | 1.7762 | 0.0284 | 0.0001 | 0.3268 | 0.0001 |
| 2.0533 | 0.2553 | 2600 | 2.0227 | 0.0284 | 0.0 | 0.2670 | 0.0 |
| 1.662 | 0.2651 | 2700 | 1.6329 | 0.0284 | 0.0008 | 0.3565 | 0.0008 |
| 1.6183 | 0.2749 | 2800 | 1.6258 | 0.0284 | 0.0004 | 0.3597 | 0.0004 |
| 1.6211 | 0.2847 | 2900 | 1.6357 | 0.0284 | 0.0002 | 0.3532 | 0.0002 |
| 1.7325 | 0.2945 | 3000 | 1.7830 | 0.0284 | 0.0002 | 0.3256 | 0.0002 |
| 1.5988 | 0.3043 | 3100 | 1.6332 | 0.0284 | 0.0011 | 0.3550 | 0.0011 |
| 1.6765 | 0.3142 | 3200 | 1.6680 | 0.0284 | 0.0001 | 0.3485 | 0.0001 |
| 1.6235 | 0.3240 | 3300 | 1.6262 | 0.0284 | 0.0019 | 0.3564 | 0.0019 |
| 1.6338 | 0.3338 | 3400 | 1.6552 | 0.0284 | 0.0 | 0.3385 | 0.0 |
| 1.635 | 0.3436 | 3500 | 1.6406 | 0.0284 | 0.0004 | 0.3575 | 0.0004 |
| 1.6468 | 0.3534 | 3600 | 1.6206 | 0.0284 | 0.0001 | 0.3518 | 0.0001 |
| 1.6627 | 0.3632 | 3700 | 1.6251 | 0.0284 | 0.0015 | 0.3545 | 0.0015 |
| 1.6412 | 0.3731 | 3800 | 1.6224 | 0.0284 | 0.0 | 0.3570 | 0.0 |
| 2.2711 | 0.3829 | 3900 | 2.3388 | 0.0284 | 0.0003 | 0.3216 | 0.0003 |
| 1.6091 | 0.3927 | 4000 | 1.6207 | 0.0284 | 0.0002 | 0.3584 | 0.0002 |
| 2.339 | 0.4025 | 4100 | 2.2785 | 0.0284 | 0.0001 | 0.3205 | 0.0001 |
| 2.0048 | 0.4123 | 4200 | 2.0092 | 0.0284 | 0.0 | 0.2964 | 0.0 |
| 3.1865 | 0.4221 | 4300 | 3.3007 | 0.0284 | 0.0008 | 0.3108 | 0.0008 |
| 2.0961 | 0.4320 | 4400 | 2.0929 | 0.0284 | 0.0001 | 0.2832 | 0.0001 |
| 2.1274 | 0.4418 | 4500 | 2.1945 | 0.0284 | 0.0 | 0.2653 | 0.0 |
| 4.318 | 0.4516 | 4600 | 4.6243 | 0.0284 | 0.0 | 0.2507 | 0.0 |
| 2.0958 | 0.4614 | 4700 | 2.1172 | 0.0284 | 0.0004 | 0.2738 | 0.0004 |
| 2.0915 | 0.4712 | 4800 | 2.0908 | 0.0284 | 0.0002 | 0.2842 | 0.0002 |
| 2.1342 | 0.4811 | 4900 | 2.1132 | 0.0284 | 0.0011 | 0.2955 | 0.0011 |
| 1.7981 | 0.4909 | 5000 | 1.8179 | 0.0284 | 0.0001 | 0.3048 | 0.0001 |
| 1.8168 | 0.5007 | 5100 | 1.8305 | 0.0284 | 0.0 | 0.3040 | 0.0 |
| 2.8595 | 0.5105 | 5200 | 2.8229 | 0.0284 | 0.0006 | 0.2721 | 0.0006 |
| 1.7049 | 0.5203 | 5300 | 1.7105 | 0.0284 | 0.0001 | 0.3138 | 0.0001 |
| 1.7189 | 0.5301 | 5400 | 1.7360 | 0.0284 | 0.0 | 0.3074 | 0.0 |
| 1.7243 | 0.5400 | 5500 | 1.7497 | 0.0284 | 0.0 | 0.3235 | 0.0 |
| 1.7073 | 0.5498 | 5600 | 1.7499 | 0.0284 | 0.0012 | 0.3226 | 0.0012 |
| 1.7547 | 0.5596 | 5700 | 1.7342 | 0.0284 | 0.0 | 0.3195 | 0.0 |
| 1.7245 | 0.5694 | 5800 | 1.7210 | 0.0284 | 0.0 | 0.3258 | 0.0 |
| 1.7206 | 0.5792 | 5900 | 1.7074 | 0.0284 | 0.0015 | 0.3306 | 0.0015 |
| 1.9771 | 0.5890 | 6000 | 1.9495 | 0.0284 | 0.0002 | 0.3278 | 0.0002 |
| 1.6606 | 0.5989 | 6100 | 1.6839 | 0.0284 | 0.0 | 0.3273 | 0.0 |
| 1.712 | 0.6087 | 6200 | 1.6732 | 0.0284 | 0.0008 | 0.3252 | 0.0008 |
| 1.6719 | 0.6185 | 6300 | 1.6712 | 0.0284 | 0.0003 | 0.3320 | 0.0003 |
| 1.6592 | 0.6283 | 6400 | 1.6674 | 0.0284 | 0.0010 | 0.3243 | 0.0010 |
| 1.6723 | 0.6381 | 6500 | 1.6660 | 0.0284 | 0.0002 | 0.3285 | 0.0002 |
| 1.6551 | 0.6479 | 6600 | 1.6638 | 0.0284 | 0.0001 | 0.3250 | 0.0001 |
| 1.6506 | 0.6578 | 6700 | 1.6655 | 0.0284 | 0.0018 | 0.3264 | 0.0018 |
| 1.6578 | 0.6676 | 6800 | 1.6628 | 0.0284 | 0.0011 | 0.3285 | 0.0011 |
| 1.6668 | 0.6774 | 6900 | 1.6674 | 0.0284 | 0.0011 | 0.3331 | 0.0011 |
| 1.6681 | 0.6872 | 7000 | 1.6675 | 0.0284 | 0.0003 | 0.3240 | 0.0003 |
| 1.6702 | 0.6970 | 7100 | 1.6594 | 0.0284 | 0.0001 | 0.3305 | 0.0001 |
| 1.6456 | 0.7069 | 7200 | 1.6545 | 0.0284 | 0.0008 | 0.3379 | 0.0008 |
| 2.1828 | 0.7167 | 7300 | 2.2745 | 0.0284 | 0.0016 | 0.3015 | 0.0016 |
| 1.6792 | 0.7265 | 7400 | 1.7032 | 0.0284 | 0.0 | 0.3361 | 0.0 |
| 1.7602 | 0.7363 | 7500 | 1.9407 | 0.0284 | 0.0 | 0.3078 | 0.0 |
| 1.6892 | 0.7461 | 7600 | 1.6708 | 0.0284 | 0.0019 | 0.3324 | 0.0019 |
| 1.6843 | 0.7559 | 7700 | 1.6641 | 0.0284 | 0.0003 | 0.3279 | 0.0003 |
| 1.6427 | 0.7658 | 7800 | 1.6510 | 0.0284 | 0.0005 | 0.3372 | 0.0005 |
| 1.622 | 0.7756 | 7900 | 1.6432 | 0.0284 | 0.0001 | 0.3477 | 0.0001 |
| 1.6188 | 0.7854 | 8000 | 1.6426 | 0.0284 | 0.0 | 0.3508 | 0.0 |
| 1.6679 | 0.7952 | 8100 | 1.6952 | 0.0284 | 0.0013 | 0.3365 | 0.0013 |
| 1.768 | 0.8050 | 8200 | 1.8825 | 0.0284 | 0.0009 | 0.3157 | 0.0009 |
| 1.9783 | 0.8148 | 8300 | 1.9895 | 0.0284 | 0.0 | 0.3086 | 0.0 |
| 1.9042 | 0.8247 | 8400 | 1.9861 | 0.0284 | 0.0001 | 0.3220 | 0.0001 |
| 1.7839 | 0.8345 | 8500 | 1.7791 | 0.0284 | 0.0001 | 0.3247 | 0.0001 |
| 1.9037 | 0.8443 | 8600 | 1.9608 | 0.0284 | 0.0004 | 0.3194 | 0.0004 |
| 1.7279 | 0.8541 | 8700 | 1.7481 | 0.0284 | 0.0011 | 0.3350 | 0.0011 |
| 1.7755 | 0.8639 | 8800 | 1.8012 | 0.0284 | 0.0 | 0.3283 | 0.0 |
| 1.7078 | 0.8737 | 8900 | 1.7299 | 0.0284 | 0.0009 | 0.3299 | 0.0009 |
| 2.024 | 0.8836 | 9000 | 1.9605 | 0.0284 | 0.0 | 0.3181 | 0.0 |
| 1.7749 | 0.8934 | 9100 | 1.7716 | 0.0284 | 0.0002 | 0.3150 | 0.0002 |
| 1.7571 | 0.9032 | 9200 | 1.7663 | 0.0284 | 0.0 | 0.3177 | 0.0 |
| 1.9204 | 0.9130 | 9300 | 1.8473 | 0.0284 | 0.0 | 0.3027 | 0.0 |
| 1.8345 | 0.9228 | 9400 | 1.8094 | 0.0284 | 0.0 | 0.3062 | 0.0 |
| 1.8099 | 0.9327 | 9500 | 1.8055 | 0.0284 | 0.0004 | 0.3086 | 0.0004 |
| 1.7039 | 0.9425 | 9600 | 1.7407 | 0.0284 | 0.0018 | 0.3178 | 0.0018 |
| 1.7355 | 0.9523 | 9700 | 1.7300 | 0.0284 | 0.0001 | 0.3165 | 0.0001 |
| 1.6975 | 0.9621 | 9800 | 1.7018 | 0.0284 | 0.0005 | 0.3150 | 0.0005 |
| 1.768 | 0.9719 | 9900 | 1.7621 | 0.0284 | 0.0 | 0.3184 | 0.0 |
| 1.7032 | 0.9817 | 10000 | 1.7266 | 0.0284 | 0.0 | 0.3216 | 0.0 |
| 1.8761 | 0.9916 | 10100 | 1.7377 | 0.0284 | 0.0003 | 0.3185 | 0.0003 |
| 1.713 | 1.0014 | 10200 | 1.7070 | 0.0284 | 0.0 | 0.3222 | 0.0 |
| 1.8568 | 1.0112 | 10300 | 1.8196 | 0.0284 | 0.0004 | 0.3085 | 0.0004 |
| 2.0375 | 1.0210 | 10400 | 2.0177 | 0.0284 | 0.0 | 0.3063 | 0.0 |
| 1.9583 | 1.0308 | 10500 | 1.9550 | 0.0284 | 0.0003 | 0.3106 | 0.0003 |
| 1.8216 | 1.0406 | 10600 | 1.8168 | 0.0284 | 0.0 | 0.3027 | 0.0 |
Framework versions
- PEFT 0.15.2
- Transformers 4.51.3
- Pytorch 2.6.0+cu124
- Datasets 3.5.0
- Tokenizers 0.21.1
- Downloads last month
- 1
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support
Model tree for donoway/k738hh2f_20250704_111855
Base model
meta-llama/Llama-3.2-1B