udgrf609_20250704_025246

This model is a fine-tuned version of meta-llama/Llama-3.2-1B on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 1.3280
  • Model Preparation Time: 0.0077
  • Move Accuracy: 0.0219
  • Token Accuracy: 0.4912
  • Accuracy: 0.0219

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.001
  • train_batch_size: 128
  • eval_batch_size: 256
  • seed: 42
  • optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: constant_with_warmup
  • lr_scheduler_warmup_ratio: 0.001
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Model Preparation Time Move Accuracy Token Accuracy Accuracy
No log 0 0 11.9474 0.0077 0.0 0.0000 0.0
1.7249 0.0098 100 1.7275 0.0077 0.0001 0.3398 0.0001
1.5741 0.0196 200 1.5610 0.0077 0.0070 0.4066 0.0070
1.4021 0.0295 300 1.4419 0.0077 0.0093 0.4513 0.0093
1.3719 0.0393 400 1.4199 0.0077 0.0070 0.4577 0.0070
1.3511 0.0491 500 1.3810 0.0077 0.0184 0.4710 0.0184
1.2881 0.0589 600 1.3280 0.0077 0.0219 0.4912 0.0219
1.308 0.0687 700 1.3274 0.0077 0.0208 0.4895 0.0208
1.3919 0.0785 800 1.3478 0.0077 0.0189 0.4783 0.0189
1.4178 0.0884 900 1.4284 0.0077 0.0121 0.4576 0.0121
1.4408 0.0982 1000 1.4418 0.0077 0.0093 0.4347 0.0093
1.5107 0.1080 1100 1.5621 0.0077 0.0023 0.3942 0.0023
1.5004 0.1178 1200 1.4907 0.0077 0.0004 0.4004 0.0004
1.5062 0.1276 1300 1.5485 0.0077 0.0023 0.3994 0.0023
1.5707 0.1374 1400 1.5250 0.0077 0.0021 0.3953 0.0021
1.6335 0.1473 1500 1.6275 0.0077 0.0012 0.3768 0.0012
2.4746 0.1571 1600 2.5088 0.0077 0.0008 0.2987 0.0008
2.4986 0.1669 1700 4.1257 0.0077 0.0 0.1539 0.0
2.2833 0.1767 1800 2.4301 0.0077 0.0 0.1124 0.0
2.5158 0.1865 1900 2.5386 0.0077 0.0 0.2412 0.0
2.114 0.1963 2000 2.0981 0.0077 0.0 0.2817 0.0
2.0344 0.2062 2100 2.0449 0.0077 0.0 0.2517 0.0
2.0666 0.2160 2200 2.0564 0.0077 0.0 0.2705 0.0
2.5462 0.2258 2300 2.6100 0.0077 0.0 0.2328 0.0
1.9724 0.2356 2400 2.0092 0.0077 0.0 0.2819 0.0
1.9781 0.2454 2500 1.9666 0.0077 0.0 0.2799 0.0
2.0801 0.2553 2600 2.0669 0.0077 0.0 0.2506 0.0
2.1864 0.2651 2700 2.1805 0.0077 0.0019 0.2184 0.0019
2.0327 0.2749 2800 2.0283 0.0077 0.0 0.2504 0.0
1.7155 0.2847 2900 1.7586 0.0077 0.0017 0.3205 0.0017
3.4484 0.2945 3000 3.2784 0.0077 0.0 0.1990 0.0
3.7537 0.3043 3100 3.3351 0.0077 0.0 0.1639 0.0
1.7822 0.3142 3200 1.7770 0.0077 0.0 0.2973 0.0
1.965 0.3240 3300 1.9679 0.0077 0.0 0.2606 0.0
1.9091 0.3338 3400 1.9359 0.0077 0.0 0.2708 0.0
2.0503 0.3436 3500 2.0767 0.0077 0.0 0.2203 0.0
2.0584 0.3534 3600 1.9688 0.0077 0.0 0.2537 0.0
1.7356 0.3632 3700 1.7311 0.0077 0.0003 0.3087 0.0003
1.7098 0.3731 3800 1.7020 0.0077 0.0 0.3182 0.0
1.6855 0.3829 3900 1.6984 0.0077 0.0 0.3244 0.0
1.6753 0.3927 4000 1.6868 0.0077 0.0008 0.3204 0.0008
1.7065 0.4025 4100 1.6848 0.0077 0.0 0.3147 0.0
1.6817 0.4123 4200 1.6763 0.0077 0.0002 0.3240 0.0002
1.6763 0.4221 4300 1.6726 0.0077 0.0 0.3204 0.0
1.6801 0.4320 4400 1.6730 0.0077 0.0008 0.3264 0.0008
1.6679 0.4418 4500 1.6748 0.0077 0.0 0.3218 0.0
1.6549 0.4516 4600 1.6692 0.0077 0.0 0.3258 0.0
1.6951 0.4614 4700 1.6681 0.0077 0.0 0.3193 0.0
1.6724 0.4712 4800 1.6708 0.0077 0.0 0.3253 0.0
1.658 0.4811 4900 1.6711 0.0077 0.0002 0.3161 0.0002
1.6677 0.4909 5000 1.6713 0.0077 0.0007 0.3272 0.0007
1.6797 0.5007 5100 1.6752 0.0077 0.0001 0.3190 0.0001
1.6786 0.5105 5200 1.6754 0.0077 0.0 0.3153 0.0
1.661 0.5203 5300 1.6681 0.0077 0.0039 0.3196 0.0039
1.6825 0.5301 5400 1.6894 0.0077 0.0014 0.3132 0.0014
1.658 0.5400 5500 1.6743 0.0077 0.0003 0.3078 0.0003
1.677 0.5498 5600 1.6720 0.0077 0.0001 0.3159 0.0001
1.6896 0.5596 5700 1.6728 0.0077 0.0 0.3091 0.0
1.6795 0.5694 5800 1.6704 0.0077 0.0012 0.3237 0.0012
1.6646 0.5792 5900 1.6657 0.0077 0.0001 0.3249 0.0001
1.6619 0.5890 6000 1.6738 0.0077 0.0 0.3193 0.0
1.6475 0.5989 6100 1.6734 0.0077 0.0 0.3198 0.0
1.7044 0.6087 6200 1.6690 0.0077 0.0 0.3178 0.0
1.6921 0.6185 6300 1.6898 0.0077 0.0 0.3245 0.0
1.6757 0.6283 6400 1.6668 0.0077 0.0001 0.3267 0.0001
1.8259 0.6381 6500 1.8418 0.0077 0.0001 0.3122 0.0001
1.9607 0.6479 6600 1.9611 0.0077 0.0 0.3098 0.0
1.6544 0.6578 6700 1.6741 0.0077 0.0006 0.3154 0.0006
1.6718 0.6676 6800 1.6760 0.0077 0.0001 0.3200 0.0001
2.1974 0.6774 6900 2.1437 0.0077 0.0002 0.3086 0.0002
1.9982 0.6872 7000 2.0012 0.0077 0.0001 0.2950 0.0001
1.7201 0.6970 7100 1.7016 0.0077 0.0 0.3156 0.0
1.7066 0.7069 7200 1.6970 0.0077 0.0 0.3155 0.0
1.6948 0.7167 7300 1.6821 0.0077 0.0 0.3244 0.0
1.6561 0.7265 7400 1.6726 0.0077 0.0 0.3150 0.0
1.6752 0.7363 7500 1.6649 0.0077 0.0010 0.3152 0.0010
1.6813 0.7461 7600 1.6649 0.0077 0.0002 0.3159 0.0002
1.7303 0.7559 7700 1.7081 0.0077 0.0 0.3045 0.0
1.6707 0.7658 7800 1.6670 0.0077 0.0001 0.3224 0.0001
2.067 0.7756 7900 2.0248 0.0077 0.0 0.3152 0.0
1.8621 0.7854 8000 1.7936 0.0077 0.0 0.3245 0.0
1.686 0.7952 8100 1.6674 0.0077 0.0 0.3280 0.0
1.6674 0.8050 8200 1.6914 0.0077 0.0005 0.3167 0.0005
1.6571 0.8148 8300 1.6685 0.0077 0.0 0.3106 0.0
1.657 0.8247 8400 1.6671 0.0077 0.0 0.3195 0.0
1.8014 0.8345 8500 1.8532 0.0077 0.0 0.3275 0.0
1.6653 0.8443 8600 1.6713 0.0077 0.0001 0.3248 0.0001
1.6546 0.8541 8700 1.6645 0.0077 0.0 0.3307 0.0
1.9097 0.8639 8800 1.9613 0.0077 0.0 0.3120 0.0
1.6568 0.8737 8900 1.6670 0.0077 0.0014 0.3208 0.0014
1.7859 0.8836 9000 1.8333 0.0077 0.0 0.3071 0.0
1.8024 0.8934 9100 1.7938 0.0077 0.0 0.3232 0.0
1.7572 0.9032 9200 1.7428 0.0077 0.0 0.3233 0.0
1.6981 0.9130 9300 1.7222 0.0077 0.0 0.3239 0.0
1.7396 0.9228 9400 1.7463 0.0077 0.0001 0.3098 0.0001
1.775 0.9327 9500 1.7762 0.0077 0.0007 0.3198 0.0007
1.7043 0.9425 9600 1.6886 0.0077 0.0002 0.3175 0.0002
1.7136 0.9523 9700 1.7339 0.0077 0.0 0.3190 0.0
1.751 0.9621 9800 1.7346 0.0077 0.0 0.3196 0.0
1.6863 0.9719 9900 1.7042 0.0077 0.0 0.3181 0.0
1.7162 0.9817 10000 1.7233 0.0077 0.0003 0.3166 0.0003
1.6858 0.9916 10100 1.6994 0.0077 0.0019 0.3227 0.0019
1.7365 1.0014 10200 1.7456 0.0077 0.0 0.3163 0.0
1.8212 1.0112 10300 1.8317 0.0077 0.0 0.2922 0.0
1.7881 1.0210 10400 1.7439 0.0077 0.0015 0.3125 0.0015
1.6968 1.0308 10500 1.7073 0.0077 0.0 0.3188 0.0
1.6854 1.0406 10600 1.6875 0.0077 0.0 0.3250 0.0
1.6646 1.0505 10700 1.6760 0.0077 0.0002 0.3209 0.0002

Framework versions

  • PEFT 0.15.2
  • Transformers 4.51.3
  • Pytorch 2.6.0+cu124
  • Datasets 3.5.0
  • Tokenizers 0.21.1
Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for donoway/udgrf609_20250704_025246

Adapter
(636)
this model