Llama_Instruct

Mini-Llama 3B Instruct - 0124

My base pretrain model has undergone full fine-tuning on an additional 350M tokens using portions of Tulu 3 and Nvidia Nemotron instruct sets. It is rough but functionsl, and still needs DPO training to align it with human preferences.

For the base pretrain, see: Nabbers1999/Mini-Llama-3B-Base-0124

** Special note and edit 01/27 - Ministral 3 3B specifically uses tied_word_embeddings set to true in the config.json, which was present in my base fine tune. However after instruct training this model was saved with it set to false, likely due to the trainer confusing it for another model due to the llamaficiation. I have now fixed the flag in this model. If you train on this model be aware that this is something to watch for.

Downloads last month
88
Safetensors
Model size
3B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Nabbers1999/Mini-Llama-3B-Instruct-0124

Finetuned
(25)
this model
Quantizations
1 model

Datasets used to train Nabbers1999/Mini-Llama-3B-Instruct-0124

Collection including Nabbers1999/Mini-Llama-3B-Instruct-0124