--- library_name: transformers license: apache-2.0 base_model: unsloth/Llama-3.2-3B-Instruct tags: - merged - adapter - peft --- # final_model - Merged Model This model is a merged version of the adapter [mekala-2402/model_long](https://huggingface.co/mekala-2402/model_long) with base model [unsloth/Llama-3.2-3B-Instruct](https://huggingface.co/unsloth/Llama-3.2-3B-Instruct). ## Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer import torch model = AutoModelForCausalLM.from_pretrained( "mekala-2402/final_model", torch_dtype=torch.float16, device_map="auto" ) tokenizer = AutoTokenizer.from_pretrained("mekala-2402/final_model") # Generate text inputs = tokenizer("Your prompt here", return_tensors="pt") outputs = model.generate(**inputs, max_length=100) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) ``` ## Model Details - **Original adapter**: mekala-2402/model_long - **Base model**: unsloth/Llama-3.2-3B-Instruct - **Merged using**: PEFT library ## GGUF Conversion To convert this merged model to GGUF format, you can use: 1. [llama.cpp convert script](https://github.com/ggerganov/llama.cpp) 2. [Hugging Face GGUF Space](https://huggingface.co/spaces/ggml-org/gguf-my-repo) ## License This model inherits the license from the base model: unsloth/Llama-3.2-3B-Instruct