| language: en | |
| license: apache-2.0 | |
| tags: | |
| - merged | |
| - lora | |
| - text-generation | |
| # llama-3b-instruct-ft-function-call | |
| This model is a merged version of the base model [meta-llama/Llama-3.2-3B-Instruct] with the following LoRA adapter(s): | |
| - /home/ubuntu/zona/decision-step-model/train/mar16_sembalanced_data | |
| ## Description | |
| ## Usage | |
| ```python | |
| from transformers import AutoModelForCausalLM, AutoTokenizer | |
| model_id = "matt-bcny/llama-3b-instruct-ft-function-call" | |
| # Load tokenizer and model | |
| tokenizer = AutoTokenizer.from_pretrained(model_id) | |
| model = AutoModelForCausalLM.from_pretrained( | |
| model_id, | |
| torch_dtype=torch.float16, | |
| device_map="auto" | |
| ) | |
| # Example inference | |
| prompt = "Your prompt here" | |
| inputs = tokenizer(prompt, return_tensors="pt").to(model.device) | |
| outputs = model.generate(**inputs, max_length=100) | |
| response = tokenizer.decode(outputs[0], skip_special_tokens=True) | |
| print(response) | |
| ``` | |
| ## Model creation | |
| This model was created by merging the base model with LoRA adapter(s) on 2025-03-17. | |