My Optimal Model - Merged
Full merged Mistral-7B model with LoRA weights integrated.
Model Details
- Base: mistralai/Mistral-7B-v0.1
- LoRA Adapter: yamraj047/my_optimal_model
- Format: Full model (not adapter)
- Size: ~14.5 GB
- Precision: FP16
Usage
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
model = AutoModelForCausalLM.from_pretrained(
"yamraj047/my_optimal_model-merged",
torch_dtype=torch.float16,
device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained("yamraj047/my_optimal_model-merged")
inputs = tokenizer("Your prompt here", return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_length=200)
print(tokenizer.decode(outputs[0]))
Convert to GGUF
This model can be converted to GGUF format for CPU inference:
# Use the Nepal Legal GGUF conversion script
ORIGINAL_MODEL = "yamraj047/my_optimal_model-merged"
GGUF_REPO = "yamraj047/my_optimal_model-GGUF"
# ... run conversion script
Files Structure
Same format as standard HuggingFace models:
model-*.safetensors- Model weight shardsconfig.json- Model configurationtokenizer.json,tokenizer.model- Tokenizer filesgeneration_config.json- Generation parameters
License
Apache 2.0
- Downloads last month
- -
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
馃檵
Ask for provider support
Model tree for yamraj047/my_optimal_model-merged
Base model
mistralai/Mistral-7B-v0.1