Mistral_Test / Modellora.py
eesfeg's picture
Add fine-tuned Mistral model
3ac4614
raw
history blame contribute delete
717 Bytes
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel, LoraConfig
import os
from huggingface_hub import HfApi, ModelCard
# Load base model and your fine-tuned LoRA adapters
base_model = "mistralai/Mistral-7B-v0.1"
adapter_path = "./lora-adapters" # Your trained LoRA adapters
model = AutoModelForCausalLM.from_pretrained(
base_model,
torch_dtype=torch.float16,
device_map="auto"
)
# Load adapters
model = PeftModel.from_pretrained(model, adapter_path)
# Save adapters for upload
model.save_pretrained("./hf-upload", safe_serialization=True)
# Save tokenizer
tokenizer = AutoTokenizer.from_pretrained(base_model)
tokenizer.save_pretrained("./hf-upload")