naklecha/minecraft-question-answer-700k
Viewer • Updated • 695k • 231 • 44
How to use emre570/gemma-7b-us-minecraft with PEFT:
from peft import PeftModel
from transformers import AutoModelForCausalLM
base_model = AutoModelForCausalLM.from_pretrained("google/gemma-1.1-7b-it")
model = PeftModel.from_pretrained(base_model, "emre570/gemma-7b-us-minecraft")This model is fine-tuned from Unsloth's Gemma 1.1 7B Instruct quantized model with naklecha's Minecraft Question-Answer dataset. Fine-tuned with first 100k rows from dataset with 1 epoch, it took around 2 hours 20 minutes with NVIDIA RTX 4090.
Model can now generate some good answers. But sometimes it can generate inappropriate answers. I think this problem is based on lack of data.
Base model
unsloth/gemma-1.1-7b-it-bnb-4bit