Safetensors
Persian
gemma2
gemma2-2b-fa-ft / README.md
cnababaie's picture
Update README.md
8ab707a verified
metadata
license: gemma
datasets:
  - Z-Jafari/PersianQuAD
  - SajjadAyoubi/persian_qa
  - cnababaie/180-fa-meter-qa
language:
  - fa
base_model:
  - google/gemma-2-2b

Fine-tuned Gemma 2 2b on Persian

This a Gemma 2 2b model fined tuned on Persian instrucion datasets I found out in huggingface, due to the small size, could be used in various general tasks.

Inference

This is a fine-tuned version of Gemma 2 2b model, which means it could be used in same manner:

# pip install accelerate
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch

tokenizer = AutoTokenizer.from_pretrained("cnababaie/gemma2-2b-fa-ft")
model = AutoModelForCausalLM.from_pretrained(
    "cnababaie/gemma2-2b-fa-ft",
    device_map="auto",
)

input_text = "این متنو به انگلیسی ترجمه کن: خاک از شدت باران به گل تبدیل شد."
input_ids = tokenizer(input_text, return_tensors="pt").to("cuda")

outputs = model.generate(**input_ids, max_new_tokens=32)
print(tokenizer.decode(outputs[0]))

This will correctly translate as "The soil turned to mud from the heavy rain." whereas the original gemma 2 2b model would give wrong repetitive answer.