GPT-2 Medium LoRA Adapter for Tatar Language
This is a LoRA adapter for GPT-2 medium fine-tuned on Tatar language.
π Model Details
| Property | Value |
|---|---|
| Base Model | GPT-2 medium |
| LoRA Rank | 16 |
| Training Data | 10,000 samples |
| Test Perplexity | 5.31 |
| Training Epochs | 3 |
π Usage
from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel
base_model = "gpt2-medium"
tokenizer = AutoTokenizer.from_pretrained(base_model)
tokenizer.pad_token = tokenizer.eos_token
model = AutoModelForCausalLM.from_pretrained(base_model)
model = PeftModel.from_pretrained(model, "TatarNLPWorld/gpt2-medium-tatar-lora-r16")
prompt = "ΠΠΈΠ½Π΅ΠΌ ΠΈΡΠ΅ΠΌΠ΅ΠΌ"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=50)
print(tokenizer.decode(outputs[0]))
π‘ Why GPT-2 medium?
- Lightweight (can run on CPU)
- Good baseline for comparison with 7B models
- Perplexity 5.31 β solid performance for its size
π Generation Example
Prompt: "ΠΠΈΠ½Π΅ΠΌ ΠΈΡΠ΅ΠΌΠ΅ΠΌ"
Generated: "... ΠΠΉΡΠ°Ρ. ΠΠΈΠ½ ΠΠ°Π·Π°Π½Π΄Π° ΡΡΠΈΠΌ."
π Performance
| Model | Parameters | Perplexity |
|---|---|---|
| GPT-2 medium (full) | 354M | 2.91 |
| GPT-2 medium + LoRA r16 | 1.09M | 5.31 |
| Parameter reduction | 325x | +82% |
π₯ Authors
- Arabov Mullosharaf Kurbonovich
π License
MIT (same as original GPT-2)
π€ Citation
@software{gpt2_medium_tatar_lora_2026,
title = {{GPT-2 Medium LoRA Adapter for Tatar Language}},
author = {Arabov Mullosharaf Kurbonovich},
year = {2026},
publisher = {Hugging Face},
url = {https://huggingface.co/TatarNLPWorld/gpt2-medium-tatar-lora-r16}
}
Model tree for TatarNLPWorld/gpt2-medium-tatar-lora-r16
Base model
openai-community/gpt2-medium