| # GPT-2 Fine-tuned Model by ps2program | |
| This is a GPT-2 model fine-tuned on [your dataset name, e.g., TinyStories]. | |
| ## Model Details | |
| - Base Model: GPT-2 (small) | |
| - Fine-tuning method: LoRA adapters merged into base model | |
| - Dataset: TinyStories or specify your dataset | |
| - Training epochs: 3 | |
| - Batch size: 4 | |
| - Framework: Hugging Face Transformers + PEFT | |
| ## Usage | |
| ```python | |
| from transformers import AutoTokenizer, AutoModelForCausalLM | |
| model_name = "ps2program/gpt2-finetuned-ps2prahlad" | |
| tokenizer = AutoTokenizer.from_pretrained(model_name) | |
| model = AutoModelForCausalLM.from_pretrained(model_name) | |
| prompt = "Once upon a time" | |
| inputs = tokenizer(prompt, return_tensors="pt") | |
| outputs = model.generate(**inputs, max_length=50, do_sample=True, temperature=0.7) | |
| print(tokenizer.decode(outputs[0], skip_special_tokens=True)) | |