--- library_name: transformers tags: [text2text-generation, t5, lora, peft, pytorch, github-issues] --- # Model Card for `mayankpuvvala/lora-t5-pytorch-issues` This model represents the LoRA adapter weights trained on a custom dataset of PyTorch GitHub issues. It is intended to be used with the base `t5-small` model to generate issue bodies from titles. ## Model Details ### Model Description - **Developed by:** Mayank Puvvala - **Model type:** LoRA Adapter for Text-to-Text Generation - **Language(s):** English - **License:** MIT - **Fine-tuned from model:** [t5-small](https://huggingface.co/t5-small) ### Model Sources - **Repository:** [GitHub](https://github.com/mayankpuvvala/LLM_FineTune_GenAI) ## How to Get Started with the Model ```python from transformers import AutoTokenizer, AutoModelForSeq2SeqLM from peft import PeftModel base_model = AutoModelForSeq2SeqLM.from_pretrained("t5-small") model = PeftModel.from_pretrained(base_model, "mayankpuvvala/lora-t5-pytorch-issues") tokenizer = AutoTokenizer.from_pretrained("t5-small") input_text = "Memory leak when using DataLoader with num_workers > 0" inputs = tokenizer(input_text, return_tensors="pt") outputs = model.generate(**inputs) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) ``` ## Training Details ### Training Data - Custom dataset of PyTorch GitHub issues, comprising titles and corresponding bodies. ### Training Procedure - Fine-tuned using PEFT with LoRA for 3 epochs. ### Training Hyperparameters - Epochs: 3 - Batch size: 8 ## Evaluation Metrics - ROUGE Precision: 53.12% - ROUGE F1 Score: 49.8%