mayankpuvvala's picture
Update README.md
c70a513 verified
---
library_name: transformers
tags: [text2text-generation, t5, lora, peft, pytorch, github-issues]
---
# Model Card for `mayankpuvvala/lora-t5-pytorch-issues`
This model represents the LoRA adapter weights trained on a custom dataset of PyTorch GitHub issues. It is intended to be used with the base `t5-small` model to generate issue bodies from titles.
## Model Details
### Model Description
- **Developed by:** Mayank Puvvala
- **Model type:** LoRA Adapter for Text-to-Text Generation
- **Language(s):** English
- **License:** MIT
- **Fine-tuned from model:** [t5-small](https://huggingface.co/t5-small)
### Model Sources
- **Repository:** [GitHub](https://github.com/mayankpuvvala/LLM_FineTune_GenAI)
## How to Get Started with the Model
```python
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
from peft import PeftModel
base_model = AutoModelForSeq2SeqLM.from_pretrained("t5-small")
model = PeftModel.from_pretrained(base_model, "mayankpuvvala/lora-t5-pytorch-issues")
tokenizer = AutoTokenizer.from_pretrained("t5-small")
input_text = "Memory leak when using DataLoader with num_workers > 0"
inputs = tokenizer(input_text, return_tensors="pt")
outputs = model.generate(**inputs)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
## Training Details
### Training Data
- Custom dataset of PyTorch GitHub issues, comprising titles and corresponding bodies.
### Training Procedure
- Fine-tuned using PEFT with LoRA for 3 epochs.
### Training Hyperparameters
- Epochs: 3
- Batch size: 8
## Evaluation Metrics
- ROUGE Precision: 53.12%
- ROUGE F1 Score: 49.8%