File size: 1,268 Bytes
bb27bc8 921b88d bb27bc8 921b88d bb27bc8 921b88d 2428eca bb27bc8 921b88d bb27bc8 921b88d bb27bc8 921b88d bb27bc8 921b88d bb27bc8 921b88d bb27bc8 921b88d bb27bc8 921b88d bb27bc8 921b88d | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 | ---
language:
- en
library_name: peft
pipeline_tag: text-generation
license: apache-2.0
tags:
- text generation
- multitask
- email
- stories
- qa
- summarization
- chat
base_model:
- Qwen/Qwen1.5-1.8B-Chat
---
# Gilbert-Qwen-Multitask-LoRA
LoRA fine-tuned Qwen1.5-1.8B-Chat for multiple text generation tasks including email drafting, story continuation, technical Q&A, news summarization, and chat responses.
## Model Description
This model is a LoRA (Low-Rank Adaptation) adapter fine-tuned on Qwen1.5-1.8B-Chat for multitask text generation across 5 domains:
- **✉️ Email Drafting**: Generate professional email replies
- **📖 Story Continuation**: Continue fictional narratives
- **💻 Technical Q&A**: Answer programming and technical questions
- **📰 News Summarization**: Create concise summaries of articles
- **💬 Chat Responses**: Generate conversational replies
## Training Details
- **Base Model**: `Qwen/Qwen1.5-1.8B-Chat`
- **Fine-tuning Method**: LoRA (Low-Rank Adaptation)
- **Quantization**: 4-bit (QLoRA)
- **Training Tasks**: Multi-task learning
- **Training Steps**: 15,000
- **Learning Rate**: 3e-5
- **Context Length**: 1024 tokens
## Usage
### Installation
```bash
pip install transformers peft torch accelerate bitsandbytes |