pradeep6kumar2024's picture
Upload folder using huggingface_hub
d95b348 verified
# Phi-2 QLoRA Fine-tuned Assistant
This is a fine-tuned version of Microsoft's Phi-2 model using QLoRA (Quantized Low-Rank Adaptation). The model has been trained to provide helpful responses for various tasks including coding, writing, and general assistance.
## Model Details
- **Base Model**: Microsoft Phi-2 (2.7B parameters)
- **Fine-tuning Method**: QLoRA (4-bit quantization)
- **Training Data**: Custom dataset focused on programming and professional communication
- **Hardware Used**: NVIDIA RTX 4090 (24GB VRAM)
## Usage
You can interact with the model through the Gradio interface by visiting the "Spaces" tab of this repository.
### Local Installation
To run the model locally:
1. Clone this repository
2. Install dependencies:
```bash
pip install -r requirements.txt
```
3. Run the Gradio app:
```bash
python gradio_app.py
```
### Parameters
- **Max Length**: Controls the maximum length of the generated response (64-1024 tokens)
- **Temperature**: Controls randomness in generation (0.1-1.0)
- **Top P**: Controls diversity of generated responses (0.1-1.0)
## Example Prompts
1. "Write a Python function to calculate the factorial of a number"
2. "Explain the concept of machine learning in simple terms"
3. "Write a professional email requesting a meeting with a client"
## Limitations
- The model works best with English language input
- Response quality may vary depending on the complexity of the prompt
- Maximum context length is limited to 2048 tokens
## License
This model is subject to the Microsoft Phi-2 license terms and conditions.
## Acknowledgments
- Microsoft for the Phi-2 base model
- Hugging Face for the transformers library and model hosting
- The QLoRA paper authors for the quantization technique