File size: 1,808 Bytes
d95b348
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
# Phi-2 QLoRA Fine-tuned Assistant

This is a fine-tuned version of Microsoft's Phi-2 model using QLoRA (Quantized Low-Rank Adaptation). The model has been trained to provide helpful responses for various tasks including coding, writing, and general assistance.

## Model Details

- **Base Model**: Microsoft Phi-2 (2.7B parameters)
- **Fine-tuning Method**: QLoRA (4-bit quantization)
- **Training Data**: Custom dataset focused on programming and professional communication
- **Hardware Used**: NVIDIA RTX 4090 (24GB VRAM)

## Usage

You can interact with the model through the Gradio interface by visiting the "Spaces" tab of this repository.

### Local Installation

To run the model locally:

1. Clone this repository
2. Install dependencies:
```bash

pip install -r requirements.txt

```
3. Run the Gradio app:
```bash

python gradio_app.py

```

### Parameters

- **Max Length**: Controls the maximum length of the generated response (64-1024 tokens)
- **Temperature**: Controls randomness in generation (0.1-1.0)
- **Top P**: Controls diversity of generated responses (0.1-1.0)

## Example Prompts

1. "Write a Python function to calculate the factorial of a number"
2. "Explain the concept of machine learning in simple terms"
3. "Write a professional email requesting a meeting with a client"

## Limitations

- The model works best with English language input
- Response quality may vary depending on the complexity of the prompt
- Maximum context length is limited to 2048 tokens

## License

This model is subject to the Microsoft Phi-2 license terms and conditions.

## Acknowledgments

- Microsoft for the Phi-2 base model
- Hugging Face for the transformers library and model hosting
- The QLoRA paper authors for the quantization technique