KRAFTYUX commited on
Commit
0fb5d0a
·
verified ·
1 Parent(s): ea6ce54

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +67 -0
README.md ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ library_name: gguf
4
+ tags:
5
+ - phi3
6
+ - prompt-engineering
7
+ - gguf
8
+ - quantization
9
+ base_model: microsoft/Phi-3-mini-4k-instruct
10
+ ---
11
+
12
+ # Phi-3 Prompt Engineer (GGUF)
13
+
14
+ This repository contains **GGUF** quantized versions of the **Phi-3 Prompt Engineer** model, fine-tuned to excel at refining user requests into detailed, structured prompts for LLMs.
15
+
16
+ ## Model Description
17
+
18
+ - **Base Model**: [microsoft/Phi-3-mini-4k-instruct](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct)
19
+ - **Fine-tuning**: LoRA adapters trained on a custom dataset of prompt refinement examples.
20
+ - **Purpose**: To act as a specialized "Prompt Engineer" agent, converting vague user ideas into high-quality, actionable system instructions and prompts.
21
+
22
+ ## Available Quantizations
23
+
24
+ The following GGUF files are available for download. **Q4_K_M** is recommended for most users as it balances performance and quality perfectly.
25
+
26
+ | Filename | Quantization | Size | SHA256 Checksum |
27
+ |----------|--------------|------|-----------------|
28
+ | phi3-prompt-engineer-f16.gguf | F16 | 7.12 GB | `1bc8a41027c400eda38d5dc9cf97fcf0a4617072d2c3c132b922dd2589783c16` |
29
+ | phi3-prompt-engineer-q4_k_m.gguf | Q4_K_M | 2.23 GB | `cc840e37e0f21c97bf158211a2f0dda1096294ee47ab3f199da854cc841c39f7` |
30
+ | phi3-prompt-engineer-q5_k_m.gguf | Q5_K_M | 2.62 GB | `218b8021b5f1b1efb2e0914d939f5464d8cb790ef9d9e35eaa05e1b8b3cc560f` |
31
+ | phi3-prompt-engineer-q6_k.gguf | Q6_K | 2.92 GB | `4cf033bfe14c43ebf84e47f7184f2f0f5286a30a2c87927d59aa95d9b4a455d5` |
32
+ | phi3-prompt-engineer-q8_0.gguf | Q8_0 | 3.78 GB | `2aa4f14b038c01a50ad8e3306832a25b8c6f704aa1d9efa29a88856dee924fce` |
33
+
34
+ ## Usage
35
+
36
+ ### Ollama
37
+
38
+ You can use these files directly with [Ollama](https://ollama.com).
39
+
40
+ 1. Create a `Modelfile`:
41
+ ```dockerfile
42
+ FROM ./phi3-prompt-engineer-q4_k_m.gguf
43
+ SYSTEM "You are an Expert Prompt Engineer. Your goal is to refine the following user request into a clear, structured, and detailed prompt."
44
+ ```
45
+
46
+ 2. Create and run the model:
47
+ ```bash
48
+ ollama create phi3-pe -f Modelfile
49
+ ollama run phi3-pe
50
+ ```
51
+
52
+ ### llama.cpp
53
+
54
+ ```bash
55
+ ./main -m phi3-prompt-engineer-q4_k_m.gguf -p "You are an Expert Prompt Engineer. Refine this: make a snake game in python"
56
+ ```
57
+
58
+ ## Training Data
59
+
60
+ The model was trained on a proprietary dataset (`raw_dataset.jsonl`) consisting of:
61
+ - **Input**: Raw, often vague user requests (e.g., "make a login page").
62
+ - **Output**: Detailed, structured Prompt Engineering responses including personas, constraints, and specific requirements.
63
+ - **Focus**: Web development, Python scripting, creative writing, and system design tasks.
64
+
65
+ ## License
66
+
67
+ MIT