|
|
--- |
|
|
license: mit |
|
|
base_model: microsoft/Phi-3-mini-4k-instruct |
|
|
tags: |
|
|
- text-to-sql |
|
|
- natural-language-to-sql |
|
|
- phi-3 |
|
|
- unsloth |
|
|
- gguf |
|
|
- ollama |
|
|
- logic |
|
|
datasets: |
|
|
- b-mc2/sql-create-context |
|
|
language: |
|
|
- en |
|
|
library_name: unsloth |
|
|
pipeline_tag: text-generation |
|
|
--- |
|
|
|
|
|
# Phi-3-SQL-Expert (GGUF) |
|
|
|
|
|
This is a specialized **Text-to-SQL** model fine-tuned from the **Microsoft Phi-3-mini-4k-instruct** architecture. It has been optimized using **Unsloth** to provide high-accuracy SQL generation while remaining lightweight enough to run on consumer hardware. |
|
|
|
|
|
## Key Features |
|
|
- **Architecture:** Phi-3-mini (3.8B parameters) |
|
|
- **Quantization:** Q4_K_M GGUF (Optimized balance of speed and logic) |
|
|
- **Training Technique:** Fine-tuned using Lora with [Unsloth](https://github.com/unslothai/unsloth). |
|
|
- **Format:** GGUF (Ready for Ollama, LM Studio, and llama.cpp) |
|
|
|
|
|
## Usage Instructions |
|
|
|
|
|
### Ollama (Recommended) |
|
|
To deploy locally: |
|
|
|
|
|
1. Download the `.gguf` file. |
|
|
2. Create a file named `Modelfile` and paste the following: |
|
|
```dockerfile |
|
|
FROM ./phi3-sql-expert.Q4_K_M.gguf |
|
|
|
|
|
TEMPLATE """<|system|> |
|
|
You are a helpful assistant that writes SQL queries. Given a user question and a table schema, output only the SQL code.<|end|> |
|
|
<|user|> |
|
|
{{ .Prompt }}<|end|> |
|
|
<|assistant|> |
|
|
""" |
|
|
|
|
|
PARAMETER stop "<|end|>" |
|
|
PARAMETER temperature 0.1 |
|
|
PARAMETER num_ctx 2048 |
|
|
``` |
|
|
|
|
|
3. Run ollama create sql-expert -f Modelfile |
|
|
|
|
|
4. Run ollama run sql-expert |
|
|
|
|
|
|
|
|
## Evaluation Data |
|
|
The model was fine-tuned on the sql-create-context dataset, focusing on: |
|
|
|
|
|
Mapping natural language to complex SELECT, WHERE, and JOIN statements. |
|
|
|
|
|
Understanding table schemas provided in the prompt. |
|
|
|
|
|
Maintaining strict SQL syntax. |
|
|
|
|
|
## Recommended Settings |
|
|
Temperature: 0.0 or 0.1 (SQL requires deterministic output). |
|
|
|
|
|
Stop Tokens: Ensure <|end|> is set as a stop sequence to prevent "infinite looping" generation. |
|
|
|
|
|
Context Window: 2048 or 4096 tokens. |
|
|
|
|
|
Model Developer: mrcmilo |
|
|
|
|
|
Base Model: Phi-3-mini-4k-instruct |