phi3-text2sql-lora / README.md
mrcmilo's picture
Update README.md
66d2627 verified
|
raw
history blame
1.95 kB
---
license: mit
base_model: microsoft/Phi-3-mini-4k-instruct
tags:
- text-to-sql
- natural-language-to-sql
- phi-3
- unsloth
- gguf
- ollama
- logic
datasets:
- b-mc2/sql-create-context
language:
- en
library_name: unsloth
pipeline_tag: text-generation
---
# Phi-3-SQL-Expert (GGUF)
This is a specialized **Text-to-SQL** model fine-tuned from the **Microsoft Phi-3-mini-4k-instruct** architecture. It has been optimized using **Unsloth** to provide high-accuracy SQL generation while remaining lightweight enough to run on consumer hardware.
## Key Features
- **Architecture:** Phi-3-mini (3.8B parameters)
- **Quantization:** Q4_K_M GGUF (Optimized balance of speed and logic)
- **Training Technique:** Fine-tuned using Lora with [Unsloth](https://github.com/unslothai/unsloth).
- **Format:** GGUF (Ready for Ollama, LM Studio, and llama.cpp)
## Usage Instructions
### Ollama (Recommended)
To deploy locally:
1. Download the `.gguf` file.
2. Create a file named `Modelfile` and paste the following:
```dockerfile
FROM ./phi3-sql-expert.Q4_K_M.gguf
TEMPLATE """<|system|>
You are a helpful assistant that writes SQL queries. Given a user question and a table schema, output only the SQL code.<|end|>
<|user|>
{{ .Prompt }}<|end|>
<|assistant|>
"""
PARAMETER stop "<|end|>"
PARAMETER temperature 0.1
PARAMETER num_ctx 2048
```
3. Run ollama create sql-expert -f Modelfile
4. Run ollama run sql-expert
## Evaluation Data
The model was fine-tuned on the sql-create-context dataset, focusing on:
Mapping natural language to complex SELECT, WHERE, and JOIN statements.
Understanding table schemas provided in the prompt.
Maintaining strict SQL syntax.
## Recommended Settings
Temperature: 0.0 or 0.1 (SQL requires deterministic output).
Stop Tokens: Ensure <|end|> is set as a stop sequence to prevent "infinite looping" generation.
Context Window: 2048 or 4096 tokens.
Model Developer: mrcmilo
Base Model: Phi-3-mini-4k-instruct