SQLGPT / README.md
Awais009's picture
Update README.md
5d4f704 verified
---
title: SQLGPT
sdk: docker
emoji: ๐Ÿš€
colorFrom: yellow
colorTo: indigo
short_description: Give the context of the table and ask question from model.
---
# SQLGPT
SQLGPT is a powerful model designed to generate SQL queries based on your table information and specific questions. Simply provide the context of your table, ask a question, and SQLGPT will generate the corresponding SQL query for you.
### Live
You can interact with it live here: `https://sqlgpt-hazel.vercel.app/`
But as its deployed on huggingface spaces with 1 thread available and its running on CPU so be patient ;) it can take time (secret!! it can take upto 1 min)
## Features
- **SQL Query Generation:** Input table details and your query; the model generates the appropriate SQL command.
- **Fine-Tunning:** The model is fine-tuned on Google's Gemma 2b using the dataset available [here](https://huggingface.co/datasets/b-mc2/sql-create-context) on Hugging Face.
- **Model Availability:** The model is available on both Kaggle and Hugging Face.
- **Quantization:** The finetunned model is being quantized to 4-bit in GGUF format using llama.cpp
## Getting Started
### Running the UI Interface on Unix Distributions (Linux, macOS)
1. **Clone the Repository:**
```bash
git clone https://github.com/awaistahseen009/SQLGPT
```
2. **Install the Requirements:**
```bash
pip install -r requirements.txt
```
3. **Download the Quantized Model Setup:**
Download the quantized model from [Hugging Face](https://huggingface.co/spaces/awais009/SQLGPT/tree/main).
4. **Run the UI Interface:**
- Update the API request URL in `App.jsx`:
```javascript
// Change this line in App.jsx
const apiUrl = "http://localhost:8000/query";
```
- Start the server:
```bash
uvicorn main:app
```
5. **Launch the UI:**
Run npm run dev in ui folder's terminal and
Open the UI in your browser to interact with the model.
on `http://localhost:8000`
### Windows Users
If you're using Windows, the `llama-cpp` package is not available, so you will need to follow these steps:
1. **Clone the llama.cpp Repository:**
```bash
git clone https://github.com/ggerganov/llama.cpp
```
2. **Download the Quantized Model:**
Download the quantized model from [Hugging Face](https://huggingface.co/spaces/awais009/SQLGPT/tree/main/quantized_model).
3. **Run the Model:**
In your terminal, execute the following command:
```bash
./llama.cpp/llama-cli -m ./quantized_model/sql_gpt_quantized.gguf -n 256 -p "### QUESTION:\n{question_here}\n\n### CONTEXT:\n{context_here}\n\n### [RESPONSE]:\n"
```
4. **Prompt Template:**
Use the following prompt template when interacting with the model:
```text
### QUESTION:
{question_here}
### CONTEXT:
{context_here}
### [RESPONSE]:
```
## Fine-Tuned and Quantization Files
You can download the fine-tuned model and quantization files from the [SQLGPT Fine Tune Material Repository](https://github.com/awaistahseen009/SQLGPTFineTuneMaterial).
## Contributing
Contributions are welcome! Feel free to fork the project, make improvements, and submit a pull request.
---
Happy querying with SQLGPT!