File size: 979 Bytes
7705459
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
---
language:
- en
library_name: llama.cpp
pipeline_tag: text-generation
license: apache-2.0
quantization: q4_k_m
---

# LLaMa_3.2_3B_Instruct_Text2SQL-Q4_K_M-GGUF.gguf

This is a GGUF quantized version of the LLaMa 3.2 3B Text2SQL model.

## Model Details
- **Architecture:** LLaMa 3.2 3B
- **Task:** Text to SQL Generation
- **Quantization:** Q4_K_M
- **Context Length:** 65536 tokens (2^16)
- **Format:** GGUF (Compatible with llama.cpp)

## Usage

```python
from llama_cpp import Llama

# Initialize model
llm = Llama(
    model_path="downloaded_model.gguf",
    n_ctx=65536,  # 64K context
    n_threads=8   # Adjust based on your CPU
)

# Generate SQL
response = llm(
    "Convert this to SQL: Find all users who signed up in January 2024",
    max_tokens=1024,
    temperature=0.7
)

print(response['choices'][0]['text'])
```

## Model Source
This is a quantized version of [XeAI/LLaMa_3.2_3B_Instruct_Text2SQL](https://huggingface.co/XeAI/LLaMa_3.2_3B_Instruct_Text2SQL)