File size: 1,680 Bytes
d8bdce3 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 |
---
license: apache-2.0
base_model: unsloth/gemma-3-27b-it-bnb-4bit
tags:
- text-to-sql
- gemma-3
- unsloth
- trl
- finance
- real-estate
- nlp
datasets:
- dan-text2sql/seoul-realestate-sql-v1
library_name: transformers
---
# seoul-realestate-sql-agent-v2
**Developed by:** dan-text2sql
**License:** apache-2.0
**Finetuned from model:** [unsloth/gemma-3-27b-it-bnb-4bit](https://huggingface.co/unsloth/gemma-3-27b-it-bnb-4bit)
This model is a **Text-to-SQL agent** specialized in **Korean Real Estate (Seoul)** data.
It was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
## Model Description (v2)
This is the **v2** version of the Seoul Real Estate SQL Agent.
* **Base Model:** Gemma-3 27B (IT)
* **Improvement:** Unlike v1 (Mistral-7B), this model leverages the massive 27B parameter size of Gemma-3.
* **Objective:** Translate natural language queries about Seoul apartment real estate data into executable SQL queries.
## Usage Example
```python
from unsloth import FastLanguageModel
# Load the model
model, tokenizer = FastLanguageModel.from_pretrained(
model_name = "dan-text2sql/seoul-realestate-sql-agent-v2",
max_seq_length = 2048,
dtype = None,
load_in_4bit = True,
)
FastLanguageModel.for_inference(model)
# Test Prompt
prompt = """์๋ ์ง๋ฌธ์ ๋ํ ์ฌ๋ฐ๋ฅธ SQL ์ฟผ๋ฆฌ๋ฅผ ์์ฑํด์ฃผ์ธ์.
### ์ง๋ฌธ:
์์ธ์ ๊ฐ๋จ๊ตฌ ์ผ์ฑ๋์ 20์ต ์ดํ ์ํํธ ๋งค๋ฌผ์ ์ฐพ์์ค.
### SQL:
"""
inputs = tokenizer([prompt], return_tensors = "pt").to("cuda")
outputs = model.generate(**inputs, max_new_tokens = 128, use_cache = True)
print(tokenizer.batch_decode(outputs)[0])
|