File size: 3,446 Bytes
8f8b4ac 596ae95 8f8b4ac 596ae95 8f8b4ac ca53f1c d29a032 46c8420 d29a032 46c8420 d29a032 ca53f1c d29a032 ca53f1c d29a032 46c8420 d29a032 46c8420 d29a032 46c8420 d29a032 46c8420 d29a032 46c8420 d29a032 fc7f696 d29a032 46c8420 abc576f 1ed5057 d29a032 46c8420 1ed5057 d29a032 46c8420 fc7f696 46c8420 c21a390 46c8420 c21a390 46c8420 d29a032 ca53f1c 46c8420 ca53f1c 596ae95 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 | ---
license: apache-2.0
library_name: transformers
base_model: Qwen/Qwen2-0.5B-Instruct
tags:
- generated_from_trainer
- trl
- sft
- sql
- text
model_name: Qwen2-0.5B-Instruct-SQL-generator
datasets:
- gretelai/synthetic_text_to_sql
language:
- en
metrics:
- bleu
- chrf
- rouge
---
# Model Card for Qwen2-0.5B-Instruct-SQL-generator
This model is a fine-tuned version of [Qwen/Qwen2-0.5B-Instruct](https://huggingface.co/Qwen/Qwen2-0.5B-Instruct).
It has been trained using [TRL (Transformer Reinforcement Learning)](https://github.com/huggingface/trl) for SQL generation tasks.
## Quick Start
```python
from transformers import pipeline
question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="onkolahmet/Qwen2-0.5B-Instruct-SQL-generator", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])
```
## Training Procedure
This model was trained with Supervised Fine-Tuning (SFT) using the [gretelai/synthetic_text_to_sql](https://huggingface.co/datasets/gretelai/synthetic_text_to_sql) dataset.
The goal was to fine-tune the model to better translate natural language queries into SQL statements.
### Framework Versions
- TRL: 0.12.2
- Transformers: 4.46.3
- PyTorch: 2.6.0
- Datasets: 3.4.1
- Tokenizers: 0.20.3
## Evaluation Results
The model was evaluated using standard text generation metrics (BLEU, ROUGE-L F1, CHRF) in both zero-shot and few-shot prompting scenarios.
### 🔹 Zero-shot Prompting (on `gretelai/synthetic_text_to_sql/test`)
**After Post-processing:**
- **BLEU Score:** 0.5195
- **ROUGE-L F1:** 0.7031
- **CHRF Score:** 70.0409
**Before Post-processing:**
- **BLEU Score:** 0.1452
- **ROUGE-L F1:** 0.3009
- **CHRF Score:** 47.8182
**SQL-Specific Metrics:**
- **Exact Match (case insensitive):** 0.1600
- **Normalized Exact Match:** 0.1500
- **Average Component Match:** 0.4528
- **Average Entity Match:** 0.8807
**Query Quality Distribution:**
- **High Quality (≥80% component match):** 18 (18.0%)
- **Medium Quality (50-79% component match):** 28 (28.0%)
- **Low Quality (<50% component match):** 54 (54.0%)
---
### 🔹 Few-shot Prompting (on `gretelai/synthetic_text_to_sql/test`)
**After Post-processing:**
- **BLEU Score:** 0.2680
- **ROUGE-L F1:** 0.4975
- **CHRF Score:** 57.1704
**Before Post-processing:**
- **BLEU Score:** 0.1272
- **ROUGE-L F1:** 0.2816
- **CHRF Score:** 46.1643
**SQL-Specific Metrics:**
- **Exact Match (case insensitive):** 0.0000
- **Normalized Exact Match:** 0.0000
- **Average Component Match:** 0.2140
- **Average Entity Match:** 0.8067
**Query Quality Distribution:**
- **High Quality (≥80% component match):** 4 (4.0%)
- **Medium Quality (50-79% component match):** 17 (17.0%)
- **Low Quality (<50% component match):** 79 (79.0%)
## Citations
Cite TRL as:
```bibtex
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
``` |