GPT-2 Medium โ€” SQL Query Generator

Fine-tuned GPT-2 Medium on the Spider text-to-SQL dataset to generate SQL queries from natural language questions.

Training

  • Base model: GPT-2 Medium (354M parameters)
  • Dataset: Spider (7000 train / 1034 validation examples)
  • Method: Full fine-tuning
  • Best checkpoint: Epoch 1 (val loss 1.410)

Usage

from transformers import GPT2LMHeadModel, GPT2Tokenizer

model = GPT2LMHeadModel.from_pretrained("your-username/gpt2-medium-sql-generator")
tokenizer = GPT2Tokenizer.from_pretrained("your-username/gpt2-medium-sql-generator")

prompt = "Question: How many singers are there?\nSQL:"
inputs = tokenizer(prompt, return_tensors="pt")
output = model.generate(**inputs, max_new_tokens=64)
print(tokenizer.decode(output[0], skip_special_tokens=True))

Limitations

  • GPT-2 is a small model โ€” output SQL may hallucinate table/column names
  • No schema awareness โ€” works best on Singer/Concert domain from Spider training data
  • Intended as a learning project demonstrating full fine-tuning pipeline
Downloads last month
-
Safetensors
Model size
0.4B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Dataset used to train poseidon1113/gpt2-medium-sql-generator