πŸ€– T5-NL2SQL-Gen Specialist

This model is a fine-tuned T5-Small architecture specialized for converting Natural Language questions into precise PostgreSQL queries. It serves as the primary "Reasoning Specialist" within the NLP2SQL Autonomous Intelligence Layer.

πŸš€ Model Details

  • Architecture: T5 (Text-to-Text Transfer Transformer)
  • Specialization: PostgreSQL Query Generation
  • Training Data: Fine-tuned on SQL-specific datasets (Spider/WikiSQL) and custom schema-mapped samples.
  • Project Role: Acts as the initial SQL Generator in a Hybrid Agentic loop.

πŸ”„ Hybrid Agentic Flow

This model is designed to work in tandem with Large Language Models (like Gemini 2.0 Flash) in a structured multi-agent workflow:

  1. Local ML (T5): Generates the initial high-speed SQL draft.
  2. Gemini Auditor: Validates the draft against the actual schema, adds double-quotes, and fixes hallucinations.
  3. Self-Healing Loop: If execution fails, the agents use this model's logic to refine the plan.

πŸ”— Project Context

This model is the engine for the NLP2SQL Platform.

πŸ›  Usage (Hugging Face Transformers)

from transformers import T5Tokenizer, T5ForConditionalGeneration

model_name = "Karan6124/t5-nl2sql-gen"
tokenizer = T5Tokenizer.from_pretrained(model_name)
model = T5ForConditionalGeneration.from_pretrained(model_name)

input_text = "translate English to SQL: How many users signed up in the last 30 days? \n Context: Table users (id, username, created_at)"
inputs = tokenizer(input_text, return_tensors="pt")
outputs = model.generate(**inputs, max_length=512)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Downloads last month
-
Safetensors
Model size
60.5M params
Tensor type
F32
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Dataset used to train Karan6124/t5-nl2sql-gen