Unsprawl-4B-SFT

Fine-tuned LoRA adapter for the Unsprawl platform's local inference pipeline.

Training

  • Base model: Qwen3-4B-Instruct (4-bit quantized via Unsloth)
  • Method: QLoRA (r=16, alpha=32, 0.81% trainable params)
  • Data: 195 curated pairs (entity extraction, strategy generation, quality judging, tone classification)
  • Training: 1 epoch, batch size 8, AdamW 8-bit, lr=2e-4
  • Final loss: 1.69 (from 2.65 starting)
  • Hardware: NVIDIA RTX 3060 12GB, 164 seconds

Tasks

Task Description
Entity extraction Structured JSON entity extraction from legal documents
Quality judging 5-dimension scoring with APPROVE/REJECT verdicts
Tone classification 6-category tone classification (NEUTRAL, INFLAMMATORY, etc.)
Claim verification Factual claim extraction for KG cross-reference

Usage

from unsloth import FastLanguageModel

model, tokenizer = FastLanguageModel.from_pretrained(
    model_name="Mansib/Unsprawl-4B-SFT",
    max_seq_length=4096,
    load_in_4bit=True,
)
FastLanguageModel.for_inference(model)

Part of Unsprawl

Unsprawl is a domain-agnostic, mission-driven compound AI platform for autonomous infrastructure resilience. This adapter enables local inference for structured tasks, reducing API dependency on cloud LLMs.

Downloads last month
18
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Mansib/Unsprawl-4B-SFT

Adapter
(3)
this model