structeval-lora / README.md
Mani124124's picture
Update README.md
b862633 verified

LoRA adapter Repo ID: Mani124124/structeval-lora Base model ID used for training: unsloth/Qwen3-4B-Instruct-2507

This repository provides a LoRA adapter fine-tuned from unsloth/Qwen3-4B-Instruct-2507.

This repository contains LoRA adapter weights only. The base model must be loaded separately.

Training Objective

This adapter is trained to improve structured output accuracy (JSON / YAML / XML / TOML / CSV).

Training Configuration

Base model: unsloth/Qwen3-4B-Instruct-2507 Method: LoRA (PEFT) Max sequence length: 256 Epochs: 1 Learning rate: 5e-05 LoRA: r=16, alpha=32

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer from peft import PeftModel import torch

base = "unsloth/Qwen3-4B-Instruct-2507" adapter = "Mani124124/structeval-lora"

tokenizer = AutoTokenizer.from_pretrained(base, trust_remote_code=True) model = AutoModelForCausalLM.from_pretrained( base, torch_dtype=torch.float16, device_map="auto", trust_remote_code=True, ) model = PeftModel.from_pretrained(model, adapter)

Sources & Terms (IMPORTANT)

Training data: u-10bei/structured_data_with_cot_dataset_512_v5