LoRA adapter Repo ID: Mani124124/structeval-lora Base model ID used for training: unsloth/Qwen3-4B-Instruct-2507 This repository provides a LoRA adapter fine-tuned from unsloth/Qwen3-4B-Instruct-2507. This repository contains LoRA adapter weights only. The base model must be loaded separately. ## Training Objective This adapter is trained to improve structured output accuracy (JSON / YAML / XML / TOML / CSV). ## Training Configuration Base model: unsloth/Qwen3-4B-Instruct-2507 Method: LoRA (PEFT) Max sequence length: 256 Epochs: 1 Learning rate: 5e-05 LoRA: r=16, alpha=32 ## Usage from transformers import AutoModelForCausalLM, AutoTokenizer from peft import PeftModel import torch base = "unsloth/Qwen3-4B-Instruct-2507" adapter = "Mani124124/structeval-lora" tokenizer = AutoTokenizer.from_pretrained(base, trust_remote_code=True) model = AutoModelForCausalLM.from_pretrained( base, torch_dtype=torch.float16, device_map="auto", trust_remote_code=True, ) model = PeftModel.from_pretrained(model, adapter) ## Sources & Terms (IMPORTANT) Training data: u-10bei/structured_data_with_cot_dataset_512_v5