structeval-lora / README.md
Mani124124's picture
Update README.md
b862633 verified
LoRA adapter Repo ID: Mani124124/structeval-lora
Base model ID used for training: unsloth/Qwen3-4B-Instruct-2507
This repository provides a LoRA adapter fine-tuned from unsloth/Qwen3-4B-Instruct-2507.
This repository contains LoRA adapter weights only. The base model must be loaded separately.
## Training Objective
This adapter is trained to improve structured output accuracy (JSON / YAML / XML / TOML / CSV).
## Training Configuration
Base model: unsloth/Qwen3-4B-Instruct-2507
Method: LoRA (PEFT)
Max sequence length: 256
Epochs: 1
Learning rate: 5e-05
LoRA: r=16, alpha=32
## Usage
from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel
import torch
base = "unsloth/Qwen3-4B-Instruct-2507"
adapter = "Mani124124/structeval-lora"
tokenizer = AutoTokenizer.from_pretrained(base, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(
base,
torch_dtype=torch.float16,
device_map="auto",
trust_remote_code=True,
)
model = PeftModel.from_pretrained(model, adapter)
## Sources & Terms (IMPORTANT)
Training data: u-10bei/structured_data_with_cot_dataset_512_v5