| LoRA adapter Repo ID: Mani124124/structeval-lora | |
| Base model ID used for training: unsloth/Qwen3-4B-Instruct-2507 | |
| This repository provides a LoRA adapter fine-tuned from unsloth/Qwen3-4B-Instruct-2507. | |
| This repository contains LoRA adapter weights only. The base model must be loaded separately. | |
| ## Training Objective | |
| This adapter is trained to improve structured output accuracy (JSON / YAML / XML / TOML / CSV). | |
| ## Training Configuration | |
| Base model: unsloth/Qwen3-4B-Instruct-2507 | |
| Method: LoRA (PEFT) | |
| Max sequence length: 256 | |
| Epochs: 1 | |
| Learning rate: 5e-05 | |
| LoRA: r=16, alpha=32 | |
| ## Usage | |
| from transformers import AutoModelForCausalLM, AutoTokenizer | |
| from peft import PeftModel | |
| import torch | |
| base = "unsloth/Qwen3-4B-Instruct-2507" | |
| adapter = "Mani124124/structeval-lora" | |
| tokenizer = AutoTokenizer.from_pretrained(base, trust_remote_code=True) | |
| model = AutoModelForCausalLM.from_pretrained( | |
| base, | |
| torch_dtype=torch.float16, | |
| device_map="auto", | |
| trust_remote_code=True, | |
| ) | |
| model = PeftModel.from_pretrained(model, adapter) | |
| ## Sources & Terms (IMPORTANT) | |
| Training data: u-10bei/structured_data_with_cot_dataset_512_v5 | |