--- base_model: Qwen/Qwen3-4B-Instruct-2507 datasets: - u-10bei/structured_data_with_cot_dataset_512_v2 language: - en - ja license: apache-2.0 library_name: peft pipeline_tag: text-generation tags: - qlora - lora - structured-output - structeval --- # Qwen3-4B StructEval exp006 - structured_data_with_cot_dataset_512_v2 This repository provides a **LoRA adapter** fine-tuned from **Qwen/Qwen3-4B-Instruct-2507** using **QLoRA (4-bit, Unsloth)**. **This repository contains LoRA adapter weights only**. The base model must be loaded separately. ## Training Objective This adapter is trained to improve **structured output accuracy** (JSON / YAML / XML / TOML / CSV). Loss is applied only to the final assistant output, while intermediate reasoning (Chain-of-Thought) is masked. ## Training Configuration - **Experiment ID**: exp006 - **Base model**: Qwen/Qwen3-4B-Instruct-2507 - **Training dataset**: u-10bei/structured_data_with_cot_dataset_512_v2 - **Method**: QLoRA (4-bit) - **Max sequence length**: 512 - **Epochs**: 2 - **Learning rate**: 1e-06 - **LoRA parameters**: r=8, alpha=8 ## Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer from peft import PeftModel import torch base = "Qwen/Qwen3-4B-Instruct-2507" adapter = "junfukuda/qwen3-structeval-exp006-u10bei" tokenizer = AutoTokenizer.from_pretrained(base) model = AutoModelForCausalLM.from_pretrained( base, torch_dtype=torch.float16, device_map="auto", ) model = PeftModel.from_pretrained(model, adapter) ``` ## Sources & Terms (IMPORTANT) **Training data**: u-10bei/structured_data_with_cot_dataset_512_v2 **Dataset License**: The dataset used for training is subject to its original license terms. Please refer to the dataset repository for specific license information. **Compliance**: Users must comply with both the dataset's license terms and the base model's original terms of use. ## Competition Context This model was developed as part of the StructEval competition, focusing on accurate structured output generation.