junfukuda's picture
Upload LoRA adapter - exp006 (structured_data_with_cot_dataset_512_v2)
09d8f59 verified
metadata
base_model: Qwen/Qwen3-4B-Instruct-2507
datasets:
  - u-10bei/structured_data_with_cot_dataset_512_v2
language:
  - en
  - ja
license: apache-2.0
library_name: peft
pipeline_tag: text-generation
tags:
  - qlora
  - lora
  - structured-output
  - structeval

Qwen3-4B StructEval exp006 - structured_data_with_cot_dataset_512_v2

This repository provides a LoRA adapter fine-tuned from Qwen/Qwen3-4B-Instruct-2507 using QLoRA (4-bit, Unsloth).

This repository contains LoRA adapter weights only. The base model must be loaded separately.

Training Objective

This adapter is trained to improve structured output accuracy (JSON / YAML / XML / TOML / CSV).

Loss is applied only to the final assistant output, while intermediate reasoning (Chain-of-Thought) is masked.

Training Configuration

  • Experiment ID: exp006
  • Base model: Qwen/Qwen3-4B-Instruct-2507
  • Training dataset: u-10bei/structured_data_with_cot_dataset_512_v2
  • Method: QLoRA (4-bit)
  • Max sequence length: 512
  • Epochs: 2
  • Learning rate: 1e-06
  • LoRA parameters: r=8, alpha=8

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel
import torch

base = "Qwen/Qwen3-4B-Instruct-2507"
adapter = "junfukuda/qwen3-structeval-exp006-u10bei"

tokenizer = AutoTokenizer.from_pretrained(base)
model = AutoModelForCausalLM.from_pretrained(
    base,
    torch_dtype=torch.float16,
    device_map="auto",
)
model = PeftModel.from_pretrained(model, adapter)

Sources & Terms (IMPORTANT)

Training data: u-10bei/structured_data_with_cot_dataset_512_v2

Dataset License: The dataset used for training is subject to its original license terms. Please refer to the dataset repository for specific license information.

Compliance: Users must comply with both the dataset's license terms and the base model's original terms of use.

Competition Context

This model was developed as part of the StructEval competition, focusing on accurate structured output generation.