|
|
--- |
|
|
base_model: Qwen/Qwen3-4B-Instruct-2507 |
|
|
datasets: |
|
|
- u-10bei/structured_data_with_cot_dataset_512_v2 |
|
|
language: |
|
|
- en |
|
|
- ja |
|
|
license: apache-2.0 |
|
|
library_name: peft |
|
|
pipeline_tag: text-generation |
|
|
tags: |
|
|
- qlora |
|
|
- lora |
|
|
- structured-output |
|
|
- structeval |
|
|
--- |
|
|
|
|
|
# Qwen3-4B StructEval exp006 - structured_data_with_cot_dataset_512_v2 |
|
|
|
|
|
This repository provides a **LoRA adapter** fine-tuned from |
|
|
**Qwen/Qwen3-4B-Instruct-2507** using **QLoRA (4-bit, Unsloth)**. |
|
|
|
|
|
**This repository contains LoRA adapter weights only**. |
|
|
The base model must be loaded separately. |
|
|
|
|
|
## Training Objective |
|
|
|
|
|
This adapter is trained to improve **structured output accuracy** |
|
|
(JSON / YAML / XML / TOML / CSV). |
|
|
|
|
|
Loss is applied only to the final assistant output, |
|
|
while intermediate reasoning (Chain-of-Thought) is masked. |
|
|
|
|
|
## Training Configuration |
|
|
|
|
|
- **Experiment ID**: exp006 |
|
|
- **Base model**: Qwen/Qwen3-4B-Instruct-2507 |
|
|
- **Training dataset**: u-10bei/structured_data_with_cot_dataset_512_v2 |
|
|
- **Method**: QLoRA (4-bit) |
|
|
- **Max sequence length**: 512 |
|
|
- **Epochs**: 2 |
|
|
- **Learning rate**: 1e-06 |
|
|
- **LoRA parameters**: r=8, alpha=8 |
|
|
|
|
|
## Usage |
|
|
|
|
|
```python |
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
from peft import PeftModel |
|
|
import torch |
|
|
|
|
|
base = "Qwen/Qwen3-4B-Instruct-2507" |
|
|
adapter = "junfukuda/qwen3-structeval-exp006-u10bei" |
|
|
|
|
|
tokenizer = AutoTokenizer.from_pretrained(base) |
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
|
base, |
|
|
torch_dtype=torch.float16, |
|
|
device_map="auto", |
|
|
) |
|
|
model = PeftModel.from_pretrained(model, adapter) |
|
|
``` |
|
|
|
|
|
## Sources & Terms (IMPORTANT) |
|
|
|
|
|
**Training data**: u-10bei/structured_data_with_cot_dataset_512_v2 |
|
|
|
|
|
**Dataset License**: The dataset used for training is subject to its original license terms. |
|
|
Please refer to the dataset repository for specific license information. |
|
|
|
|
|
**Compliance**: Users must comply with both the dataset's license terms and the base model's original terms of use. |
|
|
|
|
|
## Competition Context |
|
|
|
|
|
This model was developed as part of the StructEval competition, focusing on accurate structured output generation. |
|
|
|