| --- |
| base_model: Qwen/Qwen3-4B-Instruct-2507 |
| datasets: |
| - u-10bei/structured_data_with_cot_dataset_512_v5 |
| language: |
| - en |
| license: apache-2.0 |
| library_name: transformers |
| pipeline_tag: text-generation |
| tags: |
| - full-finetune |
| - structured-output |
| --- |
| |
| qwen3-4b-structured-output-lora |
|
|
| This repository provides a **full fine-tuned model** based on |
| **Qwen/Qwen3-4B-Instruct-2507** using **BF16 full fine-tuning + NEFTune**. |
|
|
| This repository contains the **complete model weights**. |
|
|
| ## Training Objective |
|
|
| This model is trained to improve **structured output accuracy** |
| (JSON / YAML / XML / TOML / CSV). |
|
|
| CoT (Chain-of-Thought) is removed from training data during preprocessing. |
|
|
| ## Training Configuration |
|
|
| - Base model: Qwen/Qwen3-4B-Instruct-2507 |
| - Method: Full fine-tuning (BF16) |
| - Max sequence length: 2048 |
| - Epochs: 1 |
| - Learning rate: 2e-05 |
| - NEFTune noise alpha: 5.0 |
|
|
| ## Usage |
|
|
| ```python |
| from transformers import AutoModelForCausalLM, AutoTokenizer |
| import torch |
| |
| model_id = "84basi/lora-10-2" |
| |
| tokenizer = AutoTokenizer.from_pretrained(model_id) |
| model = AutoModelForCausalLM.from_pretrained( |
| model_id, |
| torch_dtype=torch.bfloat16, |
| device_map="auto", |
| ) |
| ``` |
|
|
| ## Sources & Terms (IMPORTANT) |
|
|
| Training data: u-10bei/structured_data_with_cot_dataset_512_v5 |
|
|
| Dataset License: MIT License. This dataset is used and distributed under the terms of the MIT License. |
| Compliance: Users must comply with the MIT license (including copyright notice) and the base model's original terms of use. |
|
|