| | --- |
| | base_model: Qwen/Qwen2.5-3B-Instruct |
| | library_name: peft |
| | license: mit |
| | tags: |
| | - lora |
| | - json |
| | - structured-output |
| | - text-generation |
| | - qwen2 |
| | - sft |
| | - trl |
| | language: |
| | - en |
| | pipeline_tag: text-generation |
| | --- |
| | |
| | # Qwen2.5-3B JSON Structured Output (LoRA) |
| |
|
| | A LoRA adapter that enhances Qwen2.5-3B-Instruct for extracting structured JSON from unstructured text. |
| |
|
| | ## Training Details |
| |
|
| | | Parameter | Value | |
| | |-----------|-------| |
| | | Base Model | Qwen/Qwen2.5-3B-Instruct | |
| | | Method | QLoRA (4-bit NF4) | |
| | | LoRA Rank | 16 | |
| | | LoRA Alpha | 32 | |
| | | Target Modules | q,k,v,o,gate,up,down | |
| | | Trainable Params | 29.9M (1.76%) | |
| | | Training Examples | 20 curated | |
| | | Epochs | 5 | |
| | | Learning Rate | 5e-5 | |
| | | Final Loss | 1.506 | |
| | | Hardware | NVIDIA RTX 3090 x2 | |
| |
|
| | ## Quick Start |
| |
|
| | ```python |
| | from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig |
| | from peft import PeftModel |
| | import torch, json |
| | |
| | bnb = BitsAndBytesConfig(load_in_4bit=True, bnb_4bit_quant_type="nf4", bnb_4bit_compute_dtype=torch.bfloat16) |
| | tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-3B-Instruct") |
| | model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen2.5-3B-Instruct", quantization_config=bnb, device_map="auto") |
| | model = PeftModel.from_pretrained(model, "2reb/Qwen2.5-3B-JSON-StructuredOutput") |
| | ``` |
| |
|
| | ## License |
| |
|
| | MIT |
| |
|