--- base_model: Qwen/Qwen3-4B-Instruct-2507 datasets: - u-10bei/dpo-dataset-qwen-cot language: - en license: apache-2.0 library_name: transformers pipeline_tag: text-generation tags: - dpo - unsloth - qwen - alignment --- # jm03 This model is a fine-tuned version of **Qwen/Qwen3-4B-Instruct-2507** using **Direct Preference Optimization (DPO)** via the **Unsloth** library. This repository contains the **full-merged 16-bit weights**. No adapter loading is required. ## Training Objective This model has been optimized using DPO to align its responses with preferred outputs, focusing on improving reasoning (Chain-of-Thought) and structured response quality based on the provided preference dataset. ## Training Configuration - **Base model**: Qwen/Qwen3-4B-Instruct-2507 - **Method**: DPO (Direct Preference Optimization) - **Epochs**: 1 - **Learning rate**: 1e-07 - **Beta**: 0.1 - **Max sequence length**: 1024 - **LoRA Config**: r=8, alpha=16 (merged into base) ## Usage Since this is a merged model, you can use it directly with `transformers`. ```python from transformers import AutoModelForCausalLM, AutoTokenizer import torch model_id = "your_id/your-repo-name" tokenizer = AutoTokenizer.from_pretrained(model_id) model = AutoModelForCausalLM.from_pretrained( model_id, torch_dtype=torch.float16, device_map="auto" ) # Test inference prompt = "Your question here" inputs = tokenizer( prompt, return_tensors="pt" ).to("cuda") outputs = model.generate( **inputs, max_new_tokens=512 ) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) outputs = model.generate(**inputs, max_new_tokens=512) print(tokenizer.decode(outputs[0])) ``` ## Sources & License (IMPORTANT) * **Training Data**: [u-10bei/dpo-dataset-qwen-cot] * **License**: MIT License. (As per dataset terms). * **Compliance**: Users must follow the original base model's license terms.