Uploaded model

  • Developed by: finnianx
  • License: apache-2.0
  • Finetuned from model : unsloth/Qwen3-4B-Instruct-2507-unsloth-bnb-4bit

This qwen3 model was trained 2x faster with Unsloth and Huggingface's TRL library.

A version of Qwen3 4b Instruct finetuned on the ytz20/LMSYS-Chat-GPT-5-Chat-Response dataset. Mimics the behavior and response style of Chatgpt 5. Trained on responses only to increase accuracy.

Training Parameters

  • Lora Rank: r = 32
  • Lora Alpha: lora_alpha = 32
  • Learning rate: learning_rate = 2e-4
  • Training epochs: num_train_epochs = 1
Downloads last month
226
GGUF
Model size
4B params
Architecture
qwen3
Hardware compatibility
Log In to view the estimation

1-bit

2-bit

4-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for finnianx/GPT-5-Qwen-3-4b-Instruct-Distill

Dataset used to train finnianx/GPT-5-Qwen-3-4b-Instruct-Distill