Model Overview:

A distilled 4B parameter instruction model, designed to be a lightweight and efficient conversationalist.

Purpose:

Follows the style and formatting of GLM4.6, providing much more concise answers than the base Qwen model.

Base Architecture:

Qwen/Qwen3-4B-Instruct-2507

Distillation Source:

The capabilities and response style are derived from GLM-4.6.

Training Dataset:

Fine-tuned on finnianx/glm4.6-2.16k.

Edit:

I dont like this model that much anymore, please try any other model from me or someone else, dont use this!!!

Downloads last month
1,040
GGUF
Model size
4B params
Architecture
qwen3
Hardware compatibility
Log In to view the estimation

1-bit

2-bit

4-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for finnianx/GLM4.6-Qwen3-4b-Instruct-Distill

Quantized
(168)
this model