VLLM example inference server usage:

pip install vllm>=0.17.0
pip install huggingface-hub>=1.6.0 transformers>=5.3.0

vllm serve hhzm/qwen3.5-4b-meow --reasoning-parser qwen3 --enable-auto-tool-choice --tool-call-parser qwen3_coder
Downloads last month
3
Safetensors
Model size
5B params
Tensor type
F16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for hhzm/qwen3.5-4b-meow

Finetuned
Qwen/Qwen3.5-4B
Finetuned
(1)
this model
Finetunes
1 model

Dataset used to train hhzm/qwen3.5-4b-meow