Highly experimental prune of GLM 4.6V from 128 experts to 96. Vision works, as tested in GGUF format. I would greatly appreciate feedback.

Downloads last month
3
Safetensors
Model size
83B params
Tensor type
BF16
·
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for blascotobasco/GLM-4.6V-96E

Base model

zai-org/GLM-4.6V
Finetuned
(6)
this model
Quantizations
3 models