| | --- |
| | license: mit |
| | base_model: zai-org/GLM-4.1V-9B-Thinking |
| | base_model_relation: quantized |
| | quantized_by: turboderp |
| | tags: |
| | - exl3 |
| | --- |
| | |
| | EXL3 quants of [GLM-4.1V-9B-Thinking](https://huggingface.co/zai-org/GLM-4.1V-9B-Thinking) |
| |
|
| | ⚠️ Requires ExLlamaV3 v0.0.15 (or v0.0.14 `dev` branch) |
| |
|
| | [2.00 bits per weight](https://huggingface.co/turboderp/GLM-4.1V-9B-Thinking-exl3/tree/2.00bpw) |
| | [2.50 bits per weight](https://huggingface.co/turboderp/GLM-4.1V-9B-Thinking-exl3/tree/2.50bpw) |
| | [3.00 bits per weight](https://huggingface.co/turboderp/GLM-4.1V-9B-Thinking-exl3/tree/3.00bpw) |
| | [3.50 bits per weight](https://huggingface.co/turboderp/GLM-4.1V-9B-Thinking-exl3/tree/3.50bpw) |
| | [4.00 bits per weight](https://huggingface.co/turboderp/GLM-4.1V-9B-Thinking-exl3/tree/4.00bpw) |
| | [5.00 bits per weight](https://huggingface.co/turboderp/GLM-4.1V-9B-Thinking-exl3/tree/5.00bpw) |
| | [6.00 bits per weight](https://huggingface.co/turboderp/GLM-4.1V-9B-Thinking-exl3/tree/6.00bpw) |
| |
|