AutoGLM-Phone-9B GGUF Quantized Model Collection/AutoGLM-Phone-9B GGUF 量化模型集合

Congratulations! This is the most complete and fully usable collection of AutoGLM-Phone-9B model GGUF quantized versions you can find.🎉🎉🎉
恭喜你!这是你能找到最完整,并且绝对可用的 AutoGLM-Phone-9B 模型 GGUF 量化版本集合。🎉🎉🎉

Model Introduction/模型简介

Phone Agent is a mobile intelligent assistant framework built on AutoGLM, capable of understanding smartphone screens through multimodal perception and executing automated operations to complete tasks.
AutoGLM-Phone-9B 是基于 GLM-4V-9B 的多模态视觉语言模型,专门针对手机自动化场景进行了优化。该模型能够理解手机屏幕截图并生成相应的操作指令。

⚠️Please note! This is a multimodal vision language model, so in addition to the model itself, you also need the mmproj file. Please be sure to download this file for use!
⚠️请注意!这是多模态视觉语言模型,所以除了模型本身,你还需要mmproj文件,请务必下载这个文件一起使用!

Available quantization versions/可用的量化版本

Quantization Type Size Memory Requirement Notes Download Link
Q2_K 3.73 GB ~4 GB Not recommended 不推荐 Download
Q3_K_S 4.28 GB ~5 GB Not recommended 不推荐 Download
Q3_K_M 4.63 GB ~5 GB Lower quality 质量较低 Download
Q3_K_L 4.84 GB ~6 GB Lower quality 质量较低 Download
Q4_0 5.08 GB ~6 GB Minimum available 最低可用 Download
Q4_1 5.60 GB ~6 GB Fast, recommended 快速,推荐 Download
Q4_K_S 5.36 GB ~6 GB Fast, recommended 快速,推荐 Download
Q4_K_M 5.74 GB ~7 GB ⭐Most Recommended, balanced 最推荐,平衡⭐ Download
Q5_0 6.11 GB ~7 GB Not recommended 不推荐 Download
Q5_1 6.62 GB ~8 GB Not recommended 不推荐 Download
Q5_K_S 6.24 GB ~7 GB Good quality 质量不错 Download
Q5_K_M 6.57 GB ~8 GB Good quality 质量不错 Download
Q6_K 7.70 GB ~9 GB Very good quality 质量非常好 Download
Q8_0 9.31 GB ~11 GB ⭐Fast, best quality 快速,质量最好⭐ Download
F16 17.52 GB ~20 GB 16 bpw, overkill 16 bpw,过量 Download

Quick Start/快速开始

Using llama.cpp/使用 llama.cpp

# Download the model and visual projector
# 下载模型和视觉投影器
wget https://huggingface.co/BlcaCola/AutoGLM-Phone-9B-GGUF.gguf/resolve/main/AutoGLM-Phone-9B-Q8_0.gguf
wget https://huggingface.co/BlcaCola/AutoGLM-Phone-9B-GGUF.gguf/resolve/main/AutoGLM-Phone-9B-mmproj.gguf

# Start Server
# 启动服务器
./llama-server -m AutoGLM-Phone-9B-Q8_0.gguf --mmproj AutoGLM-Phone-9B-mmproj.gguf --host 0.0.0.0 --port 8080

Performance Comparison/性能对比

Here is a chart by ikawrakow comparing the performance levels of partially quantized models (below Q5):
这里有一张 ikawrakow 的图表,比较了部分量化的性能水平(低于Q5):

image.png

Related Resources/相关资源

License Agreement/使用许可

This model is licensed under the MIT License. Please refer to the license terms of the original model.
本模型遵循 MIT 许可证。请查看原始模型的许可证条款。


Downloads last month
1,203
GGUF
Model size
9B params
Architecture
glm4
Hardware compatibility
Log In to view the estimation

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for BlcaCola/AutoGLM-Phone-9B-GGUF

Quantized
(8)
this model