File size: 4,374 Bytes
4279bf7
9831a4c
 
 
4279bf7
 
 
 
 
 
9831a4c
4279bf7
 
 
 
 
a28c66f
4279bf7
5a89a69
 
4279bf7
a28c66f
4279bf7
a28c66f
 
4279bf7
 
7427d8c
 
 
 
a28c66f
4279bf7
f54f66e
4279bf7
f54f66e
 
5119bf3
 
f54f66e
5119bf3
 
445b39e
f54f66e
 
5119bf3
 
 
445b39e
392eb7c
18fde0d
a28c66f
18fde0d
a28c66f
4279bf7
 
18fde0d
4279bf7
7427a66
 
4279bf7
18fde0d
4279bf7
18fde0d
4279bf7
 
 
a28c66f
4279bf7
92d0acc
a9bcf41
 
c18c7c7
a9bcf41
a28c66f
4279bf7
18fde0d
4279bf7
 
a28c66f
4279bf7
92d0acc
f4d0149
4279bf7
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
---
license: mit
base_model:
- zai-org/AutoGLM-Phone-9B
tags:
- multimodal
- vision-language
- phone-automation
- gguf
- quantized
- agent
language:
- zh
pipeline_tag: image-text-to-text
---

# AutoGLM-Phone-9B GGUF Quantized Model Collection/AutoGLM-Phone-9B GGUF 量化模型集合

Congratulations! This is the most complete and fully usable collection of AutoGLM-Phone-9B model GGUF quantized versions you can find.🎉🎉🎉   
恭喜你!这是你能找到最完整,并且绝对可用的 AutoGLM-Phone-9B 模型 GGUF 量化版本集合。🎉🎉🎉

## Model Introduction/模型简介

Phone Agent is a mobile intelligent assistant framework built on AutoGLM, 
capable of understanding smartphone screens through multimodal perception and executing automated operations to complete tasks.   
AutoGLM-Phone-9B 是基于 GLM-4V-9B 的多模态视觉语言模型,专门针对手机自动化场景进行了优化。该模型能够理解手机屏幕截图并生成相应的操作指令。

⚠️Please note! This is a multimodal vision language model, 
so in addition to the model itself, you also need the mmproj file. Please be sure to download this file for use!   
⚠️请注意!这是多模态视觉语言模型,所以除了模型本身,你还需要mmproj文件,请务必下载这个文件一起使用!

## Available quantization versions/可用的量化版本

| Quantization Type | Size | Memory Requirement | Notes | Download Link |
|---------|---------|---------|----------|----------|
| **Q2_K** | 3.73 GB | ~4 GB | Not recommended 不推荐 | [Download](AutoGLM-Phone-9B-Q2_K.gguf) |
| **Q3_K_S** | 4.28 GB | ~5 GB | Not recommended 不推荐 | [Download](AutoGLM-Phone-9B-Q3_K_S.gguf) |
| **Q3_K_M** | 4.63 GB | ~5 GB | Lower quality 质量较低 | [Download](AutoGLM-Phone-9B-Q3_K_M.gguf) |
| **Q3_K_L** | 4.84 GB | ~6 GB | Lower quality 质量较低 | [Download](AutoGLM-Phone-9B-Q3_K_L.gguf) |
| **Q4_0** | 5.08 GB | ~6 GB | Minimum available 最低可用 | [Download](AutoGLM-Phone-9B-Q4_0.gguf) |
| **Q4_1** | 5.60 GB | ~6 GB | 	Fast, recommended 快速,推荐 | [Download](AutoGLM-Phone-9B-Q4_1.gguf) |
| **Q4_K_S** | 5.36 GB | ~6 GB | Fast, recommended 快速,推荐 | [Download](AutoGLM-Phone-9B-Q4_K_S.gguf) |
| **Q4_K_M** | 5.74 GB | ~7 GB | ⭐Most Recommended, balanced 最推荐,平衡⭐ | [Download](AutoGLM-Phone-9B-Q4_K_M.gguf) |
| **Q5_0** | 6.11 GB | ~7 GB | Not recommended 不推荐 | [Download](AutoGLM-Phone-9B-Q5_0.gguf) |
| **Q5_1** | 6.62 GB | ~8 GB | Not recommended 不推荐 | [Download](AutoGLM-Phone-9B-Q5_1.gguf) |
| **Q5_K_S** | 6.24 GB | ~7 GB | Good quality 质量不错 | [Download](AutoGLM-Phone-9B-Q5_K_S.gguf) |
| **Q5_K_M** | 6.57 GB | ~8 GB | Good quality 质量不错 | [Download](AutoGLM-Phone-9B-Q5_K_M.gguf) |
| **Q6_K** | 7.70 GB | ~9 GB | Very good quality  质量非常好 | [Download](AutoGLM-Phone-9B-Q6_K.gguf) |
| **Q8_0** | 9.31 GB | ~11 GB | ⭐Fast, best quality 快速,质量最好⭐ | [Download](AutoGLM-Phone-9B-Q8_0.gguf) |
| **F16** | 17.52 GB | ~20 GB | 16 bpw, overkill  16 bpw,过量 | [Download](AutoGLM-Phone-9B-F16.gguf) |

## Quick Start/快速开始

### Using llama.cpp/使用 llama.cpp

```bash
# Download the model and visual projector
# 下载模型和视觉投影器
wget https://huggingface.co/BlcaCola/AutoGLM-Phone-9B-GGUF.gguf/resolve/main/AutoGLM-Phone-9B-Q8_0.gguf
wget https://huggingface.co/BlcaCola/AutoGLM-Phone-9B-GGUF.gguf/resolve/main/AutoGLM-Phone-9B-mmproj.gguf

# Start Server
# 启动服务器
./llama-server -m AutoGLM-Phone-9B-Q8_0.gguf --mmproj AutoGLM-Phone-9B-mmproj.gguf --host 0.0.0.0 --port 8080
```


## Performance Comparison/性能对比

Here is a chart by ikawrakow comparing the performance levels of partially quantized models (below Q5):   
这里有一张 ikawrakow 的图表,比较了部分量化的性能水平(低于Q5):

![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png)

## Related Resources/相关资源

- **原项目/Original project**: [Open-AutoGLM](https://github.com/zai-org/Open-AutoGLM)
- **llama.cpp**: [GitHub](https://github.com/ggerganov/llama.cpp)

## License Agreement/使用许可

This model is licensed under the MIT License. Please refer to the license terms of the original model.   
本模型遵循 MIT 许可证。请查看原始模型的许可证条款。


---