Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
marksverdhei
/
GLM-4.7-Flash-FP8
like
19
Text Generation
Transformers
Safetensors
glm4_moe_lite
fp8
quantized
glm4
Mixture of Experts
conversational
License:
mit
Model card
Files
Files and versions
xet
Community
4
Deploy
Use this model
5d2df64
GLM-4.7-Flash-FP8
32.2 GB
Ctrl+K
Ctrl+K
1 contributor
History:
22 commits
marksverdhei
Add vLLM fork link for MLA detection support
5d2df64
verified
4 months ago
.gitattributes
Safe
1.57 kB
Upload FP8 quantized GLM-4.7-Flash
4 months ago
README.md
1.83 kB
Add vLLM fork link for MLA detection support
4 months ago
chat_template.jinja
Safe
3.12 kB
Upload FP8 quantized GLM-4.7-Flash
4 months ago
config.json
Safe
1.25 kB
Upload folder using huggingface_hub
4 months ago
model.safetensors
Safe
32.2 GB
xet
Upload folder using huggingface_hub
4 months ago
tokenizer.json
Safe
20.2 MB
xet
Upload FP8 quantized GLM-4.7-Flash
4 months ago
tokenizer_config.json
Safe
7.23 kB
Upload folder using huggingface_hub
4 months ago