This is zai-org/GLM-4.7-Flash quantized with llm-compressor to FP8. The model is compatible with vLLM (tested: v0.14.0). Tested with an L4 (Google Colab).
- Developed by: The Kaitchup
- License: lfm1.0
- Downloads last month
- 90
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
馃檵
Ask for provider support
Model tree for kaitchup/GLM-4.7-Flash-FP8-Dynamic
Base model
zai-org/GLM-4.7-Flash