File size: 964 Bytes
231c5f2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
---

license: mit
base_model:
- zai-org/GLM-4.7-Flash
tags:
- llm-compressor
---


<div align="center">
  <img 

    src="https://cdn-uploads.huggingface.co/production/uploads/64b93e6bd6c468ac7536607e/mj6xac74jHGLqymiovObc.png" 

    alt="The Kaitchup -- AI on a Budget" 

    style="width: 100%; max-width: 100%; height: auto; display: inline-block; margin-bottom: 0.5em; margin-top: 0.5em;"

  />
  <div style="display: flex; justify-content: center; gap: 0.5em; margin-bottom: 1em;">
    <a href="https://kaitchup.substack.com/subscribe"><strong>Subscribe and Support</strong></a>

  </div>

</div>



This is [zai-org/GLM-4.7-Flash](https://huggingface.co/zai-org/GLM-4.7-Flash) quantized with [llm-compressor](https://github.com/vllm-project/llm-compressor) to FP8. The model is compatible with vLLM (tested: v0.14.0). Tested with an L4 (Google Colab).

- **Developed by:** [The Kaitchup](https://kaitchup.substack.com/)
- **License:**  lfm1.0