flaviusburca commited on
Commit
0d8b9bd
·
verified ·
1 Parent(s): 3282744

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -3
README.md CHANGED
@@ -9,14 +9,16 @@ base_model:
9
 
10
  # Qwen3.5-9B
11
 
12
- <img width="400px" src="https://qianwen-res.oss-accelerate.aliyuncs.com/logo_qwen3.5.png">
13
 
14
- [![Qwen Chat](https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5)](https://chat.qwen.ai)
15
 
16
  > [!Note]
17
- > This repository contains model weights and configuration files for the post-trained model in the Hugging Face Transformers format.
18
  >
19
  > These artifacts are compatible with Hugging Face Transformers, vLLM, SGLang, KTransformers, etc.
 
 
20
 
21
  Over recent months, we have intensified our focus on developing foundation models that deliver exceptional utility and performance. Qwen3.5 represents a significant leap forward, integrating breakthroughs in multimodal learning, architectural efficiency, reinforcement learning scale, and global accessibility to empower developers and enterprises with unprecedented capability and efficiency.
22
 
 
9
 
10
  # Qwen3.5-9B
11
 
12
+ <img width="400px" src="https://huggingface.co/surogate/Qwen3.5-4B-FP8/resolve/main/Surogate-Logo-White.png">
13
 
14
+ **This Qwen3.5 variant is recommended for Surogate. Check out [http://surogate.ai](http://surogate.ai)**
15
 
16
  > [!Note]
17
+ > This repository contains FP8-quantized model weights and configuration files for the post-trained model in the Hugging Face Transformers format.
18
  >
19
  > These artifacts are compatible with Hugging Face Transformers, vLLM, SGLang, KTransformers, etc.
20
+ >
21
+ > The quantization method is fine-grained fp8 quantization with block size of 128, and its performance metrics are nearly identical to those of the original model.
22
 
23
  Over recent months, we have intensified our focus on developing foundation models that deliver exceptional utility and performance. Qwen3.5 represents a significant leap forward, integrating breakthroughs in multimodal learning, architectural efficiency, reinforcement learning scale, and global accessibility to empower developers and enterprises with unprecedented capability and efficiency.
24